var/home/core/zuul-output/0000755000175000017500000000000015154760033014531 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015155004677015503 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000375776715155004471020310 0ustar corecore9 ikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9GfͅJ~1i.߷;U/;Yw?.y7W޾n^ՠC4%_̿f\ϘקjzuQ6/㴻|]=ry+/vWŊ7 .=*EbqZnx.h{nۯSa ׋D*%(Ϗ_϶ݬvGR)$DD D~m{]iX\|U. $ॄKЗ/83Jp ώI8&xėv=E|;F}Zl8T*v (6pk**+ Le*gUWi [ӊg*XCF*JQ0Z%bb oHIl.f/M1FJdl!و4Gf#C2lIw]BPIjfkAubTI *JB4?PxQs# `LK3@g(C U {oLtiGgz֝$,z'vǛVB} eRB0R딏]dP>Li.`|!>ڌj+ACl21E^#QDuxGvZ4c$)9ӋrYWoxCNQWs]8M%3KpNGIrND}2SRCK.(^$0^@hH9%!40Jm>*Kdg?y7|&#)3+o,2s%R>!%*XC7Ln* wCƕH#FLzsѹ Xߛk׹1{,wŻ4v+(n^RϚOGO;5p Cj·1z_j( ,"z-Ee}t(QCuˠMkmi+2z5iݸ6C~z+_Ex$\}*9h>t m2m`QɢJ[a|$ᑨj:D+ʎ; 9Gacm_jY-y`)͐o΁GWo(C U ?}aK+d&?>Y;ufʕ"uZ0EyT0: =XVy#iEW&q]#v0nFNV-9JrdK\D2s&[#bE(mV9ىN囋{W5e1߯F1>9r;:J_T{*T\hVQxi0LZD T{ /WHc&)_`i=į`PÝr JovJw`纪}PSSii4wT (Dnm_`c46A>hPr0ιӦ q:Np8>R'8::8g'h"M{qd 㦿GGk\(Rh07uB^WrN_Ŏ6W>Bߔ)bQ) <4G0 C.iTEZ{(¥:-³xlՐ0A_Fݗw)(c>bugbǎ\J;tf*H7(?PЃkLM)}?=XkLd. yK>"dgӦ{ qke5@eTR BgT9(TڢKBEV*DDQ$3gFfThmIjh}iL;R:7A}Ss8ҧ ΁eor(Ё^g׬JyU{v3Fxlţ@U5$&~ay\CJ68?%tS KK3,87'T`ɻaNhIcn#T[2XDRcm0TJ#r)٧4!)'qϷכrTMiHe1[7c(+!C[KԹҤ 0q;;xG'ʐƭ5J; 6M^ CL3EQXy0Hy[``Xm635o,j&X}6$=}0vJ{*.Jw *nacԇ&~hb[nӉ>'݌6od NN&DǭZrb5Iffe6Rh&C4F;D3T\[ bk5̕@UFB1/ z/}KXg%q3Ifq CXReQP2$TbgK ء#AZ9 K>UHkZ;oﴍ8MEDa3[p1>m`XYB[9% E*:`cBCIqC(1&b f]fNhdQvݸCVA/e.# Okx܍>М>ӗom$rۇnu~Y݇̇TIwӜ'}׃nxuoỴRZ&Yzbm ]) %1(Y^9{q"4e?x+ [Vz;E|d1&ږ/0-Vb=SSO|k1A[|gbͧɇد;:X:@;afU=Sru CK >Y%LwM*t{zƝ$;ȾjHim @tBODɆj>0st\t@HTu( v e`H*1aK`3CmF1K>*Mk{_'֜dN${OT-n,'}6ȴ .#Sqη9]5zoX#ZVOy4%-Lq6dACYm*H@:FUф(vcD%F"i ' VVdmcOTKpwq.M?m12N[=tuw}opYG]2u<ΰ+a1tHayɒ aY(P*aaʨ@ΰ<pX X{k[%Egl1$9  ֲQ$'dJVE%mT{z`R$77.N|b>harNJ(Bň0ae3V#b,PY0TEu1L/]MTB4$`H6NI\nbǛ*AyA\(u|@ [h-,j7gDTÎ4oWJ$j!f;嶑, }t&&\5u17\I@ 5O? ʴ(aPqP-K4<'$VO"d.wEр%}5zWˬQOS)ZbF p$^(2JцQImuzhpyXڈ2ͤh}/[g1ieQ*-=hiך5J))OrT 9KaͶ{߰+ednU$YD',jߎmc\cN#0"",tw>]rύW7$߇v{vzM$VbήdsOw<}#b[E7imH'Y`;5{$ь'gISzp; AQvDIyHcI(>z`{|puB"8#YkrZ .`h(eek[?̱ՒOOc&!dVzMEHH*V"MC Q=1Omsz/v0vȌJBIG,CNˆ-L{L #cNqgVR2r뭲⭊ڰ08uirQ qNUӛ<|߈$m뫷dùB Z^-_dsz=F8jH˽&DUh+9k̈́W^̤F˖.kL5̻wS"!5<@&] WE\wMc%={_bD&k 5:lb69OBCC*F!6~ö9M^\r\ߺnqZV@z%=\#|-3ڝa$ΫM|-LsXY r# v&讳YE 6X̀v"@L'aE p6mD[%ZZv'#NC5CL]5ݶI5XK.N)Q!>zt?zpPC ¶.vBTcm"Bsp rjWhUuta^xN@˭d- T5 $4ذufw}}1L @5DO'h‡k;q 7= `!6зd B0C?]lja~ luq=T#>1k*Dls6vP9hS  ehC.3 @6ijvUuBY hBnb[ Fr#D7ćlA!:X lYE>#0JvʈɌ|\u,'Y˲.,;oOwOj-25Hݳ7 li0bSlbw=IsxhRbd+I]Y]JP}@.供SЃ??w w@KvKts[TSa /ZaDžPAEư07>~w3n:U/.P珀Yaٳ5Ʈ]խ4 ~fh.8C>n@>T%TDNIGW .Z#YmDvS|]F)5vSsiExţ=8#r&ᘡĩDȈ\d cRKwUS=~O:*xj Y %dRwoJarExfKB4t@y[6Om *#n/Q!m b0X3i)\IN˭% Y&cKoG w 9p/Ǜ>Y+yn~F8I !6WB3C%X)ybLFB%X2U6vw8uUF+X|YukXoa%ĉUHSR0=Rxxl3{c` 1nhJzQHv?hbºܞz=73qSO0}Dc D]ͺjgw07'㤸z YJ\Hb9ƉwaB4s=Oi$LXIG zPzMD{]4ü Q̦ Q^Ղu ;` .Тr yFBQ#C`Jyn,m9~w{gV!hp gX|`tXjpgjpF)*=}C#M_cE&mި/?=9$wndX kBb#"VHآRyGCrB{cTtL!!,CKYýD({͘JA@-539gj#':.iȚ݋.+?}643Xs-јAf0ެu4]M:b3UiF Óaip||XTu;K kդB>n~8'hYǬSiYf)<(j%\IG.`ʌ=}㻮";?ޮ}Qrb=OI$(_F:lb6sJ $ oa"$d#HXu`\mzޮ63=nUu u~xUF2Q~Ӈ/( [ycy`ðSmn_O;3=Av3LA׊$AZLk;3qUlWU Ry==qٻҦ62L0ډ"ܺ_z9JNȯ=@oUI y4z%lOONRѦmDVmxюݏX}K6"Qi32\-V_kR(I-wtSJR^m{d a|y,F9$^@mdH֙toN1 < ҷBq/ ۓ,j|z6OSu;BKŨʐPqO K\*P'Hl4 ,ŸqMHDCYZz Qnz܁$Jp04ȴIL΃.0FiO-qy)i_TA|S2G4miBȨHM(2hys|F 94 DNlϒòκ-q|xC ,gKDzHR%t+E/wd#礱ºȄWEz o\JξB.wLKZ39(M +(PWՇfR6#ю3Ȋt ݪbh]MTw䀩S]'qf&)-_G;"1qz퇛0,#yiq$ՁɄ)KٮޓJ|̖D?:3mhW=rOf'/wѹ8BS8]`;=?,ڼ"ϴq*(A7? /W= #^ub"6q f+=^OI@߱^F[n4A#bYѤwd)J^Z{*ǥzw73LuaVad=$6)iI gC~.1%YmҪ+2gSt!8iIۛ*JgE7LGoş\bC}O i ycK1YhO6 /g:KT sPv6l+uN|!"VS^΄t*3b\N7dYܞLcn3rnNd8"is"1- ޑܧd[]~:'#;N(NknfV('I rcj2J1G<5 Nj̒Qh]ꍾZBn&Un' CyUM0nCj.&Oڣg\q0^Ϻ%4i" ZZG>Xr'XKc$2iσֹH<6N8HSg>uMik{Fm(W F@@{W+ߑ?X2hS4-=^YgpUHެbZ!y!ul@ڼ63" ۩:6=TZõ$E,ϓRV|G&$rr;J TtIHFE=RȬ]P pLm|?$%>Eü%mWO[>Xmw,*9.[G n >X8Ī;xW%dT:`ٓ~:QO,}j6j!yڦʲT:Pqҋh] H+&=>g| Z;D8ܶb:! Å{2:+au 6:!fF+0#+̬NY"!6a7#񕪰%:r|o5Znڧs?si/W qEU馥˟^_޶oڷOj'?nc]Rn\t3^邳塨Lɏ"8k8M~?M}:L@D+dˠUHs[hiҕ|֏G/G`' m5p|:9U8PZ7Yݷ/7cs=v{lLHڕo?("lX-Ǐ_whaE 98 (oѢ/Р΅ 7ցl6618ł_1/=fu).s¯?.S[{'g=Ҥ):d8h\y6]t1T7IUV:;.1& ,5΀j:<< +Y?58In'bXIǣO{&V\DŽ0,9f O_"[l:h¢8wݓ19\:f6:+ .3}=uvKc ٹeS<>ij(o'ciS<{1$E[nP b?8E'xv[K+E{,Qƙ1*dcs_Z'407|qBOgYU|U--sG8`u! qGYܷw;ȌCPce*CK*"pXR[كrq IH!6=Ocnи%G"|ڔ^kПy׏<:n:!d#[7>^.hd/}ӾP'k2MؤYy/{!ca /^wT j˚ب|MLE7Ee/I lu//j8MoGqdDt^_Y\-8!ד|$@D.ݮl`p48io^.š{_f>O)J=iwwӑ؇n-i3,1׿5'odۆ3(h>1UW蚍R$W>u t_uM Wi·yT"^'~i6֬:v~m!׭dֲcUh=Ɩ9b&2} -/f;M.~dhÓ5¨LIa6PnzɗBQiG'CXt!*<0U-(qc;}*CiKe@p&Em&x!i6ٱ˭K& FCfJ9%ٕQ·BD-]R1#]TROr}S [;Zcq6xMY 6seAU9c>Xf~TTX)QӅtӚe~=WtX-sJb?U'3X7J4l+Cj%LPFxŰAVG Y%.9Vnd8? ǫjU3k%E)OD:"Ϳ%E)=}l/'O"Q_4ILAٍKK7'lWQVm0c:%UEhZ].1lcazn2ͦ_DQP/2 re%_bR~r9_7*vrv |S.Z!rV%¢EN$i^B^rX؆ z1ǡXtiK`uk&LO./!Z&p:ˏ!_B{{s1>"=b'K=}|+: :8au"N@#=Ugzy]sTv||Aec Xi.gL'—Ʃb4AUqػ< &}BIrwZ\"t%>6ES5oaPqobb,v 2w s1,jX4W->L!NUy*Gݓ KmmlTbc[O`uxOp  |T!|ik3cL_ AvG i\fs$<;uI\XAV{ˍlJsŅjЙNhwfG8>Vڇg18 O3E*dt:|X`Z)|z&V*"9U_R=Wd<)tc(߯)Y]g5>.1C( .K3g&_P9&`|8|Ldl?6o AMҪ1EzyNAtRuxyn\]q_ߍ&zk.)Eu{_rjuWݚ;*6mMq!R{QWR=oVbmyanUn.Uqsy.?W8 r[zW*8nؿ[;vmcoW]"U;gm>?Z֒Z6`!2XY]-Zcp˿˘ɲ}MV<в~!?YXV+lx)RRfb-I7p)3XɯEr^,bfbKJ'@hX><[@ ,&,]$*բk-Yv5 '1T9!(*t 0'b@񲱥-kc6VnR0h& 0Z|ђ8 CGV[4xIIWN?Yt>lf@ Vi`D~ڇŁQLLkY <ZPKoma_u` !>Z;3F\dEB n+0Z ?&s{ 6(E|<ޭLk1Yn(F!%sx]>CTl9"و5 |ݹր|/#.w0ޒx"khD?O`-9C| &8֨O8VH5uH)28 Ǿ-R9~ +#e;U6]aD6Xzqd5y n';)VKL]O@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ;`uy-u.M>9VsWٔo RS`S#m8k;(WAXq 8@+S@+' 8U˜z+ZU;=eTtX->9U-q .AV/|\ǔ%&$]1YINJ2]:a0OWvI.O6xMY0/M$ *s5x{gsəL3{$)ՆbG(}1wt!wVf;I&Xi43غgR 6 ݩJ$)}Ta@ nS*X#r#v6*;WJ-_@q.+?DK១btMp1 1Gȩ f,M`,Lr6E} m"8_SK$_#O;V 7=xLOu-ȹ2NKLjp*: 'SasyrFrcC0 ѱ LKV:U} -:U8t[=EAV$=i[mhm"roe5jqf$i>;V0eOޞ4ccc2J1TN.7q;"sդSP) 0v3-)-ٕAg"pZ: "ka+n!e߮lɹL V3Os\ဝ+A= 2䣔AzG\ ` \vc"Kj61O Px"3Pc /' PW*3GX liWv-6W&)cX |]O;C%8@*Z1%8Gk@5^NtY"Fbi8D'+_1&1 7U^k6v읨gQ`LRx+I&s5Www` q:cdʰ H`X;"}B=-/M~C>''1R[sdJm RD3Q{)bJatdq>*Ct/GǍ-`2:u)"\**dPdvc& HwMlF@a5`+F>ΰ-q>0*s%Q)L>$ćYV\dsEGز/:ٕycZtO 2ze31cDB/eWy!A/V4cbpWaPBIpqS<(lȣ'3K?e Z?ڠ8VSZM}pnqL f2D?mzq*a[~;DY〩b𻾋-]f8dBմVs6傊zF"daeY(R+q%sor|.v\sfa:TX%;3Xl= \k>kqBbB;t@/Cԍ)Ga[ r=nl-w/38ѮI*/=2!j\FW+[3=`BZWX Zd>t*Uǖ\*Fu6Y3[yBPj|LcwaIuR;uݷ㺾|47ߍezk﯊溺^TW^T\*6eqr/^T77WNZ7F_}-򲺺VWQ77V\_v?9?"Th $LqQjiXMlk1=VzpO֠24hf 1hi D{q:v%̈#v^nBi~MefZF >:/?Ac 1M'I`22؆DT!/j璓P åiw@wgRCsT~$U>ceއE)BI>UljO|Ty$ŋrwOtZ7$ "i 8U 7bSem'k?I+/Shػ߶%_BG ``'3xqd &)qkx[ERnYʼn+yxWu (@pϓ-%r f7 զ(eqJ/=iePՠV`U!J)JwdYq%7M9ơ*J%.ExB 5/}10+nϒOyTH?!^xo?X5S'R<*CUEd!v:XxZ7ay[n#C*63R+^Gz2(h2\7v9TuA|U O=¹<>Jq<A2=fjvi}BgB.)`arC%J WNI={:%e+ONݰB?6 7B^_>w<P-[+p#˕1 I?y#^𺸲lC2^WDKt ߵCK8j.q#"hҚm,Ÿq>hE^Ows ]zIlۇ= z/˗\0P FM6`= rDZ8LHv6 q.兖oWIӒH'73"a`I?_<}}]i&=}Oiض!M..h ah BSkg0tGxb`u/VZ!9w/ 嶰6d$|6ҴkZ R(y" FBx39B.b>r'PXa:?wzE2M뺨NjRꪚk- ֺb/y:Ey9Ja-FqW-pel IפT׉&}KJO>nR[U#_r1' i1Q _.:g~uVh]`pA!8) b ~A,'Y3^w#} Ddg廞~>9a9:zJ %ÃdT)q'EY*/E̓*_OyB5=Gs&wL7<\qIx9G-Ksdx2ϲ1o n"aH{zeB Mk^ 9Oƶŭ8~=4:MFoxB yvQe;p΢<7>808/8NV77Mvl6fL`^8_ʆ. Dkmں0rZHb$j a`3k8˯qY7l=wg䤣uΛ%8 UKKi[ޝN:23'0$jʐRO͒M #w3IY<2!jKG|)$_'t.ëaaU7))8~t"ҸVwpiQݙJYIV۰v7~`Ix- K݋c>ClK/(f WV8{wޖHll?ge3W^ω# / -1iF œP4F*6@LBR_*O ) C= ,ۗS?qum>.뙆K54|lN$Pn_{6LR,+8 wki]S Óe+UbyDX`kٴ\,V͈LuW 8/O~};4.莏q#5)%I~-d$6uDUtN۞kqf>&`Kɢ`h> Vl k-m/Ѥ(pVarMɤ9O'XРAj5ƮG :\k2}FD<".7O[]4ڮr3lD=>^FDM G$RpoCh3T$L!i:難g-UDmhGch"{ ]qv5qŁm?i,s!Rrg_tXt")W fo ȥݧ̦ѵsVpwx)RDjۓG?= 4@ζK @K|?>ᷭn-M런a-X8Wsn[˲3G& }B/͈;  ] Ow"_Ueݕ7A/ AwpbNuޟf%5`ktw ba,UG=}70GJuuc zHIcs,Ui n u\fYV&pM]Ob.SYUy-[?nZAR=P'U7,U"%6ZĭgD{zHj~"CU+>ڕD&:/$DT 2 \Ϋ0ǖUZFSp?FAu~2Ӂ4Dq͘a>k-gw?24lrzl 睃\lgO~e6a'lO .sca3\ݗ-} {/8m.΄֪5xEG]φ5WjNaV;ͯ )~kHp(.VF#}~[ ef('Vy#J>`,='yGy9O\ŵl."x&#,z|v?5~6e}*3vc^yq%a <$\gf#jJ6?Zl{ 'rZ"t} 'TEs G@]}0k/(;r?"]Cm"x mE\3nfV*U6TN .s bmO_>hON>ӳ}FV`69fYE.CX49:em1l@~q,n=Sv{z{Db=E s&-pQNoLNj^úّQŤɱp!n9ODZXE~ľ!nfd OgٝQvt|qFZW<׋xy ioB9ߡ|PqP ߱AXa/bʦ>'!vQ"҈z(6,UuYBPN;qr ^/J`$c hX;O1x Puʊ4,A$ݣ{(Eeb;HX~DD!-AHE(۫jH4s@!{΁j]sa/ t,Y5*YO: !!P\rb^#g@/]_߉ptl5tN2 FMwt/q΁N.HлP ȓzO`y0V؋ ݾV$Hx5=T'з3޿hhT0&4":7-c@7^ O.ahWSu;v`׍ Q KtN5W t [t X~yJu^+45niB}uɌn˂P qwiNʥٙ0$7s pNʫ%:*]6/)%M:^HH@ї/Uh۫)%z5[E+g!k҆T#5C갞ޅZ \W8; R]タHH5T.s(^O2hJ̵7b!zXcD>7w\KD)S,LZ`Û"U)2$IY9ƐL ؙ>v臇X\ʖ{?±LfIZ+>SWSa&FK]UdyYaƳ\m8"b|JxmK*:{1ʹ {|yn?TDIJe"|ZX!l'1)[r:bf!#]f7a6OKZs4$Рmh4m7!Oc?1;\zr֤<{ ?Y:{emsAdI(b+ =4 j\: ',Б86ިD| 8=7l_v_5"w '3bFA`X`^#^dɼtbQݟAԯAvI{prHnD?>A8` p!~]3 33r)LO@B\ |HH{F}4@ Z'l\$Ż'A~G񝎨ql&ZȎ=ӷ %^f~4LM;RMf44tIhPW?fQ'GLGlCX~ݤ#j9mn_u/$ (q}N/_OS칌5eQ2 ia\gOg㚌@G֤AZV{BKvcnnn@yԋr ̮CbM@!lՐ:9V :);1Ow1l(a3ʁe1t3X<җkG\;p|g!޿ tB퓃_"-8bz(_k8(&K\G~rbU0 Q6-{4g$wba4rPYnoP{E\5/GŸ{( Wa0.+ ʛǜV:q4Dz]7)EU `uY\rV"q0mOsJrkP%S 8ƖwӶ7ܓl1`m/B[Ե+m U΋A4  AI7b8fPJ(ۂP=egʶ Z%ڂPk{Bj=Pk BUB-'~3 Y%قPg{B.s\N=wcp?&q4JvFժ&D2?}3cA6 /]].z1h՗i|@fU^gXq)c: Z.@Qk‚Ӡj1?+ 'F+:oRxvncdBvd8dj"/t|8'.y^g)<C 򍶤yc"[x~ݾG2DeԳ, sYY{Ƅ}~¬qmK}0h`_B%s2>xa2AU;78wLE>$Phz03E@v8l>[v90O,8^Lˡ?R1ʒ6%ϡ3>J2|2w0gʔcuX#*d`NalvM*En.=ޔ+v>ē }7x. 53pgP3ϮO,baCÛK0_PS'\K 3$K&q>tNA3w\ `B9|2m}. /lxgWf<< Inz"=Cg7fY8xhs?ހ> HO6lh: SL9氳94>:0>"!~/IyVs\y`vpMpFA x2hQ n8tmjRn%3^pɜ^Xdϯ @F 0"S +S2< UUw\wl3XѷaJ d=7WQl$1HG cS^m VJ)aa\L;w'޵0TP493I}_$\ujP>f!0x ~’O9 ZF<ς51DH:R;?ߠst^%He,{qqA[m|2XO8$a0!֬;t2nH=-,!Ԃ|x.=wrT 65Au3*q*tܲg?]Pe _"ޏ]'Ca>p+D݋XF냕9 `|d<]蹓4P$"YNPs)g%P<=W)*hd(ps \ : Ixv#<=?[ u3zk^H[~^*avw)&RE\/>Y7a2Y2޿:p'1pfdd oZ>ERUnxmȟyDj<nOSqm`(cߛZD)YeR O]䒷$vך`YRi邀puRzp\-n*4=T^za;Hw3=H]R=Hao 5N+;ܮ5mQmR 6eMKA=LWi%L'sDndҥ7ghxV{LM:gc_t<[f2y${0o@Smxul4X6ᷚ)OU1p6j@GhQբuuFjvU{<1MEZ{Ƣy#[X/7StPH4C1zh@uˆprҗM񔴕1d# ȧJk &XRx%hl1AoАj=,J:,9Adi&Ck|Lvi{+Rދ!¢аp%g%܂.?&NÓ,,\bcᗿbWGpXu[ vƸ<Ҡ7,!"YwtpR** BchprfJul7dGqotmm+K "ѹo\`"쾕uPa쟳+gAZf钛(5pxVq&Z#*HeDb͐/5C)R/Eb2AZǮEU&_:`Ahۂ8,|NdKk," .qeIPK9R.,`s1ut?wвt5/KMة3%<ĐRFkpD`5s@b\} ecױ362Ŷ9,tMOfVܘ8ݫ}O{i([O9Cg(yv|p Jι) |\bSǽ!%4(,K+'5GH&kZX:Zo\LQ$4sTL QFH n:aLKx ia]`ԤWb |t6r):zezeYeĚ@!UJ!#<8D;)Nתv&&5EƍJDbN5>8S.DJ:P\"E,k6 ,Ps(҈IǻL o! Ֆ(-h]gUrk^->΋ VE߷NlRQ)˨*.8 3MskP]մl{&M-MT\N;$+ %dJ%ItA+Lj%ؕ`usʍq6UbyzF ڱ1Jt\`~{ਙ|BP([c Yڤ|P+[:q>..=8:/\ L "Zy[-}{f eD% %E0ːnީG*J}ژ7[lڎýAά5c,s )f?rE>51p'@Ѱ,gy-Çǥ8I-mEsWeFaFi4"u^Kǎ$?.8NX  jJoC3^2*V*X[F8x.ǦlY Y1$LS KԞ뱔&[TU#0DG(U|MbQXr;s a]Ƨ~ ν_!S@;x>S jzBa[Vw!3M+VcDUZ 6^qo pS,+{(enzꭈ%7"h2x|G|pM6ѾTАh7g)97Rp“Z Rnr6u.ȸܷaɶO{؇lS~oç'!w"3}D;듦RÇ9Rw/Ocل˟ 'ă޽{}O/o~e˟yw?i{;\%?˧ߗZVO>˗m"r%|ۏ۝XT6ﻰͽBmZ܂O,n&W+M1 K&T->_Nvlߴ~/iYm_lٽտşjV֗׺v44[Y~ o5xٷx=rޡ5%1Y1jlm5)"qU9B+XјsjR.xzu{7Z;tsJ@2ɔVפWeuC m4jR5R<.8έtyvuk{zΘ{M"#vBM*zãBW&zdö^%upL ^6/s,҇Js6>.SZ;a7XY6;ԃui)iXR#>=|ѫG!tHǛp,.:ٝ$DFRB}dw4~q$$72 x"*&#Sk3A8_sLXM@t|8\%IbA/r`gX.EN =Xܖ)-/wK]LG#!+Yc$ԛh5 F2R!|bWx+mXS,`肣vj|YP 8skˀ.pҹ0"B?l+ZDw;alt" (ۯً2.8f/1=2(-coe& A'(Z P-5SPpK/L:bZbxu¦́%9s4)~yG޲6=T{tu$=9i,V@K)k[NHB(<>yPHEϚOv jWn[i}l^7𔞹v}RW\%cGu_1ƕR \;a=-@߽Y|k{BOyU֚P|.'\N򨒯%+&Tdk Ms,:/jWE Ζ~J& 6pvMI}I)w99dCw(]מd{Uri >Е`ڴR9L5Ӹ4bTANoc`\*qw1-pm) Ur!Tn':.(ŪLkS%7Q7 Q$92rJ1}7k\6AᎫm0t5 CV&P%&3`R,B􅭑!hۗ%8` 5Afx߃JK^Ϭl=O2XfԎ(aL+9)Ȓ%HYr4(=%cHWQn22N.8:9V&uGnovIu?_f0Kgd}&e,w6 iT9@}1f~߸e)tNsH>iy K\F y5c)5^O28xQ ^SXE|쁣cӒ=(Hƾ1 rpPJp8d* =Г>ƙtճP~H^z[.9o%GpKbQ3auJJ}q㝸y]pL`)DD MD_/{eS킣W<C|M['+.5e^RzCh.oegm?M肧6jiܗT4nrЉtk>xJҹ-Ք.zߢA]'!(OpRoR~0 IgQ+ea"tWYh἟?т,$D! x4Q#k#Q|VihZi M-1`뇳O2m%Jcf N)sS 8j_/LB\?WWhTdL@%cp)A]ddc:L~~ysrσO.3x.# v NK&N,? 8VK`??//:CAGѤ'KP, Kay_MQg?8>:JN !ri(c4}P%2>j]j'?A>9} 6d4;qadwQu9T+R6~?#w4Fʹ +P߆&<>{=ƵbL=஄m=@4'3*N"S=KGTv-Lh\ЌY(kiHbYzffZKTcSU_0>gPBW<_E fc?o>XB7ÍuimNWu9_!wU!W_T`mypi(;7Tu uīþu{$i%tkρ37g^y@>'JM.qqeTs}HOr2@_FWe'D]7gȔy0fgs_NeGɄ +:g9/ї>(g anϜtP ʜ%q J %|q@*(ITY!U2 S_2a6T[eM阧8Z0iu^G.h㈾.PKT!hcS`˵aV6vb޳i#ª 6<h;mIUAR: nv, p.I)-'8ZEtX&A!%KG&S.(-iTGlhz#хjqu-#d SO-[H v)$6óER^q՛.~D"c-)ӒL?!6pԬ^{mq [mXJ5ȧ`| Zh~0E_y1)~>~6Xҍ/QH94ĕq ۣ_j-G FEzʲBU/BUtFbY@$A7MEq_O%_*  CRRԕ-{w O_Q19HEIGY"V;GJQ&~_᪹Gײ6S3̪bw#6mxfMvv7=6:};O֒;F/ѳĥXbaCf|ϫ5Y}p:a?ӘR BZyقq?lKwiqɽj ?Z c< ;iDګysb3$HG n\1#m. sI3g!ҝ w>I(|RmT2tPWqp,=S9{Js&upү*NW܅_t\P-/_E%'؛( F˂6rg J"\)M=F4ƙ@15]%Ẏ~ζ7 [bh)x> T&e=#;oD27ج8#RQ^LCT0< FHQWՌҌ?\3@G>f,K75i%U"V#I圶e" drF-:W֌Ҍ?\3@(X؛ 6b!XGYL<,| P-[2Y @T#&M>dM/)~IWJ1}v9﷌.G*X{F `~ ֯azW}SkQDjǥvq(9G̃ #SłvCi;[7c[TZW4ޜ}C7h3e%;W[GVl)\ǿix-W9EzyqISB5$H|sp3 :ad|,!*<~-&-Z,yh y}JhMNf°5[A)i[Zbjqwl׾?.kmw}uҥyc'.p).4UXR"c Qd3'1[Ygx~n"#lp@b>0XwOKJ<0:2wV"^ Ct8"C0c4X@K!Ǩ$R8܆$"&N!cY%TaZ$ am*n O PB2)xAx DY˃tAH<8ib i4DũrGXr#@  ) :"n1GXH0sR.~1 .Nȣr#!jG+!0ǁT!iPSoU^iLhZ̈́i6ة.:U̽; QZK+:iCK+Ohbp)OT* D(H0QD,E!!2tCxkV@mkCJ`m֘u!2Pĕ;Nn@8RA,8gd#eϓ8ekpǫr|BB^]?K4 k?kڤ5:AZlA{N&/[RXsTvX$({e 2x@C8c<éʅ;qZ@ z%6<0 C =ǜ- 121F#DKT{VY5ش.LJ ǔA2`gR@:qAEy:`oOS |A QT1DBDT9J081Zfqj6On+RiIiOQCn,]bI+:p`-72YlpA; "[ZIN"j=l7--*-"Y)W 4D};BfLdL V@BkEMFUnD\SIe.I||Q"jM$wG_&VgE?sNE?v_Eë9:߯on'dvٛ~MuOw]vqrMݪu4m2w1C&(O |>#2˫fҝ=u7@+~h-vs|ώ8Z{q' }]bޛ#7gJW36VŅ"WK 6Ɉ6ZT\#9:hQi 䈴hs֞K FFܕrGi\eQZ~?]Pe#NCћA9a8uW .A\p{U5"Vaϱ#+Օ$4aZq#TTRRT՜K" iQiC݃rA V:УqsCпEH]`Bm `j] lj0h,uHTڠ֬ "Zhu@7aakO=oyRtR~;;,4Ik|{}?3KY}xӝxuN.x IaK4T3o:B<z({1ˁ|k0_~*sIǕlNLZ|hBb_??N.ܳ;\'W(l߽':Z|d1܎/:E;w/w#4(ԯ^4S˳0b} /ueAާ3?3>kh*ƨ㏗Wrwy;&G0o6ut fg3:xvo$.[U:nr4Ѳv0\S0jWнtwB(Y C`Rbv50 VIppwJp n޸Zxp4m's*.~|rQWGۚUխ.7_bۯ8Sqޙ4[?@qK3Vl;`mv)2KW h}7A=[EzngĤc?>iK'=wR;W ı1H6~5s;^e^%j^zp/JX˸2X^'{4ѫ10J9zpED뱪v P4jO&Md:u2aSYnL,{o ә5W/.>% |Ur|}SLw7f&P =ۡӪ|Uz|,ty/ 4UռM0*;eM(}52^j8.pZFH8Cz7w=dT۷ZC+0\Juɺ $aL+-i{;%b4nMm U3OsuO?qZꆠcPJQvJ/O۟5?㚤Q"f++W"5Ndlݼ[{lCpْJW 8^nL(/w-Wi3Y>}Ns"kFNpO,B&,XV$Urߓ2yOɛ )?X0zg7S}9BkwNk>}JDǥWܕ}nOALwŭJ8{#j`oOa/pJh FkY\tVvK۔} X?8`mڔGʏr Ri߄u&5^ 59!$g-h{pX/s\)^2*7zQo%vV>5>2V:e\|,czyA=e2Ynygw@򢩾T_~aU"8!Ξ1ۇgOg!愭rsgg3W?2Ϻ_M:;Qs{m^pe{Mn7w׃[4?wo1lky᚟7/g?p}9b_2x⨢euuU4[`Sq<;(Jf<  Ռ$\$gu G\ij=STЁ* ׵4uae \ DxY%z]|u.;qQpD Z=Eknnlt8%Q vNq~|}@dD<}`_h%h0Ƈ`#PA\)xN$J0+4r:lTBX] h͒!rDJ2enFT?Q#!_EAfRӱFRI>$u˺8H,@Yz ntH*иˋ-f6 rS 0u%U^6>oD>/N3l)&Okzz/ȷϥTSKQ9h5@(lnH9al%g\m(^C!BrOm^#V[Ic.CR#6xz Tl3ՠkUJBғx׫AҒq~B$ 9xzF1TSbfL3Az7J bFS D͹JR)aXobDͳHAER'TIs¡zz5a4 єrHxz5>*OZ@mh^!gCE3UPH6^o^Р6COZ.+(^'UƗ-zQ2&/1J= kgޠ]f0$ 7=6}] $)D'JTCk+tT(LR|lxhN醴92ץy{j̢xջv5.uܹˮeA6lv_%T.4et3N6)XSY5&D3 :'2#:I b\%>c~B:e5ŽPa0 !A+J(UFN%v9 se'YmrO"-"PǡZ9hoU(03KE&QɔL ZQXRI`\&BrYR[ƬIN !UkT'R19Kj* f(UѪW9X[o )3t$#BiAlig0a2W(#yz5{,ws] ~`B'eݯؒҔ8- bKl6?V{` ^aXlץ HUDN0v)MnPࢲ-6z05I~_=Sp1k㯄RY-͙ri,، KQQ)E*߆#QP XvPoCo68dLAn+{ZܻYW5WieQJ P]#Sʊ_F  &_[9X(R(Ec "ŴFh v/ -~+{ToR"* h]x)CjL@!7[[52)Ihh}L"rL٤MqSUIߞ[Uh>EJ.CK1 Ч.Wکֺf;DAqi3:55-dZVJi5iz7ЈV-T֊8Ho@BaZ iɀH8"(~eUP,}zPPX[4n0v#R))G6+8 )΁?%A[TB&IHo/@JJ|=)!; 9DIvj2)AR ˷8f?(!&7^i:@K !NmFY!2Pu%)ad{mnķxxAfx볻̸_5o/MG 3m)NRs"bGl"]ؠn %Ŭ1~9坅ևM>G?Xl6A{UK)m4Rq3 |7} tk7xɋ 7?Hj9h7#BLݾOO%vx3,mx-Pp 2iuN,tV:I KXed_V9(G(UZy`/n;i2n_:?`aZ-o5Xg 7][_|x4z?f&6x/<}|Wz&`~If~瞼%jȼ|זC7zѢ,Q"C%G_(Zom?/TYV9ˬL;:yE냤 f?#ӫZ0#QBbܩw[QgA7PFW;/_KC/op/(:< H~1KEw]ZBDe1.gfP ˭&]m.h墒v1]ۑ4U:Bvb6gF oBm)YUysx w*(4(6Q Loڍ7N[^H:wh!}/"ey}!J$G9*jBE Rm/?He*/5؂Il,%ŧV#5RQU,)OfE FMX$158BmQ!(Iry*'R_[8קOd:nfsM?cCB dڍ2aZHݩB%J&a v V ZeqP")cR:gk:" tK9XI'pȱk2AZZF/tvڦ%&zi6\- :NFz#s¦2 {SЌi|lW ooհ9_7 a:$r  U ܹ-AtaV Iv#A@L ݗuX^j!W. iV2VU\eNQZ= s5RXl5J˓/|UN2, JIHQj5LIgW>J>YtnN"4Qhr*qee*-{|#.{N:fYKdGHo`>eemޤeevlكg46 yȜ*w>m}8xic=[ 3YfR9?_V%ۙYZh B,H%c SէOupB%@E-ʹndȮ F M##zmXحӺhPZkD 6rnfJ_7s\ofMCUo]MQh5h(Ai*RdU*Uk@haf. #q6y; 4PVWq\b5aQ4F@cޤx߇ kOh;(PMHVp,M! IHΡΑ|a( &FBHoH*|!Ug박v@q" 6E6Ek"rie>!@, #4 Jv !f)HobD£)'BFS_ui4TM0f%V#AK~x<ReNqvOпv낶F$l 9n @heDApE@$Ho b72ҐKo.2|zS×T~p $ fdENš(#a iw oP3{m( vz+j&$; F -H'P,0+t@q,Ps PsjиbGv#}BZ-^.}Wo}ߗ?F! (AU|OȄb,W PE,Fʒu(\#F VtRV[#K@60@4fPGkϔwLe t쿭}T ю\CZd("RL/ì"mő@BI-;и& > (ڍ709kP .mb('C̭BV=J2 <۲rĔ.PHkynԠ"ɮ:r٢3rVy!RkhP$cn7BF\$I.MV& fhV# BJ߅yAS0 3pIFRr;kA#G5bL-H$=c {E$記&~ 4$4` %UXZ,d_P:,ds`0eX" /؅$dj Ƕj- FF#i2%mUie铤wI:rSpF-+q1|HڀB:cI<ӘfD(޾L.RIquQȊ+le7Vܣ6h+Y-0+WY)ĕ\~GZx!mHKp"QFz#hg ڸ0(w:})Ó0-Ho&ʩYПۿ! \ܼ>V#:_/O{zB%ԅ%%苐3riLQ$mBs6E(%X-j7B)UhY޿[y_]YD.]-Jr͆~"TD؉c  cIz dy-0Nsd2ӆCU?l0WW%h{ R} \nvϮGyL#S"ԂW>ߎs6  <0"|('`8+/=}Ta/(\VlTa^Rll5-sݣp&3| T"l_vn地o}R80r :ǾV+ ~/2ߵ\n.U8IRu=K[JOb;/յ,B+.0Й}VYT]ToYedy0(-I {| 0mG `mڭ Ko쇯lVܦ#N'鈓8aDpƆ(0;lvDoɐ\vGZ* GPBZA BtfqZK=PL}1]8Z~?.n 'y0,^[8Q{_OۚԽ?܂tހpkӯ!կ.-9ӛ).{i-nevL0i5.ѺT;F dTZ'2rc]ipw3r2X<3噹{\#% m8m>mOT`++9fF{xo8t>PZHd6'Y E߯ ~}3*n Ϧx^ƃ!7r+[KuW!TYyir{YGT}v oQ-=6tsz/wv[OwVըo1?M?C-5w帨7?Z[>c>ÖĿl4rT~Z#͘gL_0i>˽xa.x: I498,A0kFun~6Ll?-~3XfK 1}1|zwlG=o'*}a]ٸ D^xWǫ;m>gq|øJ;Mc+?duUH&SUџj|5#ۓDai/sˈiۻ=ۡypg;4 1/!lEu]_لf $ s Dh(H:VBQK-EHYC) )ܗ'؋ÝbhUq/jiꜵ+nWEm[힭c&iU9ίL7 zoQlo^Nvnap!P-vX-fN%{%CFh!b-p6ٸRVjx K?h(;@y(E9h3uEю](̜qm6&gu6ˎg'75BؚmM/LRT.xA# ]:7+A@T!F"J V$n!r8T`і$AFʤ%Zt}`++|M&,n<(b}vR=9oó0|X3KMgy3](Bɕ~ѱW3ٟ}mP7Ra.:fM9S]b&eG1/`N5cݷTjk3?5B`FL6^_K%̈ f˚kxKhstc 9HXRĮ3x2dڨ^dOUxi6Ȟ=}!tҢ%٣5)J+#a%}63΀"]G9]Tl| ?U5T )*@e#hbUꪈ⩬ e QXzMު(Rظ~e7RPgLpyŻ93.Q2+! nIRjZͯ/tz>p0zZ?S򵞲+.NRNy43syj=j䯏qЏʂF[mKVZv$U$t$V@3PLvL Îcܬ/bԑb4VR {L'OHxt=%|O򀟆l~]H #QUi&B gEiU(P28Omއ]<#xfٺ=Jq,9FKv,v+ ϯ-+HT\M--~0gЖ.I=6|vbϤeɣ2ȢjڽYˆɁf.f`NR\Ugqjd5v˥)jra_mM[zա)Y7鯓~T]S{>w'̎3jI>O'ag_:_FRL Ԗ^9%>Z4ß6 ޝl\V7镭Yi:[bi2uBॆ Av/uݳvkuBțvouBZ=lGˡxPH>Hp m4ɷF.B7ZEpD0h!=7@kuM3IUl;4_GΝ3;qν[y)dv/<zGۗ z>ʄzajeΝ%DwK;z^%WȘT-L6K޵`aͣwJhwbE3Z /. C~;kx㏚PM~.Ǽ})zFIcTBvVB됢V# /ڢ+A4`-l>Ў[DjzАKДɀ+ƴ0Lnf7EM tt]a>k&s'[#ф=rw믦]c&Mn(o,pըou9}-oI1M51P\H2=g$yY`;8.) VI+_r+YPVӖnWbG!uVn6[x磶‹LnQZ_ iN<4g*h֎>\\~OA^R##Jgq]Жv{O3Kk@M´8GcDJ8YSBYO<hLRbs) ]7 9ѯlJOpQd5/L~ 턲Ub efjjrgUȬ[ +0I'wMxbI3L_4~zo"'SPn}}eP+ۆs9Q[jܰLŸG5տ~[%ԣo0QߎS$oZ$A3&JU_g-0XŻluqV^M8G̍OshU ϏC6'SVm>B. қQs!K?޽{7c~Os3oyhgsj.Y{%}7WκkxoyץK6b !/7yltvNӚ_\ m4G_N>~`ҡ~fsWM҂^bR?nyqwl/=\lьGfćy߬gbNŭO-_vc "sނ,kVa2mVKIy tI|&xmYǎ6b6ց!~<.oN?X2L{yᢗd+[zDlA[ ^A±E5ZUz%x$ZN/ ?ԌmOcvy5u}{: 䵬L2ئ3B}ImS nTAU )/)6tuՆB۴Ƃ0pA' /Pætd̳p+ N8B 1W20:&1I 9A+5/1(Xq@yw$ Pd"EFͅLə+4Ѱ<{^;A87Pn66͖wƝA 8Դ.χxi/T01awPT;o6P!/sP6D":X K[oڔ)#Xy+[51EY>t7Xct,!=m9TB1", W"# 2vwY+Ռ!H1&xB vD td !N;׻"0Π$1$zMqi R6r$Ԛm _tz,3,3eMF(8TfeM[T*b-y-UNV Z&HH>o#NsH^1ES@Kn+8t," MQ=EJdp J`Q @'0*HXoجaX V#-"0$EELY[*P'' BN`|?$Qq[ 0ja#+(ưɅOBň CzZ U @c6ij%pk1׊ح D5hփ*z(`(Ky$i!Dzj\'5N=8EvʴeƲ<|›R D=G}D DK (tH3(A˂Guu(u!ҭzhJ7ҡ#Н5>d֞ #JOnQ!,68A/;#@.* bwY\[M!b!M P&T$nBŇN F !Kkt@9/SBWf( FVE`AA0ꕷAw][>p=fds'#n;ex;e6_JS B4kKV8Z@<_)q.Fv[G^SRȒl:,)ǦcZ^M[׮6Ǧ#yզSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:G@3iZMQI :MG󊊱ަRVhAmcYTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNTNT+h-IQ4tWbl:yz$M{/Um:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զSm:զJl:wKEjͯӼ4wv(})rS<_MOmgi_OOKx|49pwۣoAXeX  UXje!`%-7UsS XݒB< k-=KKOְ#rRJ@m + 3eBj¥, AXJZ"=.X'Xv8-,͏Pey5Zg-,,ĬSXET#H!`*Y Xʴ>n=BT٨'I!`UBr5y7n)1+%ŀUԪ#KYèeLy~V VkR͑˚" 9dm!`ySg6LKU Xôc< ͎qbYyd=?X 0WH!`DBj.Pp+R9f՞` J[XʑmU!`ĊBr*% Pۚ#Vrt)lrC]ZPCKah!`k zOA*,<eBM SfK+4BV %i)`ЂZhX!`0#Ș=1]{gncQZŮg4y-.?)/1|'V|UK=GXFkɑkRzܙ9zRFʖ ~3Zff 9-w蠏hVHB {MWsd~ /X0] ]eBW22ʭU~JiAlI &UkRhW2-+B֒*ֲpM1sW} (֓Z+]}?tOl5 `mʙ\p)tzٵ?]e ~te DWU[h_튠UtE~" 7ӳ lq>WoeFIzt#K'54gQ2qʍϻZm-C>a/8{tا֝ޢXz3D>Hcec6O?K!qoon]O.Owo.&n-lJ ~`+_Mʧ!& nW$>ve/>D/t6_t:?|hpҢ- ~{\_~4o_h"VHyho}OK4Yk|eדB]5`KnBAHۜᖍ/r6I3L[] ?zw1y~ v-8иA0ѱ { %"fqy1igmxW#%M}lau4rO4u#}lQܱ.߮m֒!=*sSޙ֌9h2 /bx&ʞ1??&:mKιo. #ֺ6&ےՠ'A>COn텪C CocnF!e5X-EaN&˄ g D dMY`1 B&aMmD [|`^'+V #C]W/[{(z-KD΃3}"ءr\y:}d.6y~*^3g̹ ڤHsmRdi Ց$\uA59[k>@-ڹq}oS; zG?慆 ߌ[r61k/ߜܴӤW+/9/.ghMC$_i}P0jy@ɰV7>Jz!4B7psۻtM![m?>kX3 +7Q.+_VH.'Mɸ~ʸ*\cOǞ7zp朔l[oϫP_oWƱOk}%͕is]tyq-[l1|/ح]^q=_~<d5]o.7ۀz8^MVG{t费}vݗ,V_>VlOմ[_oSl, +h7D7ﺋxusWK>.X!,\44¿;&DsJ>PGkm+GEϣ+b1bF^4"F@SRty;bxyyXܟN&^].$bCR1m?VVV淡! ?^.|\_OG'{F|zc3\n{S>^lyall7>s›xq{{qnDdRm<;S9s,\_^/קnpnO>yb0onzo~n//w'ZzdRd+^/˸ax89%3X XK:Yy^\x m׾*Wx5{p/r/~HZ]^v! i1͏H?چ+6OBڋi_<=Pcoml,AxA騧{Q-;0Y~qd0g$vsO_0/O_K?%jp1|6 OB{N%y*έLXBW >oNj~1j9+UdE! 9#5Q*s4T2kU#Ȫ'+4dc)5Re}\8G9|%˔Roe`69gi"#/<{n LTliocqccߤUϾ}R;c/]PrMCj䵢kE֊ޜR-XFtX$z󵛮4Unyם-u|8E* w<CT>.hRJJHe9EdiVW֖sڼ4 2r)iQZbrP\JʰT"h ұ i9>OaLJ#ɑ3M<>CnȠ@I )w|F=;ks3<7Cھ rkncͮy@1gH4l1 fejb {/k?ȡ5F84]vQN0󸘯/VY ~PVٵGNdzmnB۫ߺgd009L|[mF3.HX{5g 1f*[17W/|p8yA6JFos%ɜ{5ݏ mxB|VPs42$m "lM>J2j[Է_؊i؇Z-i[X: u}2WRFBYmT|mM=9B4Ze絥dQMҴ2U{nRL$}"$T(#k-b eq18kոyɒF~_aS4JcFzJK-mҺY` d̒N1A0\f J$뒲BAӧ2޽\ǧ#T1U#@dit2Q0E!'aZsy-'BΌi  Y Ī֜t;⒘X*uDǢ%OKt#CuHJt;hJ6iʋ[Z kR!TT2B >qDëV*"91[ӯ/B,=A{.FLlGH+[C'Dɐ>% losL.kyK'JdUKuH9QJ8ʌU/r5yio ~xA `Ke,/p>į#t<_GLHeǤ `F`VĈjH)B"P+59K}2IeSv9LŸzҼȚY{E/:$%EVXˁ%Q#b& y@yPXLY)dUH!-cI쒓T":˄whZ=&7d!RN),"PT(@t[ 8gsh[R#oݤfD)R[ʃln< Ƣ%Wp ߪBiLX8Y' qAo( b#U3rbOe]8b'u0 #|(*[` bW S!Za򔭋` b\XKY i3XEfҡJ  ̶ 7eϠPg2BQ"!\͑u0 MY;J@+hl;B6 9GCZRfP}Y8ؚPԄ yDS!_JA(D40P(6 jel $zj a/![k?R6h8֙py[Jb],B[140_c!)'A; B LDK0ӎ*lL0庝Qu(=Tm3ئOC%P4#ҧMPdn@q87Vp1[$;ƅ5&:.%ЃSF<@H W1 iIQ)0¸0GƁS@^|D$H.:/ejnGe@܊`m]Kw V&ESY_~^[QH9m^V`$aª`|ݶ',ڏ#Or8X9XR זXk#p`WHc֣.)1H/&9 HTq]_.$_1#TM ^HAʀv桔4KpFKIsa @AK^πB vDŽ|0bF f,Ւ0Zj"΃#H[E[cn\Xdә4HHRh r `3ZF 7Qd%CIaIVc9"agDH! c5Zsf6ʶB:f0M `fՀ RTi޲({)E8]G 7)-Y !) fM{-[ D{0drɦ6mMºvܮ^yM/Wקak><]esmc&Ѵ[Y#W@7SJD3 N-&7 FݵQp2Hu5jZRU Q"eѠƤ@Wb|6L@ŒP YD,V@mr )#1pyВ]vn@)8w׌LԃX&R␐ B<^ra XߴYaӢe>c6 "($+)CUxrQpŎ9{?"w +ϘV0 .)&95ZBjJ xT##: 0YY:@)ErO@s6")΄y x193Ekgfxn4H5i*E\<{ٶ']&0-L2k듵O:E'5kDv \-o*b{f1CYBU|`mqa :X*:L:Z8`A+jυ4S"p#&D9a8AN=(9,3;kĪTiX.-{N옄C@F%f|$!8@I- Y1)GtEB@({R p0@5[6˺ފV"dtsρ$ړVם`_Җ:'yA̝Sp2K!n3V]u{WnוOa{uUiJciu߶::iitSxua{l3@`:nm;n=njFV=t+]KVZ^Kkz-Q%DZ^Kkz-Q%DZ^Kkz-Q%DZ^Kkz-Q%DZ^Kkz-Q%DZ^Kkz-s% w?2l@\'Lzze:2P/ `fCEbq-/q_4X0"f{~INr nrܙzTFNY[|@/DbJ}$#cR11J*{0'ݖ譲߆~;h~Ëۿ bSHQP葃ڇm1K~2L >נK/z6v?lʵz$7Mrz$7Mrz$7Mrz$7Mrz$7Mrz$7Mrz$7Mrz$7Mrz$7Mrz$7Mrz$7Mrz$y6<&9 4Mr2K&ϽINlo6Q@WfvɑcvaW&mѳX.=ode$ ߠRGJVZ!t_QV08GEV_ٛ4eUڥ#ߞ>:Y܀4qe9EinC}Vv~Q^juޒW`wҍw9zѣ_o]ϱn)u]}6ܽs'ktS5ݹ;t\ӝksMw5ݹ;t\ӝksMw5ݹ;t\ӝksMw5ݹ;t\ӝksMw5ݹ;t\ӝksMw5ݹ;t\ӝk'Z(e#tKqjq505L\#tbk,HuFRu{(cLp*Nь\}x(AeCAq:euK<=ʊO}%m[Td9&ӕs2Mj#KYF_1-i^T&\߅fM~W,ͷk9~[<ݽvykg| `U.ggէ%)_{<T>bY~M.rfrGi\,`۳ tq_VwO2x` yl$ah"l7x_ZgV]z*trg<#Y0``ܬէw,# !0eY>6L-Gk0;ɔ^zLjn_~@L!{}$[(J!YMI˝;',vb@RRIuRak CRϴB»ak>_y9i@_`0ý<.g_"vfn~r1#~ a|@[{qn1DU}Lj"ǹ㕍vxg_WaCdזr;Zn?nxj@U08!6gpV8e :([C+ OG/iL<# yB/ yB/ yB/ yB/ yB/ yB/ yB/ yB/ yB/ yB/ yB/ yD/{żwy_)ռvMZVuRͦ|C}E?Ҹx`.>θx\?̸xZ5qq?[PISq3+W&g7*#XL*z8z#aT,2"xRNMوyęʠqu,Ex|-֫t8I`q=GhGU͛wʑ7_S!wWfm5[V 0>q_8ͦ4]\ܬ.>&'=ki8`y5o~]|^<J߃8M _I+Op"eJW":yd죍:+I3~AS K uǥJ4HG:G@n;+fN=OE\g]Ya]4E=a&iܞ^q)4~7oKޮs,Xg +S4)4ONçƇx Ӄv'x9v+=uM#W&~}^m.}՛Cz1}]tvUS`v{%_e"ykTZj¦x$Xc#] j9ī; |_ߟqpV_ X8K+Ca4YҲǟcD_Dr`$o* Ee3":M 3BI.MLm1+Łw-||)qV)m'\pKP"'.#Je18$ǁw$0l0o," NK⤐$EQ{jb4HM`<pT2JNbQŔHj@d!V99ƻ.fi8Uց t'%bXq$JD*v@,j*xGj;pr]/]ң ML.E'kIT6{@4p(6{gc݇1*HX$W!JEf&9A/fm8nc 4Vޏ`-\lQ,jޑ5ڥ)3zڥ:ĊK"tziSN/ӕ)(%/ɌX$3 #.G~H!( Ɓ'֫)Ea;=(Fw6gtuTZO( "dATߘro#3fu8dm  0?ʦ6(2륢pj"A\xG{O*$xB˄%>HK,(JSf^X;̢&QQ :f毈AU *Zx1RY,챿 p4 ($z8=QQ- G)"sXHiofu,- Ôp1^XidGt q݈ 1^4ky|W  `UStN3Y=v$hޑR)܄UQ1MIjJD!5RQ`Kݕ%8p,mPޑ,.s=g+N5B[ ++մ`bDhF8Q1ӉpbZ1ZAW܁JZ" 6Cx"icGr qIsnn.hx̋Ա%:b+ 2EOL cی_"/PiXݤ1 #g=7W:CĀ3V8'R+ҲnѼ;/ܤ7^KNpO8p52xj'x4gg L> QXJt-O?2c ؤ .|Y/IHLX5e0E1ł=oW1\O d |AQB2HH@G3ІHr[ h[-\syԜ\Xڣ8!9˚xx{6.:M5czxDğ~nLB9@0P忟qYt>eyKq*nX>Gct^ʟrScY?ddņ)&$roYl{fۘ57cU1W42h( =Pm:}tI\<,î1m1Ԗe{Xv|Uix=ߛ&/Ew3Zp<\}(vHYU`uEyYU[Wn@%y:Mv%Wl1[pWp=Vx}345x'rEd+$4plg+l w&8q^*dq'Àw%9-w&׎`@E1^p=)}%G)g>*vR8-Vug2Q{H 9+xiP,LF  ^a#4ފRO!=$Ghޕ!Jhyib-bY aNnɊ&A>H,*N;`58ˇfi0^;J`'&1h5cY߼_Li8-8\/dZ -x)qX ,mJEwoJ,$EicºzWvo$eЅs%ʄ@DPU*YPg`"p4 & +pzZ'+NցK5Q!N_t_ƭ#g{"7{;͡]6Yw1]I+ |oa?=&1| d&5%8\! k] DYڴ(ùvx;[SÝ÷+&t>zh+fZk+-rDiDw:7\O #SOl*cV:y'Ppl>[#YvRXqС:[j_8]RXJZ6!ܶ׹Nx-r6n܂$1<  ݞM^buರLTwN|Te4kwEGxWR=q_ ~b6:L uT Gb cÜ$*ZJrrvG[mq]i=y:R ȧShPV,7Ha*]D/8a5kP]IFJ+r0U-{gJOyxrWsLZuX6VjuFMmc@h6PI.> >A\.3 8"Vz/]LePz>ьu(o3ҬOO8IQ /5P .`Rg =ru]"r_j>k5jATxW8mE} l"@5`/bfJ +upJ ŮvbBn]Is#r+/|n>pV-TlGEQJb%Psf3*2Dny/8OvoSa4x"bPӊu,< j!r+VbXOªGx Nw^;"7_uA+TCJ#[LN"+zG&<9nzVN4v+'LJ0P:ᵘ2D_Ok ؈6|Ld+67}^@RV(Xu'+?Seuv:Cep{ VuD}I$LJYʆ4tH"JX[:Vl04 &*[MAiC`GǶ[=>T䴩>3 {'zͽڹ7ܙHrr7+w!2rcl@=f~qiZl>d<<#hL8,8,"g F*t {7P1KTmͶf5'/.h t`ӮzZ% N#k=]JH8eG6Q =>T貝H jakU2WĒ6ˠuB"l(7[t)cuk]} z|و)p;u. o^W%8e +TH<<=m y|ȹhϋ5uz 1.IzMER* D'[eT}=>TԪfHZp447fȘ| ^f|{ň&I=2Җ/Jw 48:τϛn|uZq=RfgEKE$$[y]yE.+l˞o7~6 #tВWb{d,{([&WfJ#o'DBȟtfC4v!LJɖ Y! ,mN7UhٿG~~&xuB'*hvnqzZ88ch|zᱽ@++OmᶖW&>&0>1S;Pa6~8 zO'~9ܽTU&k4+g M8(CFY.4 O]ZMKy|5o]x豊47ZYupKEAT'7 Mj|?֗<p30]r#t1/{S]|Ϫe5n*§ ,M~YhYP],bBӉ롿JRbq"r)4QlRTH)[ A6e|f :o۷O}|ʦ@@Hb 񚍿GA;"&!&g4Β'#Փ(4Npƫ5N_˵;:x4ܸ9֣lz1шsT ׾ʗkA/E齼(EOf;={}Y(;':x4BnjzTFkojnFh0!1nr-zR6 Q,GpS(9rX:0䫻 CkLSx S# "sEyh9@£(a(#>L" {a1>hgCE^qEE7cZYfwO}% -4NtuhUk:8btCCPL;x4_kv#]of6[f1QJ}<Ϯ'Ȧ{b6wD {|kB_8%]S2ΞQ]LBHt@&^ nz_ ]@joT;1WayHHDq}A=AsŢ%IPTRBPSqN3TMCzױ(AaX٬o:z]V+@LmG.jPtD2cU˾miX۠kVU`nJЊ/5$N憀K hJ68/h Gxy,Vv?_웘'W5\N\W{-G5_@ .rPGu!XLlX~zwC@whx#lQ=D.#kjDi50)!DVg//;c m;n;"{!Ҳ#w!m_$e#{ 'd4|o;$Mbno0 E;>l>5[@ԜOxrׇ 9<nJ?J6E4GDm"=Z#Za2IG=}MZ١ ]p5 > `  AwhwAO`I/2[2b7y-(Leb>EY2K}PkcýRiL0 s6GrRG2L54Ե')]I!O\L_W}1OueُA)GUHT`RKe!X.6;w)+|^~AH[~V att4g0/t^<߯O[\'83f32#>ٔVmtdD^?5Gfm؉t[)e:Y*1yv|5nm'hpT8O٘k]RHp[r9 ier/G!^%e<n9yVL~:Ĉ{iHrA9PAGJ^N חȃ[kvɣ˜Oc<9CD< ,_˽쐳"E7yZYW 6Og~ 9f5hY.~/p ёaf sEzn6R^w&J;sDW |Oy^LVArh!eIP^z3b'ъRtDP!_Nݦ->ɰs971؂9sKt\Tr>=Г9hTZ?=E[6&sy#lo$ՈfOX%S`}_@fzG GE<9(}}ɡ.0cx\{G e80Kc^f= @h?Օu.[fFJ1,ߗ?k D\ KP$Wm9d-&++CⷢKaqG[ (IK[fzdaS$5 o9^e}]NxSU:x\3w2W]9$~+vErYɵ~~>㏁}ғVB@68״싡]5Êɚg LI`s+ ¼Ϩyr3gX-$(;瞵Qmt~w-G&Z+EH,Ɏ%un Cdru{3CA%ΰ]#a9CƬl{e\~wKvu#wfCyVo>b).7}N|؅Q+r7@D Խsb1Yl[z\"!$e GtwZ~agFX_Ne0k>f@ԙ]\h<j= MkRzܱKr0 HV"*;x`Rm;&mid@c(k! pCk>bl3*QMۜCor&zt^qIE/6/rJjd A<^Qyy?cG/ÄsvyG',cƌf+Gu@ KV{w[Q ܯҫ_T:x)ۨr5QY8tcb d{GVp"cr;+VwbXWB3CHd"H`A:VKDt+8=R` gY#| J)yEL&vȫ|Ů!(9udYxV*q,^UZ~iBg/-Z%&kp8`<27qʤWi_ _WpKyDr9xX$8D䈋!h* ²S=#Bu /3w;TE2Ei(I I?K]]dP diL1aqZYguT.Ox'^{l,G!xV2pbI ‘ÍF`W=f[! E* i$r$e4N,+:xT6_us#do`<|<YuSLk٤Lt2,{0ʹt`+Mk zKT :͢hpH tBg` J!Z\StiF8=F,>y%؃8Zd./ a]`bGCG+KC\5G/j{\Lc 痤\l`,4f5"H u,)Fۿ<]Yj|g䑍S{޸q$`,pwŧ sXM>$7$vA%$weHYdUUŪ`Pߍ CN݇LQ?|Źzv< C}sn+n}>W <<r:FנI? eQy2I*B/wBNq^H3'f2ؾy4fƢ}n9]Nyo)VemҠ6, UJ3f2JTTF"W vE Q`': ~< 2o_XJZ[LhBFLsq٩U/yFFWCV3Ã)R~ /V_[[ > EșŧYWg\ s<o֟i,ٍi|7lxڂ.ey'uEs{5(!?vxˉdI2i %-uTP*$ 90> !U g&Mhi0p\>se$v%Xo+v`7zDSSBZ{+b7]۠d?tؖCm_թ6gBr UR:=O`9:c yuy0OVL;Q:&8 ly]/3H2z /$lC)^, 7`e1Hi0:K.`0) -yÌ+kQգ8'X߇nYVc7_O:z?a`o4_j ;˽)0%`%_[_;U\шy.yoB[wN-w{y⑑'A`G_EhCuGl0bl8 D#:9bO#2ߤ bUѪLk ,(#Ie"4F*aDl! N;D`&ݙ7g˔L}SR~qYnLYР6SoVxl 7F{xh1 m]]y#GWiϲ8K3wS6rU:"oP0HECXHY6EW,~l`y(߬gaWvZ莄zN1BWj(O^'1u2 +`-tI_pCb o;bpɋ!rOuB\yH ex `1F"B ZœȠDP1uIyh%5WM@?Э#$%,1$~e>hPwj6C?/!-obN4P<1!(/}F5S:#"L&Eoj^&ze#~ 0nS-un8Ƕ<7 ]6ՈtB+ЕTy^5+wT_G$Z04 s}qǫOg< )n B;䶣`۔$pX0eO=)ѠE6[Ygi2R*yib%eo.(yJ (_mC]hVqp5Ā :iY:u%S&cAO5iƘެâi:7Fi8^%?qpW*Pճy.j#;}W8pW6J c,}Jٔ;uJ6 Ckv@U(9QUL/F,D/cnzf WL,a_n WK?9+N]ݦu$a*c2|Yyv4N"cf 3JeYLb?J@mUq"HH J& 2FGB6h`3;QTz܆[|O srא"tux-xQ_C%~0P>kLCe^SᬎUOuDlPP4K93%XGw`]߬ QȸCx40̷Q1# P^o+ߠ`K1F2 F'ZB#f6(%.mWE7a`,E[4CQ$O^+ ⼐JQ.ϊK&=2F &hy9@V;ʰB Z׵{:1*qYѮI(g 'dRCy8m'#pC)㺢)-89QYh@(뜽M--Oq2]̊>9 |ls܎1(JŎlٲ+SEk|{piBժ'|hm~ͷݵ j@9fFx뫾ҳٴw{s_]ݛEN`].W,jvc~?ͪ(eѹ6o)LkPVDu](/Y?ŸN|;{ݜ[ w{(Mz}HtT- :h- xǕ2qOfyGf}_/+?Უ[oncZN~ qjmG(~*ŝ&&> ?0mC )vL 2Uv2<%`(J02wZQHagYsal9ʪx$KLf#M0-tB똈dƥHo2}&m܏RZf* dlG'5zckWY 0vck8¸uM ,H&7;h"0#_d#o1Y\dʇ %l٠q*ڡ_>j!uW +O7*eꉻ'>ra]_r~t[+ίݎ1IEC3K,TZD _F*HL&u6;WCE1EOzqECh?Tq:N)AL#M$Iw+_(hQd>qD5վ8:>etghطfmeoWgC 0:ԲA(*[;Qp*S;{5J2i|yýޏ"2riHx3kIFFA~O$zĀoHbP3+RnEU8n#b.g2@p%68\6v@/Tnvdf~dٲ-H=^d]fQU"YUqœPXe'V)*3>g/R>U"c9 b!GqvЋkX^b2e{B4FQI(iKo-e&)VI]6<1V9>!$!}%PO警?^E擠U-[; ŬxlL^GߪyGћIM/R_KP23<%6A1ą:zi&Oۭ[g3adQT"M3Rny: dM x~݇WsW[M4ct}s83lm!8$Eh逡mo*oֈ :O ̿~JpNGf%ʬ3G}PxF; yǃ꿛7\xHy䌧~/9T1ӳ@ kP9}v*jlFKPcbVVv|:k~>^4$X"B@8~|͗v]IcJFcM΅BYssksT|aq^,%udkE`rHńT (FrrF)%zp_#+@ -EZ,zUt cTڙta|1kqF4FA1p &򀺛onN%[SFl?>V,Bgkp/_&Jxe]RQb]|ZUQG/jF?ٗ 8)iD3y܁NfRK£qU chi9r</KrOܹL>HZE6*L-Fi2R #"JRsVPJ|_Pރ}{ISb7`_[JźEcn4"G,@&q'=s2Wm7dJʂY !_a Sg昡SiQ m$df E-ྟWxrE]6q21z,у4: "irE .;ޯ?vm-1̜[ )b4s6Rb{me>Kv \: \`FgC|~¢3Ա6V¹:ĸ0u]gI¹%ynTAXH#RR*φ:k%~Ѐ~ݞ>4iq]fgZnMçǫd0G]'^gOrL'"]}ѩ-fك]١dn'zM?+u5vV(E,YZuJ[cQ"L4O<(Nhz+U1P,5qJ@g))^iJ١kpx9ءwjF5 lٽʾz&|-qBSX_f\Γ0rzt f]υݜ /T?FlK|r {Zy-Zg/8?&nSE?q ƺ%%\̜\C&'"T&" #NUB6MuL΁N!н :p~ GmSA-hǻW/lŷS_.$0^<2wx˦8~l,pN8ͲqK˙I۔\F" =tMWkCb/<#۲Mʪv][=taڪŌ*!ng4yRs}\}Tͨ՞bP+vT~aEUV1}O+,~#GӍєȽ/z?{Vx TG郚%,]:gljZGxUðPs"ֽ؃MhIP F3'JU}nFCכimb%j;ؐG˶GpoEt~:y-h4}wCݚP|M__~74aFk:xVx3ٹi6#d^m ~ !ð,g;P-q/y΃NIL/j㼱LcABf(0jk(ƍz"ѻ*yg{jkvu^0L]PθQwv,%sfzcu$!_?O2ݐ9dV%cƿL*wݑYa/Ʊ ZTۭY\ =lLdjѣ]YzJ g Agެ!82+2kKWV;iw'ڦUJJ_.tn:8(C! ṕߦ]k <<7qdX=~_0@W犻3wU!W:fW>[ Lu x^ NXsf'<ס-ret t50"nxXqW^(Ns;]| Q{훾c;{:ZkYu5hJk6;Xn{:T6o|Oc[{{,?RsRT'?y…VW}VņybC,m©]ïRhI`OiMf6^qk .re_ yuF=BDҹ8" q%8!`Y`W,Gn=+Dy^x+4yWR?4P$)%ɋ$ĸM#-ijyk=j`)K oV)Ŝ:AJ}ʐ8lzeשU;c_iԭZPk>ˉ.i#6׬ ^WAWAlh< qKJ&)mE[7(+|4k"z ^-J3 RjR ,s"ҊhIȘfEov1x-F(`V/VL ?nAec;cUbXE~~8'@ax| mgmW$JNrOï"vHZ9L_왧Wh9)dlO=T.6;/"\VƼF p]z'&y.U'6x<#Vr[u)Z@,LE{vG'wvھZvd9ͽ^ -BۤN`c55 yI*hM 㢗q%HG1 W!Yjk9}&r&V>%![bqv+|<.v5F9¡L[!+A|7( V튝P$Z#αہNP*ZX*v{7zT>HgP&CB/f4fz:vC.Wߍ_͊h[twl9:&ٹ/x$ʞZ7^&er-32oG t*ߕjR ?nT$·ߍ~9K#0(a(Cp33v?=~gaXѿcum2?.n|>jBֺͥJL^ij%%c{;k?ir fz7(v΍[s![zH72ǘSy<9] sW%/z] _Οq;S4D'9Akgi`Kp[Ffo2/nݹ2)|0e>-~[mv%Ht OEbΨ@< dqSdl{Jʦnޕuk2MLm幭Ѿ iApG#ɒf=hVI6)W(p,aޏa%5{M Qbپ(P|Pl/_"8S좖1؀]b/#p3h@QKt9( #Ʌ9e)ӽZ<e{{ 1 '~ngl΁wIr\tv ߼W{".beI*F8#dDLJ9ǜb)DJ)vƮh_5I$A;eO~N15 !Phצ:yhpNf Aie!PW g`X4-,ZN/⤢GP"pIBbpݬ)s-!FוEv9M ȕ9`wو$`2 `Yt\Oezbbqe5akƄ;"k;HJ.Spq^`f&&.@$0>$DF O2*tֽu$:p`s`Wpt{ӂYɡM@%.^"%>*U15PlڃxQq0ִQNNhgx"Z^ ᇻh].CNzjg hbN<PHdAsTH Pc%]$Vd@W>XpwhKI󨥗8 20D0yDq`u@O-EWht0o C@2aNg:kTv@t:&-M)ñ1C(2Iy2W%vQ8[Ln~R4:s"rMX1Mji1ш<)r S ,G8^ƾ"ucP-?ҔIC и1`AbX Y+צTgtU! }vIPNP\L`e 6v c䗓JK`LHhǽJx$2BFogqrO,d ^/=c\4;Y\?J_@@zX0`悢*>IeVt弦+ܟ R1@ި΄K6vVEOO}I䐍TI9_c D'b 2l*Qf`dMa#䬵xGw϶-WFO?GM!yRMɇhub/]^ Bc'qWQL.\IB #Evxog{P=x9GoΤ"` L!@}46 Q {s}56Ѳ>F'a t(Ҭg㫳dg{QB`/>G-&J-JԔWa/xgA9C#Jm4^ˇh5.;HqIvKbLFf|! #7zr.o7,(制 Ӄ'_` 2iKtZXg<> R wVgGSyfgS8w8c9fTv8,mk/d<.O{WhMξڋhyB3@6ӫ|u8D7\k̗8]yg4}l_'<{=_ͿN ޼OS?o_߽~r DV~~Ïݛ*o~K?|zn2߯ӟ_܏q8]hzyB7b` d .`"0Q/?fK9u, ȴ4+.DXP/{_ْwjan/q V+[QI8q];c \ɉcjwvvv; *R]ָb{dRpZ8WmvmazWT_< ]h8i f giV]?REuթ冮ig3.mZ滎i2[k)u^M=ǫo %l 8 gN-QE1Ki f験mPO5U53u=KRmWu3̲mۨ9ulEMK+o62{2RvsqUŪniRtL_ nm :^cM iQچ™T`aTj3P msU];촚kahof}f{@в]؆ 2#?}jԔ=[ s:Wt_3P:Cb4հiP8,L"bS]L/n}?Go4̊|wqltxsg,] 7+uB7t@eˁ6];@o vU4 ,U3rvZl/^4noWotAcT|඲{HU ;u-Y@rx3=aVzmcݠ96ց07?a[wVho0t|^̖uA6Tdێ!k T3l˶|tKǢ௃_hJfP9yɾa|7u`LfS-Ӓ5vu|pq],۰KlE(le@5Pt 6b{Vc}J$n Zjel:u:A EdZ4~ε}6 G~^s7{ቶڹ%lO(^aTX0r~ioЮp ;~?x܃+7?u g|ė,Hfߵ8ėx9OjR]a"=΂3:5Y]`;P9g AjkwH3/#U.kl0rw&'a`T /jI4[킋h7qIMm|G5M`7 3a<ç 4ڎ&,I!ޞϙX7 z/#j q7V 2x)6 ?j>gl~ H G:f#gHfp2bZo*9N tXI vg '~>f;_l_\:>]Na+<;#,ͺQ{z롣o2GNw9睛.6pf3Ot#3_ndݗjXuTSm[݈ӄO!MyX(*&'Q  ctJI%[$$WIJO1{EDfb d)o:/Ne)e{> u1;H҇#ei&$]P*W[r{$N>166#.rґi:TE0EXOH}[zi8$It TɅZCzf #dfL8:#g^`'I2k9'ngN#B; mL[= F>kW^ܚdSJU+#'J:SEd%"F'6P0-N3̀OilA<qWeyDNXƯwrJE=%U HКCqv;a"?N":׎cv*N aEx2xg^ޠIX7vN%-SnվVX[  TwTY [YHH֦1rn.n,o{.f/l T^:oZUeiK(r2zgFB7zvgAwW<權Z5ϠxyF\K*/fto\<chsG7s1n["f,Q7 k&9>Q죎u:T#ftbuJq7Q8E&H|S<L,ttwv%opܑqŕlٴD&J%SU ~} nIͲΣ\}=zzohZ'9i;bq$qBG6amffb<=h<拦)BB}szY\˻g>d4ub.:誼ԉVʐ'p.Ⱦ01Μ sU&;0kqox(9U)-Y8p/OQƓw{+nZP[P'(I `g''6eM4þb6 d;b+J V%&QN hN2l}XLU!&5ƠEG-_$} zAu0NewRW>tM&u lb.^>ji.R8 U46J]9MJ|<`*&CkU-8²FKفj'e= @ Z/12 *+=x `ٚa[ʢr@e͕/*KKɄv;;44k_T!;]]<?e%ٚnyilv#8/K6rb%N6KeBt w}nE|nkUK}cKjO?-zӦ=ʺn?O!J!J!J!J8*:qxgNLNA p,u'v6[qLY5!ap+@t[gDr]9%ջw*» ࢋqa,/@%a.^G,wB]<L@s)ݿѲJ{nci":C[ey!~ki_}%&M͍̙vz뷤aT蟔!} v cfñڶ+M-Mf(\ ,M7?k `+9^ԋ"S Vފ3[qx+`VEn(l*8 5Uw\SfԱC = _7|CX9@rx^8j?Ͽ~-$Ь. l`̑Ǒ ݘHcL㘜b"4@2 ota{w=Ŏp^^ JwHFxU DK~+zUGQ D&ϟcO`$\4LLMyo*U>bU|ݺ N^D 8*tdEJc0C8xZ;:cq1X~GN<2sгRE1h@P a h`RpPA< nqȰYqQƁ(]UJn'IH=H33UR汌bO1[|P\@:w =HR #n?h]m{`DkJN^w./W]}kNr{Urs۳7TOMTfխce}7˩13_¤kMoo4X0ofa)d57'ǯS;Hlo0ݻC+׽ /:VWd]c6n[@rӗKn^}ӧ\wc)jre_l A`>ENČ !X2LA=PyOfJ7ܐKTjuZzoPj71UFіv+GE\Gۘ#.=8"2 b 1F$wH@=D>v\5$u&$ ;t]KttU&'y1_?MU&^}[u/(n8 SӦR5{ woߪ)_jw{'%W}wj/_ksԨ?/חMgP賷~OV-uqI:Wle"o5:McR3K0,}7%@8UNHRT9!UNHRT6o8x= ҽt +H ҽt +H ҽt +H ҽt +H ҽt +H ҽt +H ҽt ݫ ҽt u|SAo}(uaWؑvdYaGVؑvdT>P=NejT'KUˈtW5|A}J/k/N\ȵ,^n/I(7 ctkarUg㧽ֈ-XR]Kh0&{Gna일oNvwgK%8/^XE?6ys<|Qcy݊cam1G'ͲG&p|*}:$dڇq jƫf嗚$*9HG1=`l4|=Z+kDՖ6w0y/ӷ_4°v\gٴ_}}}ϯZ 73gKn:za}KCO2dGᣡ]!^lyY~(f͟+Ag2?ELٰ%R%!xLȉ(R.gQ$Wp+g(SF} 1Cr?ђvd$tHZ!74VÂ|gvލ\IVLZ:N;HDW}|e}'r=t@s蓱йttïrO$\f̥87Pi,狷QxkxeR:҄t9Pޗ)S:Z{Hv^YZֽi Aouߣ=ZuѰPXfclFA泋Xwo׶tJxӆ^%:$/^2eZfU)N60xr xѓH븡 ؑ'bE#G#AFG"^ qxR\ `+b8`)Gs;v;.Q ^(#Ny Hb?bPh=)}WYD'*k |֯Z.ӵ@˯+Pf(1?"f3YR0ٚ柠o5{kѭ-]@Ρi˶/@`W; w ]#UWCL$l3=_>99:ֳ4mo'f>fh-pfۉڤkfM1FXSE0햵 à-rCq0wdJ@2s{O\`"'^ 8b -uX$$ƞYbҵ'xowv\YAmJ˜h%h ~@`YhB{aS٩MV3􍦡i{ wvoh~Uqlm\9<`8{6Ѷ˹ ("!?BFW{ K[v]6ml}kw˩B>xؘlFYEcr. JPHI|̐a1e0k%t J)ҵT*>.$JI |CAye+#yaD**b4ri d8iInvGهGGg[N d7QoabYT0Il'_gk{"{)ꓷH볣Yoѿmbe\ݷ~~ݧ;''N)lw,R]ϰJN'U-nm-3g4z׏Z[[ h5 ̡Zmh'%+"3dNljk=ٴ;4 Ҩh+ #pewA֠1}C`t;Jl]`S2m]C˵β)ҩUwγZMKPXݻivb2A j4 ]lO'Og zJܚ iڣ<ǒk Fd <&DܫU@K>׹9 Lh`#0@*ח 2UY"mXC3&s9W)]WɣRsv]r7R8YC]c#Bd A*4IԅZ|*ޟ?8i|xvA{zi}mF}S^NԽw'>y)XN0EjZi|9_moN$}bIR$?I29?T`g('[5*|hq鈐  6CLDK0en}is2UQ]Փ)HmLH )WDdŮOXx!H[Br\ xv Tx< IX*(}D.eEH 0yO C1|qDd( ,AcGUEI:zgŖݳJujI˛ cÃZyny{B."QDCաNӎທ錜(Z(fTFEЦM4{'nU߾U߷vwR`|5¯q}yV?GE~F>lD95eK Szϲoz'.D)Z͚_/5SP1^c#)ȸ}Mz,2t!#~К+amlŬvaȁ'*gC޵Hk%wҩllcosCM}qyhbGIYGRC")D&4׾hub+9ǩCl$`i_%Yb7#+D>eDLGa%xa# E\q¸O1cL>WqMuCVVLOn[f%`gRgG .VPWW)l,8pIEִR+9(]*Q+YklTb{J{54mvY_ww{R,e[;|Ul&\וa&1>'6J{DY :\L98O22m04NK@?{F俊ݸػpCףA`&[O`%, {</Bz!1tJՙYάȥNaDx*{k3 k`4djsNB2"6朳ؽ*v䧢Au6Xu&Tqk{? ѡ-E;~aza"v'/'?Эm `3"̣S %bB{W* Fm>^@L'^*M5I5YQ;[|&E(ύ D-  2M蚑\?uH썇HFԿaiIn*F.VK=|OrrOˣ;Ic:fѩG_\:glmu@/A#:)v'eOe`8M (2CQʒHϲo0xڪ_p"ֳ2[#h*~uwq1MSn.y 5L^h5LG/$Ԅl^dUϏ8W'R x/4J4yr*Rka.-CM"b4͸QyLR:65/吣|dMh=ńy`kV[h%N%ӈazi\aѰP: 2jD)tf50^IEk^}Ʋ`ԛY|c4nFɿq"%[V(ln֍3V4-BqyĶ6du?N7+ vzC59y)mΩg噍ǞzO?tgOs~̢n'G?w=psdld}2@Wk~lW=:#;:#;:73G0ͩ2^1RDLJ2.E y~3ZJ ƕ%P2:삿Gϧ?8o]T=#Z(_.^GKٛ7FyҺZkX6ÖB 7*  ϵ<+ݗ(ŹP2\)x346|YԄLJ/=>trvZ]$vˏg˓U5)-C ,$T>A61#JA}hPji/td>lJ8<\@X(2T>?aHJtS0ᄛ(4c Rjx-`)l|*G'?䴎Uf{kiHreBXl\gDƭ[q 1P6f*?|S0{o(S5&8<a0 Ζ{`THtT2oO S~Sx|~xqv}8eNwF1~ycQW}A">ӻZ j7NOE}'5ea!~۵{ }Ojp,23P&VکvvjS_ZNg.,ݖP6[M6nm}Y[Y; O, Dv=Vss׫Mw7{fw S+v]M+737/Z`tY(7)7&= ȨOg -3׳9XnB Z0b4eU ݤ):ykdEw13 e%t\=`EXP<8LVrHy$Q2ӌkj˼|Z$|Rf(N9}Ȋn*'DݟcOENWT&\͘3r1u?uYu]ҙ9_#DN%%wZ٩ d? WeI1ƼR&棌k;8}4:rcFN <_uwvas|g3xϘt>b89x3Xzji!hU<ۓfƃz2 -h 7c@VT3Eni:5* LVbәIh7/T:]urH +74,o7U^e<٠5\V(e.N?`MPY,k}Wd뇲Mͺeבh%N%ӈzi\3~NaXc!(b:T:x6ݵ =öUGy}QLpo:A }RcSŝ(2R͸pR +=~bZʊ>=F}ncׂ˖s+Q+o)uէhj\``_[& Wf%6uY׌rӄ$K> Zyk0I~31>8d}nW*fWBCM GY7jDY/ 1)cδ;6, (Yv;;ŐS^u?{cns}9\9m`ۏ!"#xv9jɊ 娤W]:( Uo4DN*=#7Uv [lMiܠ waifMYcamb7?_7H{;('2zxiq)A6; UmTE\qqtz(vՍ׀# 6>iw\gI UQk)M$# ET%D{-֎+/"q(Ůw.1e24pm U@錍q>Y)&)ĽS.J^b3p* 2S((1jiuxMzoa0Nd{D"vю}=b_# Db@$6 DbO@$>Cb?$*b?$Yw聋egcت$JWlU[V%Q]6*2e,6PJEIJbت$*_pUIIK k5^0U^E"Tuz^PE"TuJV:iOԤr{I7 *Xw8ty8 `jh>v,F1m2z9JX\<"Nֶ֣6Km֣1VQ`\1rX^F3^Z]<${ӠWtnM=pl/x7a-ʝ}vI=}>u+X PZqO:i0:~x~*wCvHEڈ`5< UFLP>G虮[՟j;r`qt+a^Y"9oasSu&1&s#cFBm/(˔f5s*XNzӽ8Nluu)=٦K]aDif,!roijFo$FdMYolzG_؟QG+0hM (TԶn>V/Rgd7=kcJŶcp@HL볇 ,ʘ+< %rhB+f{aHH$ _+,')pQ>}|Dn%>;&'2'( Bo~,sun>._;]Fevv u_Ȥ33H]s׋5nF4-h1< .صc])YfmOZrte][ :|bê⑍b1T]fWӎJҔZ5lBChI=nT~u^nj\מnDzlӤ]m? Yn:]̧s(ʧz {ftAm6t67sfp$N0àbl[3޼Gfjb 9=Fݜ8PuT F&kIDF_ehBg,@y6(*8lx+8`o#fN;~n5f|Dsx"z )cFׁSmFZoxnۭ33o~R_i52CdPæzP Մykd_}_=hveJ=40nD2@.VY=RJ$`{فiڞ+Z[ ;08Ea7]S4nVqsnPi2@*njACqIqq]pSc,Fȍe$)Tt=82L_pEW\1lBЧc>f[m0? ^rtfdƴRIS.:I5v) ,u{>HzqW/d'pZ L% ZTY2@Yc9\<4]Ƿ#Yz6S}ۻEmA,I֊T8U \RbK0:ޑё jA2dzJ9{ M{fqp4XNǦlJjvTۑ_ZPV)-u -nseuM-S^&c|,WE2C0i^ XFLB>uaE.hB-3'UpIqIJ)FK-sSCc1m PcLR.0c&5)^ҍ-3fI|-ؐFHKº? tSA^X>u}<񤠊ۅ%џ#Yfi#X| HЇDVx:/dpT}2 sϏhyqؘ|)6-K`YiDD-mb ̣UDIPɃH00Y_U 6KE6ξ`^-SOvv-ûce3*_{)\]7ZnghZT$7X]2Exr^{siY&axU2]T415sGub:f'*8:9NY=2:>Ӱ62Ei gƢ2Mz{ ϴLy&K3ؔapL}L9:3&)2=#'5lmG:۝H@IiJ$a T]#M$d'wW+l9ȵmiJ:4j*Mbܕ^5DuN-tKd;4N؁PCtX.Sõ \!fY1L!C]]j{yL\[8ڢbnIJgåMA}ބ. H_<^bAW[g8쨯*8qq7@~WZ"IHQ u4UoI[C=~U@=bٺ +[P|9l`r޿>wqp8')Eݯ,>$%R Jd䤁UQTԥGFKnvfҝ%S2)NĂCXD͝qW\Cw%eֶ+IEjwݕa1X;$3J똻⮀Z]g$ĩ tWԲ mw\Bv]IjJR]@w Wt&3F7^=e+ &Pnv[I)7^G uo5%>67lL 5L2y5XRl49q{.ܡsv,k2.l|Prͥ:t1|ͳuMF#*l=d+t=rMKε&ڢ6\~PsA,ȹ3yz46o~腝CMBXc Bnۋ?!voBul|4^^8wz>>xoVugMԉ[(JÒO{HCVV+[Ki0%R|\5Vq'v@0,4ouЗG{2w 72=W7/Zo?IջX(j4R?^ݯދwck]a~5vuU5r7vu1g<{JM4R=nFMMYU)v˰rh _]|A1+AVy[`~zp羯k@LA8<>FcJf~\dNH.b?g L G =G4QHb;22xd/N{]2D/GMv.0`vE~F|$iogoZO/Z:[k]y m ~nW~7>)22n ,x-&2vl>NߟIs<M4./܏`L^S?,tĽQ\>;oN;0 MY0D,~(D%K߆7R=/o{]C߿cDFgj J#c mؘhBSq;qmɗb$u^u$ 4~wŴfHeyWiD}O'q׽R&Hq[$q2j2Fk$?7i[%=N~ ;@J1&a .0һl##^{{uf+~Qgw-9/$UaNy:%ߗs3](EB\6nm|~訵v.R0 4.zS{ (sϢY|2oxMtE,AGZW^-UD@5QEVJ$GÝlc],6b NXոXL@(T SYfNJlk ω'\lLs8oQScDu24, m@Z'Pm:s?'>o~T/_:5gr?N? uP[ $-Q_Kyxta'*g,FËeI̦u)CƓ^ .(GJEUbʱ[i1b^E\I.lOt,B)W 9|ATaIMD "}5oYiy(WP dhC;Ֆch\&ЃhI{a&3*|u}>nbyGB00\JXf .>Ŷ&rc*9 ۸QAJ|yʾog>+bbF)ur>&*KkSOni8Ҿ~E館D #sAΠ1LF{*]EᮢB^sye>Jkb/le ktr6u -4oOQsEO%W-qt+ rK@:1HĿ( ˆh|mL|kݴ"g2| C b=;B*+o0 `q~y-?\kE((g;RQU] oG+aH}c;k#;oc}J|HCٱ=P8DkXf=5U_]^# ?q3*ߦ3fߌ FUkI;J38β~lთ?xC6B/&i=R˛^~^~ަrP *_ )\Gݛ\V+#rÍ!C _~|c~y_?*8)+[ &{K1u^m.g隷DTuZfFݱn[l:;^ r>7kwysaD,5+..GWswH߹Zwsm^4+? оu0|OzE@ >̙*=Ù-NIUmg.w[]47zrKdmbH(G,b@ݹ-waފ̍>Me\wO}Pp냂wԓG|5ƆPvvWw:UZa˛Пusr+x1G O}W *Z+W)yT"~xi`j喗akxeS,_HpϥI)= k JY5{eej}l]sE!>JJ@SJ5#8]U?f=--φ92|=nrPg\/S-e@Xl 1^0+:yN}#8y7 ^\sڢ<YyQ@Ǎ3ܘ X/ 5>}!`OCcy2 (칗60Ϳxinߊe,ju'ˌ6?X %/e-&EGZj-s`va]mjsv/HiNJ{[ U@ۓq6McPjl@بrս<d=l~a8e}485n]euJmjYϧ+]QRg`E%y'1SO',<:ÂgOa0~u62 Ԋ}J0y%tTK$' 8>> ـ 9K&ʲEqT {FK׫22;}s )WVD|aGIp_n_qO!^-A__._fte$(Hy] /PtO[Dݛ 2ւ CW^Θ$̼ӌr^* yM)37 ?N@dHH}Dd.jRA0QIjPzaNAaŬ$*-rz u(>db"a=Tja9[/:ot՞sv\ -E 1z%Y c}ly8KMg Gŏ*b$K8% T+RO&Wv'kIvAB1UK,h|@gPJ!HkQohWVnIAFصƒ>aA6~[7Ç5c:0;T)Lݩ wRn-bw8R%}qXIE~a}PJ?z:P73aS 0Ou Ģ#C$1zu3~~s7𷧸ssz$J&kk Z=B嵒XI׈u3,{7/yf-,Dٗ"wbVLs谆e_;Md`FQ-K.wM>ZIZ[{+7KTiGݸZuuZ^} (ӺjpWS30LC˴]E =_y\UwWgϼχr_ϲK:.=i8-Z:ϭ{ kV_5=N'$cd(WZ> _Uq]P+ZY^ieVmBڊb^Z%j%qO-UvZy1٬jbFOwjV"%lEbMJ!pS"9MB>V[FPP+tCzC9-*]Mu9:qCF67bӣ\EZ /Z_}o Ez 367KP{ʄ94[ɴV XXKN 03Tˡr",&<`-!Z fc!*yؗuP9v-OgM=.u%=+.,{[nX dM'UyԖ|N6gd8>>hy- =]VJ8cPw^!ak"P$!R[%P_"s&mzPvK j]IgqNK@uAXacs@rV@ 0x "h8qN.$Ļoc1ϞojkZG8f!aV\ZhKB DŽcAj|#p!1)qC Kƺ?S[\Q$`C[ZG6}"rsy7qa%IKU,EaΔq?FLQwRM>ȟͧ8MÀO$ECه"(P  h sCf_e;/4,n(^[028\n^Iv8m?t\ma4k(Ma徏[lT38"Axxb#ܧ@D[_>+&^+UK/+|lyN8‡7zm6]Joo^S^A<>Yng [,,OjP#N'gC߁DZN:ȡ``S)IHQ'Q7|A$N[φև ٭mu/Cp[^AkF< ==㡫NY:5^*k̤jB!T;PH,:H|J ZDc'=U\d?R# RNpSc !DI谖Xd{HV XJ= 4EQ{=D (gs?BZ +6{$ֈ8Αe9b:avGa4mPJWRJ@bYmIO`N21䄥9B+yNBc)BzdnCQ=~2yv;`oH ,_Qҟ 8)˥RƯ}TGlLO|6fETz=`qlvg嶧 v}MӦz\~~:H:[{ʦO%[ |3zQͽgZWeﴤs8̝lWkD ZN$2#xON^.C.KK ̇%˺:x`➅W5eb7vbI[IhK6 BDLݶ dnt xykǤ\ས[%uJ{r݅g4ߖ U(|5';B}̟Q̵;4\l6 p>|ElY7hBf%U ocRTl=嚁Ƽ-T۬$@'=-O_ ;X\%( * oG|@.jn=iFQ^žN@tweIᓉ'\YR9?v>Jc1F1sɭ`)$MSte~ϑ'ۺzh7XYX^K ndI͔gsqs:wck⿜u.cc"oϧm=[ ;;3]ϛb,=د+#siΙ"q%2'+-=Hi;SS*i#k0װ{yt"<31`-Xhm 4wF-XQ9b0XJBԀELRDMilஞ=YN[Z\ccFj#Ph}i"7 iሟtbqÜ9Q [cI"p,ӽd|t]Ͽ;7:5|1jLg*Tǟ"!7 "*s5$1-oq_*J|{pHX`B<UX-zg՘IDk1hFs+%3'|E]~4\W:BLr3n ϴD,<ߢ _UL[Um3zKuo|ZiVny[کUivtsĩ*a al"Rja\vzԼݮ#ﵞwMzz?tC]#0$5AZX,_vGEwo> KKB98J낏[ sIATt2 Gkbb\$+}Ԗx93V^n_t۟^{-[fVjpe8RTPE#&jr M'y׼92 ɘV!!: $H[XA`Ir,a#v2%^v)iOZv ^|'vhQr#iBCj,Uρ7cF! _!sNI'3y0T_y;:}}^f26(9N-P g3*nm:WxPW~Xe1/}ٜ5X/V[z>:gڲz;75W!;75H&yqWRi漓(GɜisCH̑^p46k]?+ַ.z]m9: 4#_:'LvO(sjqZ}gsM8!8&M ,r7Ͻ2XF.(nvDz-^A%ҡ0PJlB O1jve`FZ¾lkO(x:[3>_F=k'Wӽӯw,nrf2k}bBbMyzISZz4κlLõ2 > (B`!cPaqczfqj6n9$03nI69fg`ztFP Ȳ̰H;$$" (O %QLX20q*vA^XuwhVJK< kD rܤADCIeRkX =?#s}mMVZEcN0 Ǩh%j I5Ɯp cAj|/p!1|U2b9nφ!3K> }gb0HK0_!IM}k545JP^&Lo˥yR#ȟ/$G-]$G  G#L?GRP훢gvx_LA 퇡gͿJ.! ) Æ&UX9]ԳK e4McqF_ldIܔPe|YÃBr#@tNjC6ZZJlitgzrtJacuV5>!Y~.p1=?x5VLFg93 Fq88>hKmEjI_ь0zjI֑_=Y0}Zg >ƥ &*|<^6. k'QYdۨmk%+!42>O\>8WXT5 UgQ8s`ǿ_Gx~~:z㫣]/@ip)}{ԝ۝&&pm~C󮆆񚡩bUfN2uj6nilo-ҏۋ?~KG$'ҙծ V G>W` gUA5MYmyVY3_v@/<;~+-kmE4p51f Svas0؀їP]'QwYYT`#v>H)ʝEi)N~=mIfS+B3 FtG$`E)썺49/EcZ{=d&ky7T֓&nPkc#JrAUQ{0h5:9*ۺo1O,8FЖnz?O+(Gn%GQ%::_*Wf'WHuס0Wj':wp9;JثR?`#=uè-Yq>\p464*yra!Pք\`b6XIS{_d7x?_{b.j4xഔswʴߛJYѽ$闈W =eBԳA"iibv]ICZRRpWe~1x5˾u"_fx\%]߃42jvkf%%6]avm3jP霎KWdnOҚ @ϖ.lӢV8U.LzcK::-dXmm&V ub< ugAUwӽ*Rs308XjbePeNq:\TϩmfPRPۏ~J  stq3*/柙-\𻺶hV M?fR1_.5-f}eK;h'^X=wZDs9꓌W>Jm[me].#r[1m >9H"f|2߇Ujminl3R 7ꐣ[nk>N*> I iowGGwTRϲѬ5E`tE`RGMv &kw鐼O"ҫk"aҍC{V2[ t/!_S 5*Cz Y@SJx->cq{ۙ‚S?Z~r p}mȆcJ5S*F_"< ϲo'm҂e{͇۫her:zmk+cOi8*ٌMrαb-;7c5bL)3 @'_?PQPBB,[JL.0u-h{,=iw.=]J<=YP3xeKdmkKW5j|v߉ rٽzpڬ:XP$2ori̞Nx\0CJo?vQ!]Gӛp6_,Fc11Fs?p9QKOl`}vu^EQʖ?ԏ.}nM{5/=9;r퓴sP赲2ra^!X3*S$Ĩ6xΨjgWk'Ώvm]u0V[R-fF0LZ-6003~ nz?lOoy=,߮lnrSfyZ0Wͷ?4m?#~&j=o& Fay^vu"0.(b8:Ym*rrw8x W8Y s9ZnRe 9h3Q#5"ʭI! Oc;b$ݛU pC=L`~RP;R!ivYRvA3ԢB R KF3K#5Ͳ7)پmRPܼ[l/%>K׳T^f:r+%7[\3gJ*dNۦr{w7ڂ} %xoT׶~o,{7ѕ5n+PM-961` `}ljσX[QSRPͺDlWj j5*u5FAzetKjFu?\BW+قa K`Pd ,%=T_B~GDm" 8&(!^ HUެXcFgkwPB;XaxHQ#nidtnvW.=EPpfA\ԴoG~Kj+V{6B* PRj, RV36&s]|^B}MYV; Dx㉎F&("8J#(ATI;C[.oV Ơ=>z㧷$ME؝Dt|Ր>WIAw*:z3 nnrNq}\z{+q3}# QoF%8z\]C\z㶹ג멲JZxlw,3b0mv4e|s?-U6oo4y( H}e=Ge %GVNSЃgEZ3tCMI>`:O96+ y_٢?fʼIMg7?5Agw,c@\l`Po 1#dTj.apCj oldXuT s$ɔljOՅ Ppr~Hie!P%|73ʪ SUc9!m.j K (cC  "ޔ (-HKE(Y P^%|7*+* OX^愉bB0%G&cf€ '%j/QwIEA#rJjd A4P"ѵ] DeKsH0a%j/;NHX͈GVp"ciK(^G{Jp4 ౑6%>*pV (T^bJj<`PTib<_eTK(`RQ LOD8Q uvCE U761]"jG $ &/PCx*<#A;5 p3K` B 0 9 Q qL ʍs (6ϼ31"0o pk''C UWs&ᘡ)xi##D'L"!}LpI 5RWKɄ2F!-`zڄ4/PCxpŌy/]j4Յ R ˎ 9i(2t M0g0d(5XAlBuur!.3- r`Ѹ#/PAxMu9++pZANE09H NϗP"µc{@\eȁjHI4&C 5'JpZ'e04ټ ! 1);:B iPYY!@IA[8C.LRf!t%j8!3LQHR9-pb3_Bډ<-=琟j H1zr]P!v# !q&LE ;| *ƒO}RxAwl6uYQM׵/ D Ko*0z4$JQ!% A@K>//uX_̦WWqv)5ZUg]% 71 !#9`[0;ֺK(>߬Z]x1q6qK$2$ܺ0&PExU}sO9W52FQĬE0툇K(TR Q{ZUHSaAaV 5VSǑԆA< rxFȹB )V7T$!xF;DY+t%qV,cqm垛_ ŰqC˼Byyʍ(aT KS U9MxWէdYyٔQ+t59m;;l2iyx4?7W ))!G*.BeۉP# XzM(06_/Wm`>vrhӕ@v7sOW2\\]vjV&vqڧ֯0?K_7m0X6`7'i7DĤp !pM 1A)FRX VϤ|03,Jyh Sg0S&C蛝M xL6U jeȬ aWlXQQ- 6H;*]jJ˥ǮAx0LV50.$4}ʝ=Xn.wwwfD8U]wlA0#jLc`栅?^;x:<6wм$])leib}#'cƳ'hTz;ugB.)@VM!ċVx†޺S7Tp֙}}-GnCӎ}hX@ySQ|q8VZ86zvyx]N#W矼tannXе~rF74fE#nE:ZR$W6tPq~Y. 4Z'!0<7w_@\^Yx aAMKna}>oF'%j/]LkʷҦr7mĈ$KB %We!K"CoRmRp^uTP!\@"_pcnBqBa 󎮿%jTՅ)"CNY12ąĩ?T *kQQRPrk@F9%*o0a[$@£Y (]|BQDƎ~%o-|Xaq,O^:; K(h\$c4Izl< Tt9%jOYVakARChƃ2E Y 5gz$8snUՂ0 UWvapL!`n4"!ʵ[ĜSj tB 9$7#e%*[DdhJ(^PU;IJ%F*49b%H18ws O0orߋay;orߋmMqINɝ7:R-;O(̮} $pup N}aMWв/p?Hzpd3+N$h_ [n18UDW-QO 2bzpڕ y/pe*sWgW%]vO2D2:\5\UfXUfWUء pup2R]՟-cWK>d8G2[s &/p-v)}-=y\nW}+V83GW{1l}eڏ]_ׁ}xc=$WlCwܟ.\1. \en=u\<eF+#̮b}̭6WcWgWSED*3#2:\z+) ǰѽ+`w㖝<\e.Ő #\ aԞ%Yr|p:Rާ f27`VS̥6\!\)=+`]ev \en:uRB`2$Wq3fN( (4L{Ӥ=⿛+fum x$†$Kqg52r8jnOGP['| ϗ,opoPɽ Q3s*e,lIJlDPO+ 7ϳ)/-s k?|oF3%[u:5KɗiͧW|-RȵE3(QAԩQכV9-=xA/,pB~ɇk9Vf,_uYMǚPbNiP7 -cZl =1N|8"IK9 ý5$Wv5cm=gmcSAn[n;mU}.wWr#xX!)%"p]&AQ%|$!oo]awnnq-S3r 'tGmbyh7dX;Xyy%Iɼ%VLV;k5ߺ-o\-B8׬17Ȱ&ȕI -RVPRMIg3 jv= ; r0C'ʻ˂q_˨# Ơb/Lx,MuTiNVӹ[+ߵA՗5b04!qdye :j9zìT6:8E {f\CZ͙F++^֡RMk{u}w>.mXvrɌ=ʩE`D{|kڻC[k 6j#:n fhՁiRw f}kc!%ѕ3tej ]ZEʢ#+"1R!JgexWP]+RtutE%SB9DWNлBWp,Zyw䍞~ste2 ]YwX3AOWGW\!CtO:EWVu? #Ez:B)I+8%rW rNWRpҕTRPЕ+th+랮v)%Еwev߻({(Jk&w戺 \+te{WTdu5Z~f `Eء6[! t-#J7B+K:AW]vz!9r,\A\+VӕEYOWGHWѧ)Wʢ%t=]!]1$cѕ3tej ]ZdʢĪ#+&!!D+ KWʢteQ>Q5=ǼhK[˂f0EvnƑeAx楗}N&Ěѫ' >SӌsAjN5XZZ W/K֐qɵb}:U>;yNY+]jT=80.1Hlj{TICP|>Y<p؆QUI0~8Ç|QM`+afQI#ɤ@Gq .-J^TFYŵ6FeƆG|4zu;|*Q?sS޳]y4} v %&nS+D {-.I! gW|BWμ׵+I3'p<.ʂ (]IxIۄGۛRc'Lzb磅^%wyAJ彷oV.7A"PQ"*XX&\"c !"%sV^W;ypc+c4&g6@>U@8CrE80b$P~"ʍ Zb"Xjk`*=0_ Ux[ϫj]( bq)+8Zo?_ඬ.,^z={_.0DUϱ~ fWe9^E2|ZAۗ%mpkTEtcEnUb0,bQ6Ed2JTbLPBFh1,BG$Wz%+h4SkLyRuBDR"%yIE$d"fKSVc a0#( Z`'5\@q?bƩi:zXzRCm8Ғ$ CSFi0!IfLȤ֚!pO541$Jc1HDJ^KH$cNx-F7 ) qCuPd8fW;x{|62^χʌ}o@.r{|8T  _dIx&hdsja}Yc&Ad^"z'wS.;sxo(bZ0m l2AqD0 >JAhdG7l$x=0#/-2%QB& À&bRlbÍo@ 2_2xiȌX\) b/&I,~27$ȉd!%EygiO9ƈl~ \iDٴcc !RKW81S^xڴx+>#~-x=]-.W ``s|U![W$@?-?_R#3~VSqƚ\jj7VSUw`9G+v$OǣUƁVumnk+3eS[1!yfWg[lS(l8.n{pr]~w/w<ݏa$0+n^5;SܠjVP߲jUsjSoMR]7hQ9DqfE)\9O'#(׻ 2mo5jV Ʊ/_`0E9QiVbb7G`[^8̫\1Zqotۛ'{%z'PZ=)} [Һf)?M!;y S3`w~mbۄgp 1!1RHjbk~dWt?七: ,T㳐T"Jh#mx's!L(2L*y샺Ύ,M.eHDvdzCX9B 0KSU{NygQ6DކP8#Õ*רk`#P(CfBkOB2 |n&FZӉ[T3-BϘ•{ey} ˜V_Hq7 \ aF{kb F@^"ńWi˛4XZ_,CPg͠ՖZcKYq< 2D+m#0Ӆ۪}1`a&@A&KVEjHJ9)6I5n.bsi`XuNYdCr=N`4Iy0{S[m:g_(8D@CܡGTQ;%cj?=ۿ}:K;FߚMv J ]Fiowډ[}z8e.sc3 C!fayF: lp6HF7g)Oc.*5~bLj #KRl`݆2YpZIχ?~:!zBA:wpm@8u[(h&bFf y/2_wD1N:o굯dAjƤḽb>(~Fn?9}ŸSl1!v,-RslK~`;K*pBXYL^jʈhA #(H89ޛn7y ~]{"TOzV BvNH~ֹ3VR&R\~05%a0#:O_1u[mz$tt|׽:(슁Q_)[uFR.t: ;vwBV8:{zsg*kv4bf<:Gc]gIڿ{FR>|.yهsBfT.O 0 F'yUܶ\?A^ i{ʌEq3J[ǐ \0BXu. N Ɇ iCz.ǯuk1Y4.׈ORCkܠ~MkZL_]Ll7&^6 u06%h1}}9čEͻwT{jv%KE~^T,M֛ԛ'ڪ=+ȺgHTe4nw+ 58&Ղj/8MJΚYY1bO1,묡 zusfJqbpբY'V0w]XvRMԶisnCȖV60..F1I>2ӌrrk[vU8^0y2 ~oZ٦x=v:S[L`YQ=^ |9;4NxO\y#" 4o}z/9ޅ2ۯo-ٜE&eh|*(c"(w!XE-Q`M{ɬA=ʏoM?\~|Gy~7ߒTLmN54 -FebȢ #XO"V`*u{7SYF[D)*{TG3eWYҸe57~X^XM ]DV7VjbG"̌M Ha8A Vyhȱ0팋$j`9͙HdlI,}4pK {9y坭_?cSH-ߐ vSoI?ܚIP=\35ĺMToq$59Z1~26jPix`,&H!^ؠLaL /rmKO[~9e:R'ur&]3ɿ{mqf\F CfR7-榃urnzҊLim]_¼8RZ,`6ƕT&2ˉJAZEʕ/`Xrv>L 6: X[ֽ݂U|a 'O[<%jc&PfTet`F刊`) NP2mR-eMƣ46Xp/:;ӯ<轢 6F6@G}SS|-҂ǂS"\PDe|@f8%Bl#os}| x$ )[X1c2b=6)5[!-l0gj6F&74dS=/[w=zan O]rhe$]`a8t&@ ⧊a}"5.-0U8gmMm׹1mkoqw{|0[OwIs<25h}{j]b7ԕeCC]^!^o+*xc_ڇf0CǤm}g U`tֺ)˭jMϳ}"xh Uݫ䞅Lެ"L]Y|_įXo C!B0;;VNgB/aHqօy#ߗRt3+=?vjӯ..RvKl,`YvK*ѕo(iَɦf =~J4Ia\߶ϋ( PxQUzf>~~P"KM|L~}p/~4~>H`EyW8 ~9 3S%Š8 {?MWbsr.y9,K%{ 'zs0zyϽKO-Ƚ;̎Q_xQETc>?~^&Z}%qa2Nk: 0E3Ț%+ȚX\edBdT %"JK$]$W %A(•Za!JsJ➔WIˇ$eJҥ$+Qgْ]:\%)ik+1&Cp݁$]$S WW:EwL JJJҪj)Y޻zGpw|BfWμվ3v2+꭯s J* tg*+pW*I+UReFJ' h@`AW +p*."\. \Zi|pRb2\]!\1N$\Z i~N|8\d W aWI\"WIZv*IYqpu=p%$u@`!#AW - dBR %vK:WI\c0IURdvupFwT3q9 ix+Rb J+%\i ڂ$.Q]$UwW SW,+ L0%ǡU椝3 hs:M}7Y\sx> 1NJa?\\.a7ui)ͽ%6{o^`1IuYkK=|^r-" `BK 17{w4~/0e9Oy LqY,CjX6x]M ꠷*iQ[~[E XjLaxh\Pf$ڹM=gO\ϛElOiyCr &6Aelbc9+"Ug*KeW*I+.ReB"ѓn WI`;WI\WIZ-/@JV͒z2\Tt]Lt@Z.H)DH9_HʜY̋:}v\BT :q[f0}*<{x\~@*%a-< t4j!?^o3j<ן>%*5{Rj{j(tspROq|nAO&KO3O0﩮v.j7o4FD`EvaX2'f/e=Ǧ}<4s2%{x*FkۍHBi硥XO;"VCTݴ,{IEu $Y}*2ĉK ܍+o]U$s)V,I2U&rѓh9R:o- A(V1J&ׂH4jLzNѤhYIm01~l"(\R6w窻mLd()YLf c"bcrѽ EXb/9%|nhB2sd[8e mbJQqhSVO٤+C)玡(= !Ƭ&ރ%v;Ea AQ! A uQEL~_Tڇ.L{qA1#3X2d]șcq_4>5ysYUȫ;;]SI{;6`PJJVĠs )'o}QܷbK$]DȝDNE[RH((E?0tb曄E)>9* X %$ZjC_22$1ڴCԨ*O|"GfֺXL)&>*!kՕkʧMpS7BCJ}Y\`X2kW5JFuT(HP]j0IQ/3ޅ*) ,ڢ % v**6(:j< uԡiӊ#XQ?k(QN[PyU벫#<*!kBjcnue+y ,&ACq!ƞ$;u{E84iZȠ xOq u3RU@1 UӉ8!bEw!w:xaU.vv5WzתW f=fC&>MG[oT^FyYUL!Jrd@h2BCT<9? z4_<+Xy=h5'HiD.h2Q*YYskHQ4Xyy`\ih^Bd<—ud PH܎m }KwEWgUS U_"rΪڃ d-їHܩic'n:mOwu4P詌 uXktdD(c ٣Q.ɧ99hP)_C߅^c {6v5d*2;"=Y%2v6%)5cEZA "+P(vm!Y5M(#{t$-٢ Ac8ZA" tdp,FJ351ƅJL T "8tD쬮ZtXt:*BURh3ҁ!2PBЌډb|؀@c/y!mhfQ%7lmՎ6< cs|_l֫ XxV۴vӮ9׍$ n=B!l$G`3)-zCq.> Te aME{)Dҽ$4PzhvM ,!ѰMʌؓ¸  <%*`Cr[SP.Oh7"fh:jUNàe6M̀z=pe㾐:?Д n F]2'ҬFLEi8 8XkC+F5VC_ p \t*N6fTS.\;.8 b!C1+5PI BG!&MSxE%a4.?!u+kp N)sTjXͦoov6,[0R#bLT^_qj )mzYr~-J+rBc)q@]_]N:V'l.%m) 'WKK:;߶r}uryJ?Yagm_= 82q8=vH/7/H 4B4XVnq_ޜ|-6[ص/xfq $Srl:c^MǩߦHINםŦ#6MGl:btĦ#6MGl:btĦ#6MGl:btĦ#6MGl:btĦ#6MGl:btĦ#6MGl:b9RXudӉ{962M(0J&=MGl:btĦ#6MGl:btĦ#6MGl:btĦ#6MGl:btĦ#6MGl:btĦ#6MGl:btMɂwd`y16ߟf>@{62G ]}Г݂ -bjռh;] ]=AN$uЕv/Z㏝Jujb:ebj,Ra)t5}t5Pzt繤fY] BWc2H3ʺX/+^ ] BW-cҒS+Ofp f;w?}?^%H7lg7שD|Wy{&f]o0NU3W>ո*wU'O;cL̲?.cw-iSyÚľǚq-YM  s)alDeA5[FKt)܌/u3s(tr5Ch x RkKa3h5b?Wv+Tl;:{W^i+;8ɫ-^}tXkvݒ Q+6@ļ*!d[Vie}=8禮f޿'޼;?\!{~{#^Cvyf;^8˓͡\_R7fqjoh;a'T\5G`sX_7zR0 6zzZe䔌?n_o;L9E Ǣ |uy!~->muMzX~y>r+.~k*M{Dbw|.|oK H&;iaLRxccx} EvzslosfZ tx]XY51Inj85eY|XJ刚96oL)M-ߒ\q]\xE<3XIF|o5Zwa &jC )ePC%jײKTJzš$c%nVd\YRb?q~:br⨌13ɬ`)$>-@G,\d>f3oIDM"Rjg7i n}kѓO׋rnzdn B\<*}Tq_t?9| -×ΏE,por \򥬽CP7#6qh9Y/"{Pv=>_c^-ݩ;yAAc,6ʌ ;LըQ1,%! j@Cc ܈^O'Yln~=A}=lr]~͆Ă >.NDB:0Xp*P3,d|@fd8%Bva*PK>/MFd/D!e!x0k5f,`Hk1ͭRmlli@gӛɰ V4y7SZfZ!.ߢ >V ]W;]O.gEv8ne+D%ƒ.4g/bȀ.-C7'D٧Yٺ, ]oy:qF"-CnP2kf_RϪ Wf̭hǩVf] 릛]fDof{3^yX_C._{CNK|0q‰\%%N{p>N5 +AٱR8̨%ȩ܉q RHg 43~ [M \ޝػR6,NX8a,HD{P4s) \+zYy -8Z֌WC}?i#p  -, Q"N]`DZʩR ˽)ׅncst0A tI.uP2zEX\PRYz飶ěU`̙KهzGovZ /"_ hG+Vj7c! _!sٹH'3yg0Ty;:塧>>*O8Cf,EI%sĖa՞kw6~PR+XV[&x=,;UW~eMvLoj6 ),N/gǏ`咍0EF3fX?-(to\LU99T]L9&r6ȹMi_udXDa -J))$XO5 7V~ZʷܭJRo>Xsw;>;8[pEL٥C1%1]-IvBUvkG_f;2$̘<3 yAcCꯇQu uyƾ8KȅWQ& gu9*"+eBYB`M:"VK %94o. (WKM)$%J;f2@A@kYys;#gClj`k瓫kKS.QkWČ&-_R 6yX#1!*"B6zO -3vSc$q$d+ j6q_ .K"(h03,KЈ`@[4@2A6j^w_}, 4Cze]Rm9[8Ғ%59n҅RAECIeRkX ?7cKSmM(VZEcN0 Ǩh%j I5Ɯp cEzHξ%70|˒( &v~~{qI;"Y4 |j!\r^%i &_lV!idXqQRҸw~/r ~Cd<cr}ŎcSD.ѩwG&Å3X/Hgջ_%P]tb քYÆ&Uh$b"\ I0%1̹f0_i 9>7KS,$K{w<\Q1bXbV`Ⱥ&S/WW{pӓ[(UH Tf_7]ΪY1B0<'9-շ]p?]||̌{PHBxHmÐa4kfQX>"0kp0q1ft~;b8[39h~zm͕0oY'{u$ A|1UW$1>ƒI陿N%^ө|=p3ǿ>|w'׏'uN>vfVԣIQh ?>ch0^=4Ulu:ی[ƽ>C- ڡ﭅Q՟ߍܥu3ҙ V0\A/|Pg+*U[܅/a@/PmJh@N} I QX1 ި{jH>g'I]nԭyxRYOAE" Y V(VElf֨?*Gou_Hg2A[E;k?].I{PˇEy xkPʤSHTááTw80tEY !k? OIQ &uDfhic>[>}{GG$ Q:ܠu![cR;l*ЭGjd^ wE%OUdz{:E܊I F]%*  m75!;BK<;n|]O}`9f `hf)1H_X`nJ;MN8] ͓g`!͊At80V 6毸m|7rrY\5d5hY?ZXVt$Yltj^6d>6(l?JUњuu7r=ցh,5&Fوqδ86j ӗ}cB˾J/ >A,9HwyOfг,OѾHPnoFt/?A}0TQ.O"FMsFV;|rܼ8>tmoaX#20GFavzRn̐KpO Ao%}M21Qj?Gq|%ӛ>sWsps|5 #W'?~(j15{xXGZ D4d*%rXͻ=y[JxP Jx) }ZݝEߪ˰m-QO{0x8inOzMNzzvŁ@g=fU>>9[xWO5Lq[]!`+kM)th5NW鎮b[N=;#Iһ&nׅ[ Q˄g Ɵq^H ӮE2i`4@ftyf^E}i`̾ɹo?@'eF3 Cp=KYڼMYnɪ~VGîɜV\@l奷Ufd^'ʃU)PB zw^xmߍk64Ƙnfw٭ܦtZ-3olObmԒaYIj)` s2Y&ee{èwiَ[®Ueӻ5Y75%.zǼ=ƻl=?WǗAZ< \L&.C+a=P#KNe{^;XXAt9}k]!Zut(W؎Np; +lh1tp-)-% Qλ:EV2 +̹(dEGNWRNBт 󧤜¥G "J::A 6DWXCWjc+D::ED3 ++*&Վh=vB+N %R [[ ]\EI#Zvt(EGWHW CK Wxmm_tp-GOWrt[N=?#p}x`ȁ\KWЪU[\}(芷+k2j)7Ί++D)th] ]1X9-Vp+$NWRu)ҕ ^]`Cˡ+e)thN;ttut%_՜v+| c+DuGWHW(XgB~pۋ޿3I*VRŒN( 7u (ʆ +Ǥ0e3ŝć 4ʈQN?|*iX2[ px; (5V\"Z\NI7dNWRR ҕQ,2J2N+E15k@)UWl=EʢAD9tpe1G_@FttЕrE}Jlu4''Vj\ȣ>y`Ŝ<"ܵe%nf%Y9c<tҕg7G~ZI;SZ8y[[H+(;b qF;Ȟg-Blхدzj4  B~Zn] ]1#`BC4+n BRCWWR ~t(-J "Zv+D@cGWCW #xAtm+gCˏttut$BCWWb ң]!J;:A2jK +lX92NWRu2"]Y"7>NGb*5NWwoS/Q2 etdhP2C2QЕlAWN=(/Ӧ\R lV(#] ]1OZ]!`K+'R v(e]"]qj.) "\J+Dˏ޻BS+={ܔ]!\+J+@+(;vBLttut%zZJW BBWVst(i]$]Y^J $Qd:?= nq'IbHUTd]d"xywH٦`rT o5&5MҸʻDJأf;5>[q&Nh וT8슳 &V)q0>(! )T#הQ<#EGiQŝf?vU*i@t6TR~5rd"WPDAdqWB #m#SLe&T48CbdKп#$J:fL!B-s|P+qmcsa!%?LXm E+ !d5r~%d¤$V9"r4BeLa5KK%?$cSZadkqk&׵h=֢*$wQ! O?ᤡpF4RtMjYo]ԕ VŔ ՙW*#>R8$MY˵F+R2^s { i1hNTZ\l+7)77~j]@,SH@UN 7[iV DI# I;F ʸ`MZ% $3$.w\eG $ -|uQ; "t?n'M~R*xs^a<aQ+p(BO쑦)hAL 2]H41V)pIGo:Er6 pr+ap5yp8H7&wU)JK,Ѧa2lpx|*Y3!;yԊ诃^!cz'ip>"7vp9ϧwSwM?ֿmJl{ q`~6\ c f=< );^*A̢wwן^nu:`6ϋ{7ف R6\k~+;II ` ,AHXˍ?Η=uTt{> խ8#5^%.n1K&B'-+xG<0G`:& b<5?,,>)Cii4R\^7uy_f08?]x-nVިww)ܤq1J{mrMfݿ  Gw5c ݊ڮ+MkkXoMrNk'~f\lIC.mļO(4~=7>M{aHl~e~*+^ F/KYu(,py1e9D+ԱmޕJY+.AU3e4v}~͗I{֞+gFc".|ZāqZ8Cot @)?+e"aܯF-B&Lp&Tl'8c 罥B$2Lye,wVz=S]9tNףS!ceײ.n&sϷp4&i:}9[_zozZ R{R϶,(|ISDW];t{b&gu=~l]59a f!e5}Nz^6~sZ[y}?}6:?0Y/tyӗ =<yk& *.uZv"}+ oN mރ]ै? kL/y\FxՉhrσWTW77ΰX7<\k2-"gt 2!KHfHtaɳvwv"'a[,X}6$ Kɱ@CTZ3.1*Gkb>eK*ŷWdaQeR2:ǘ'F: 3 81[ϢK,NQQn2^[Y:Һ7fH͇['0{b Fwdq@hYBT]d y+JDlSblf&aU)>8`-x[e! RiTG )54SL$( Z0B|2Ntes:l1k(!J pPڬ4Q ѐ|'Lh46D~:K{KK+(Tx=Cpg41?Cw] qڽOSz LP4Z0+ٞK{+T>4~|Sާ5Ⴚ;}9/Q-2W!FS&FFY ΋ٻ6$ %1FC@>8 -pXO1EʎAHZczpF5=Uuu DpC@x,s$#]SjˊGs:m6d>>M8j!z>Ln|YBʇoMcdYkWt`:p9QKOl`'¯؞6-y s=u׽8m_IڹT̃VR& 1k"QKA~}X5u[]jMr|{w]uwaPJM \V t5+uz5eҥnӖײ0;rJ |A QT1DBDFX {buhᘶY-C#d L>l~I6gy .AY6"V˾mZI%HD0 AP$@2A6JAw_y, 3(|)7N?ʔ^%XJMtvv::(c,QJ=q[S6z٣M§k-*93wgiW~-uR^xy=?x5V#3b.͟Q/vmEj7rՅwF?2fu&Wt6 iGX9<Aa4.u`ZdǣM휃tFk66k۳i?442&>O\wOp66L4krW__>|&ë^}xK: LK]ZIP&"@Gpg~S󮦆bӧ|y;}yXټC;p3Z 7K7ri:ZM t%u=o1ET-2U[ܥ/U 0 V>[1Hhb1f Sa(р?@ vw?m%AF=sIM^0oO@c|ND8"0>PmJh@N} I QX1 ީ{HRԿR~p<}o \ '\ܠ",G+z63k BkoBu U#ٷn70O,FЖaz.WPo$f=([QR˿gxf; KEW`>=t?顒j}o9q _iפ+j:Rjsfȵ1OgMLymEzO=P#!DSVwu P0vYH ^ wE% *{-m@fc`*WY"zЗ:ކl8N6>l %,%9ɲſE. o{ջy78 D֜YvO9>p8Uս́꽵v(]c# -Eyy$q ADrUAD|? γ9/jg 9]iy7 3 03WHCsKFDOAJWݿN.:|]<1xO^wb5ϙϸ_w o3v[ ^͗O] jfC;ETj91Wy^2@-OWYi nB8mOST}&wJ;٬d Ks!,Wu. iSbC9 \<<ʥ}OpΕI7̇󤌗*3"~Gu93|5mV/Kśϱ77}uMzx[jܷ>~-5Jpԝ4 /&p3z :$ Ȳߴj:.3\sە>VVܱ~:TB\ M/rY^2WHf/`eۼgKh\YzxbX[Dˋ^]dJ7u/5liAdGֶ dp!/cZnYM;vHJ*)WTߓ_ÊYVz5 S^옄|o6' S7 c.޳'mj2\jY߸ch),в@]{va`pNxSFLQ*!y55渞޼-T3։.IQEWwv'N:>!oE',d=r8m kmnsSݜ>G}O9Rs:Q4*˹H57t!XE͵Q.gI{ɬAƪK}ރhyM$EFD ld;o³jS5N,r?3oR;Kt|oB6s7N)cD 9x QiF!uilzRE'_= )@'ՠ#Tn&ݶ!dF@6)uvTR*i2)S=B&z 7Qt4W[wuǖiJ\41|'W)OnT$յ!& 8YwA^#ycsλ7)fn(-Vo$EDoamZe7u%n}k>,֋R5]sswR }HR r.;cn]؇32RX2?Uix$l8lVy sZœdsk,QV8d7::<|rwx>?2=il][^>^ oMh;xhjouM$PC|Z}?EB:,8(9ETǴDZzg3L̈́L@$Uґ~[T"L"#ƒʽf3LG eZ30{-#chnDt9[It!e4GM'7e=d&7_l{Uyk-;o1.LUg-RiØ fvе~մFKkI&+&UqdOn$Jwv'ۮaVj鋵qm]wF} xfnf5oXnf07rsO]&[n`[mqmڳu*s9 Kz|o.v3xbǾhЩ.*~ Μ%xF \e]M{2kn˄I &? ϱvVKnX  *GY|]%v3[t+p5Nb'bQwHuҏcx*n#q$ ̧ Eῷ(Pi10$5AZX,@(vGEEI~]`FʩR ˽)q3r6h~dI 'H k>rX Ҙ{b˜V!!: $S, A9t0I%ߘٖm"E>v(2ZnŦ(:nMe@)~~u{4*Er;CveUsPUB[:;~"/:/^Sa'{.̋I[vf_SN%9.232-+˽rCRciLrsߥ 3çDy<2eߨCz_;j:Rj f( Hk#ɝZV &mekûOIY6.dB6Ԙ]a#f5.Zzb3~dZӶ9s`vVn$ȲE:z@2,l]djy$#yZIIDdrO%|,wR? z0F~(^ `V(8#ʥH%B( *"%m([ַG~-g*IC0Ce8`QA@1g9R 2H=PT/G"`"RSFDD b FрG!3fNCk\" {9 /'i>s.1ϷyO-WKTYo5Ae$hϸq! )iL! HĬ ƒE*LS15t[gklٺ=>Bk׼q>z?YJ_69)jXi&CV|B=,+6dXxb,;Gi&I,mCK/C?IZQrʹR46 G cb0^LYBa62o0\ՑRf'oTPBq\s \g5-5XDýyHT"VDfrKq2<(go5F@ٷ Yœ +V YoYtfT'Fi#㐪$īĈ*FV",- 99t!^:{qFT)TΛň)F"F$B.ňF×{HYپNcdޚt FƤo&?E/ۖ~('rNޢ@n1xG3c^pƒ\IxܼѼtFc+R W iʲmW NIz0 \ Yw4-d;vV] cl9}Bg\fk{MUc%R 1ipVXg"bϽ2XF.(nvDz-, "`\B>P0!S&J;f2#ʃp-1KȳvٲxiOa:ne4cMruLP՚FakrYI.5V /ke&+q,\GQ !cPaqczfq``%g:|15YP'ꦢ XIF(aEcXV. "B'H&[D6c١^xQ31`kRiIiOQ7Ipit;BLjG퇚I+"бIN Q#5Jb X's-G@[PRf u>HE&/F^ KzЩ~p>v{j!|\2L;"isa 1<I:cBkm)axo\ۻ3JuTZU$˲GYXL&aگdXb !\ܣk)(]0pEb|m(.:mqkBR0Ifb"4*-UE_m"cqZ; ~,Ɛܹ7@2%&+Gw.Gw_c*Ja -EW Eʢ=O0W$ʇ*U&?v4SݞL8c]/G7/ϒ%if̗n?7URR3E? Bu) YSK'SF,Mf| cԠ:O3{T%]wpjS%huB֥ZWV~a2#9t J-!_/ *;WXz+^*OYۯ.w\_ÏW_~&_}Gu50 7jv?wʀF;dU 湒&IS'Ne3z\MZt_fi傶sk J?v~cE~*Y'Yc'@w`wؙD*s\i^r*&b;"P/\^c55H)ʝofx`pwm/$eS+BMGj#PBst:\H\0UrqCL碄S!ăVI xRYO,qX#D,f9.X$[e3-F!TCa=}kM7Fp[cO+*3QԉNOk[ [p֗(Idv{MKw6j 06/ˠbI-mlO#Mƞh A/9 C!faG: lp6HF/g)Oc.*51VٍCH9\2&bMBu6mi) w3fJr? \D_Uٿ>kM=p{ ue{FKSȊi!kUՐ%mɰVi8L(L Vi5ol"gB!  -,% }K2M{t[CDEy23 p!&n`>Pmನ0x:-cSjH|Mz97$_FTz@qD@ ]WA\ fhʥ(jIJdzLXaMq9h)nN;[$ؤqoR0#4'(xAcf-; \/q i.fʳ`8sy>PD/|=a=4UĿZz:O`^XĴ|H|9 S}6N)lNSn;\[t:% y/f3Ѹ+OV%FCNXZ $p`X(E!A>!{Okl|kV>+z4_,:$2=4V'at;}mg~Mȷ'*U[li67]5=t^׮Vhrꍄn ~9+^Vj^zȜm7 Mޤ7٣"@zOCa`:Ɩ5A X*{Mj1N0-.˒\-0̅u\]X-( Tӥ^94FSjnܵ38`jLm1.az|WbdF(wJx;ml;S:|[cVS!|߸,~~+*h|=5X27VcXu5,_Ǟy^d:'Xb㾓*L)9I˾e_%+KKvͰa{?f8ɰ7 E*XT ((1&? cƈ"1K9QDN=EX͟wm56XuNA + :gaMsi+\qףT)“*CA ƙARʈ$g'nlDD $$3 y`nVFk z/J}čC$JMV7#gyi ."R7%YJfkTd|zFk%%⯬[vtvD,pͷCg3#X[,`6=ƕT&Zz>!|?@0J3#4㎆#F0DP+J9(O#ޞ8&gZf6!jh9_$\3#>=[`][oG+_62R/ൽI{b_R-qMHّo"9IĒEkU_uW}C`A$h8A^*(g|Cv,8X*KQ\8Vbo^y"EpHl!P] Gȝ9VZ5G敓ZwLRIQK$QLs52e\T4F 4 N'=YSG٢o6S`(O2n׻L>+X:oCpiK'+1כ^1!6~?z_zJ%M:8)tP:&HG=YDcK 8vS99)tZkL) k880Fs (hbZ;) A8<5h j"Q^4 NT+i-PӟB΂ڳe CIj 6jX"&* Uʵ ^S*yS#+ $D&NETh0;M `d1'DV;9<8_j: I>(+_`efI?pzx ћ  lP6e҂ ; סDZ'I +vK9?TK?\|>]/[u^ 򪈑 xkYVAD#QFČ b,c]yfiʲKW;S6y8Q )lg8TIQWLNVc?)Rݜē"CU OKKɸ s(ʾvװ㛌xυrT'p@)c3(#n(*ztYIN\\v_UG%T fg6\f0;.ig} f,A"ĠIAЅZT裒,u:QKl(=-(v^ |.ozmt'Ox>cSl"(+@& K*x'ce4Ie|N^t[ SL_"j=X$(Is%eD.4$ 3%T&ectCS$kgiGs8[RƝgI"|wwxo^&yل;\|'\0_R & II DO$`g`07rn٥Vo\׾s>-FE\ )XO] MX(.lH~PqBbDn|e dv~Z]Vԝ֍+hpfߌ&K~|p}3ß\pI) (S!_h2.nWOg]^rw~)x{7n-F)A'\?i_^ ۇ(BU~7(w]dr;}#iXp$!t]5 Ɗa#bnQ8yJv},y8h2|CVZL0|cKTnC]WjszF[ #?o钛*ԉon_Ňw~|{_퇏w}w񗷸g`J"c+,BGC' -͍X=$>hG 5wD0Y[KS?vv?q':=3+(60o-J6/eApKbCl?ӸDci4&v\̷8.lNZOiv1}c~Qu0C!~6N#H2h~ܬu\,W)N KI0_xu),"`AJ3' 7Eu4ǴC2=Γ~t14J` &QKJbxeT.Ƿ7~N4^x(Q:b7I%kk;]7ULLj"#%W/Ws=Xnn_G&x]cB]vw\[1AV$b𚸞s5C51yy%р-;B: Q:iצkC0|=@y6(ߜ8xk8aoqs|V9"@yS5,qI9x.rYCh8zjӍ8zf3o#a. rQ20A0Ov }Oӹqjja=^!w/E $tzдl;mKVe) x F,Q r'1( 9 >fZΒU,SdO}v]a\EG rԟ <,I htڛ O<7/=ڹX+WD5XUXN>M Y+E] 8VuĬ/JD1:Ţ4x_$j)` s,N6uσAz +`e=X+amg;;aXJHu H%T]Gׁ1*._*:^e.kT20H^zK2A`Lx+SH(5 I2b*!"tJTģP%y)68 W5CRKű~=Ob鍦m孱7P@PO C=q?7VV])3yfuQ&[`=-QL78oxw_%xi ,AZSy#. x͵r.e")z%K:$Iu&AUI !#zj1!15eZiS)=t1㡟O9q`IY]ȄRt:'Lj? >yi5%B_~n2壸\SGi8z|R2@|k_x dϿPHEh/R^H$I"[gL! F ~]E{$7p!ysCH Y/bu1cDX1Z#wX543 Wӹ{dm.cɼ9^y1Ƙ`Amv]_к`&vsuszMtKU+q{=: kGQkOӅv'G D'%D#9KGt@,! 0 SP2pȈ1QBp!@E1G9p? V"1]4CX/ ƅEk{ՏvԷzIY!M aSD / ^P3r FB>еZpT&YRepx1 K&0QAqFq5[pm⬶8m7H{'k_QEK&\YEQ{fM_ ӕWD \ f /#γ$WyahPOθדӭINd`g\Mz%󊶿JI!>w_aa-[n4h~•9 ݯ D2VV{Wƒ\1OvĴTKv=ypYRlHw  4Vhn\RYHr9Y[^|+"WgL< h<h#e/ftk@lwIALsE'EgӚ[.\hDI0ƨ*$Q.8 DD.Jhd@U+MJVYXf.Kz8@"\1{tx+VV#pJ7_^&O/poհ|k{/<~ĂCtVSmS{,pmf^[㗾[,7[:88hȃ;,@hxg 2'[-F/TVwW.uܼhn/{h{֞j^xJ`zè|ϓc9s?!Ƒڷ>uq,Xs7Wil{tΪ*@ Fg-ϭg͵c\8[UI}F m uc;J 7}(&s3~[ 3땧}b{R i3\(gd;uRZިfmM1{Jk_mm̒ )`y1E3I$!Z" 05,[1M!ŷr>g6⑧>kI f?f7U֟1 &R h&ge.9̂5t@oH²l2 SL,h l- !>fE80y25$сX&coHƜiP]ֆh@"H>`!餹Ɇ@pVh829bn$Iژ6i A(*MzbYb-LN*&'[~oN}Kq(3q:]k ~U.Ugɉ@J;.MB#[HyJRJ(0i.J,; |:6td^߭3[􂖞woݙ61,롱zoR|~LsdSiOEFAi V݉sA3=LB dC`RfG+'aP$HRhdsICQ-ߨk/) kTjtla|/𦁒=u2Ce4<+gZm>׀gy@bG- y[Cw.b@yƀĜ cRW`' :/EP;BM]CuEO*[1uUub*ꊞ8f.]]*kSWG]706׀U!BxgP)zJq7:VUW`c' v2gWDfUCSWP]i LF]r_5"uU\]*n=+k-_U9?Z33Yaۯa_X^viށْ*C8g ϊONWHW0Id\1]3F`@~`iZIs1mگm1Ԏ 0t܋GJoU&SR1/mfJH2i)uftYSo/M:f#}1ש(d:Fai@zżI:ȒŹ5ıpԉAxF+ҙ >Rȕ|*0P 5*U# 1ӯ6|'i 7݊_ZWRVH9&iF6ـG}FWu"RDF:!.f/DxjOI0HIgp(1Zd<Ǩ,FEh!J!etI#. a(YԋI]N9} OML.'@&Q!ƛ QeY|KiͧNA0, q(K#'\FlutsSכ,|_`^Nf-dfco%:ʔ6Hu&HmWUT:Yu:0YrY_WT*'(߻!wJ@ *76L60^ƚ.0\gS*DTFk4a*tOj˯%&f&_Fh/ŝ&;y>x|UO^bsad!wK=v_:TC˚4Hm\s}ȶ7;Eo{O8,|:auYnO.k-2{( 6؂d>!M0`O0w<&{ + ş&xaZf!eRk?LDV5t\M ]_΋.I7M%y۽cgH_= {o4mElLO\idc;}9<:6|xxtw7"\82Iewfߩ=8]XR~D13y6x%;uҖ.D@' };aó)շx[;ncG)X'-i!r,UsfΊ,eQCR1Xլϙ9.1!d/w58 Ǘ7IUa몋βJjT{}1v}SC(@BFqiP%\&<'נd$oaDŔUKBgAJXiy@bd-:G/B˔$lbƪmg&DFz MDZmEzń%X*}nQȾ6*?g4RAљБ1W~zv_E e5Qq!4O HL{&gh SA+ A'딕2HCR7mf?uy$&8>$M,XZ&e %9%M wIKɧUjWNsrL&E=9dasBNXH%PW#gXYo@ooyd9qyg,K>D>\%APkإw-TZhdaY4b0ݢXtl>U9R;arkn>\C ɛ\cwVmTZ|5+ Goyn7?|s7Ǹrq4W}%<2בjzN 7}(&s3?a-~[ |uOj!mZ ̟l *vv˝ o˹|NHN;سnhS.\嗴YAeAHV# wL1-(Ez?Q&O,IULq9&.$NJG1<$$7FHUq]}l][aMˠb1Gmr\Jz'?sTĽV.b IeJq#'Lz01ŜyX)jfdRLj hkr{f1K.HPF+S9Db%r3)ښ9Ú"Ji1Mi[Z]<هuqӛb5Z(, FB6Apd^uYKt)`l ?fbNKo GԐ@d:9d2$L1E6h$zH=:k2d ɘ1mhR p \G,$472 0 #I/\RW~=kc؆_THH-/odVQlHņ+wDDPw%zGiuOJ;i QhMw*YgUЙ9ƹ9ݤ?(F$nά0s[mz$Ի{C| [8&Nj-ƫ*Igr"4f"|4ʪ#(W]+}$BC΃?}{<Cnj2 %4ڇ7¾WUYS]gaf#{(_oB eVI)C;ieAneMr.k,PT YQPEkWx"+aEnc882BxC 2W[PZtUnDʻ&b9}9:x%Cv&RaM|\be&V0VTeU\Y(1ޓ54; Wr{.ZT˜/ Gje31ՠ`gg1( 3&&+md*$3 -4%Z،z %pȅ`I%S.BSBP(gtyN%hY6o;>?|_w̻y^zҏ4Ӌ,ǛEz5{ʣLeYỨr<2G@l3fY2%AF%BഋLI aIN,@#:#05 c0uMC9k87DPKU(~E> E8 CAG+ĺdd.ixpHHI6$qǤ`Q HDƣ9/mu4mBH:I+!] #1IZ$!5w";S*(Fb j;Bj JJm,^)ތƥ|L x/ijF!4WE#r}mȁo8I<ӁJB ﯊f}CI{mA ȧ'=E5.Vp-b3.8agFK>l~z̵?̃J0K"jß #^~뇋k`%m[tKM͈f46Y~&BL8tB-ڭm?.֚|o8yԤ\:=E*zzi6p?Iֽ^9zSB, MzEƣ"?##J/IS(:uoP_}8Mc_fgZX"Vh{sn*?~hty:DkIѡ)k3l>I@ݡ#s\դ1iH9ӏqBz{i-A 2y zsI3dɩ Ü5Hĥy -,Z//oq1?>R"g؛P"Ͽ'Kgik.H}K~6Vv;ʶilx2wª-360|`ȥa&i,V. laV oWi[zY)|9ޡSZ!:xpmsSV!s2IWFi ੐6a`:$69X`Myy95P0FQ=s嵔2;M֜lt ԁ2eGQEN`D, u .zOsGonuma^X.xE37?LW~@ m 0&!:] ke* *w:`I\̘C_̘])*!TRͮIY\c,o9&l-wZ#AlU+UqÖGjJsΐ5E,B0#6Luuy)1:&%N8ƴ3o&{|=96>;DG uX u2n NmVcw[^@Kl_Mh#`$J2i*H*ITRDph˰Ce\-u6葋I6%BL|L^֑H:IgHEodR%Z#je62FL ) YJd.s>w*hn֝ f*OB*Ә -@7f`<): T=;7zt3Ykٳ;kQ:@u n[_uʎϨs*]Qzt5-zJegJVTͫ]7Pg`eJњOHz&{B/=Hӯ>Sy~UF&bk·?/iةCIVRf:2rT/J(ol2A}@Lu VR$+'*Ē$\ d sF<烝~ǥ ܻ%sx\*p盙%>4ix&KȰ3|~/='/gʺ7 v1s@b J>c˻([yH,1Ŵ02xOFћL?)0Uu"3N$ \QdQe+]B/"uPd#eubkyD򶶁(GcU1lmlsU{]ܮ֔)xIʕ7H&<uy{g0BQBfDr͜43BI@ZJmv/Z59!`S\hU&eDl j9W319cky3(zb:Wr3+ &?}z|?y*?'Z1M2!) D AC2% 0tY4 HAqҠe$^p AG'Q!bJȢ(ltbI| $qBho>Pu4'DcrƴGQ.ke2 r!B57m$ 0QmX:J'4(Z; ٍ*I^ȜDK7'1\2[D]C  VWa2XD"@Ϧ9dQcdgY'9t ;:(BH CkeK1cIgMƻWV(녻F.MU`Q`VO~T U. 8*Sp#d%PZXD Vs"dz]xVK/Y6T xw-Guz[qGJYjn aW$Scl&mZ1E*fTb)$ :8J#[$|iw p꫇S?hVaf>49c!%aƤ-1}rR\&9IVh/2f, {D.K M`]୧I:PKвlZw6YE$_׭Ʒ+:ZiǣG<GM~.6.e5*7Ê)s 6c E@!SdT.$NȔ&_=b [M(]0FL dc9x$E'B@-Uyt|Xz7mK4( :b]!֍#s9HÃCFJA$;&%\aDLd<@ `s"|?n?Ѵ !9l&qrt&i%܉sLZDӾ# ipL(R 7 o Kޙ Aj x/ijF!4WE#r}R3"qޗy >^B޵6[ٿ"K`i`>L:3X 3~AO[i[v[{YzXU,V(^\s.7ߟn̯4{O /4yM1yk6 )L & Ӌ{|q$ ;[wb?4; g!OPEtz\5!*fЕQS x0iǺUGmh 0 ж/c>ۿ[cU4\l\u̖Kiqvp~8oio;K~VD9 8@/֚ kf_|ww7RzAxaFZ@DBDHE3v: {wd>?տifsd`BK]Bܨ1SG' 6s,5\B}_xxwܔY< Cz8oZ{霃|[3R%gbͥ\T]kn\*[)XUz*V1xgzahF" }9|6 8ACK2'2 #x!5ʅ6AĹ5dܧq -~b/K.^@iMAMĒ\T߼3&H e0{6eD[ܵ6E/t%]!ӥ//0KoC'7 z\O>Zl-x[m.L?\jv19 m~%c8;3Baayjs!.;<}6MfB -cBxӓ;l59Q|XJ4͜ah"~V YG.`$XCs(Xo"59 w_-'^mUW=>;+Jw2[qR [^!#b-):<2uD1P)YHF K93WSclr1μ xCqy&9}.oYm0' a2ny[fA"GrmCaxW:4e$F$(,N.M#*uZ  ʰ`%)FpNMqT u8|<0^[03.'w3:v}oJ Y*GMnBK7^Ҝ<+IVr3t0$bD1ƚ7;euBثrbc˥g\!ՂK6t(20| YKιhp D_&(As(Hg8 G^kjE4[QO^G1B~TErQœɮ/7gV_вVHuH1S.3A fgSU\jZUkг7X3},yאN?+#-saU "E%( 'PdsT5|UOٗys m\ g+K9DL!ODO,hn0KeПW K趯IGc"Â3ur/ۖ=V(2a4l[{FBgR9jA!A %3=+;jaHRXRTgI?5^0)EeFXXv)im>)D"xE.D`,16p*)c'2$9bO1Zks J*RBbN ЯSsjj.rҌIXr0)DIIdq̸hu6E߾x]"efGQ2L#gQ$+eřggZC2k5L(JNŸ$Dm :l$ YX2$KM ٴڑH od,.L-CzիIJ]F=kI- ns ,ږOnzyRg>rk[u{6?bN@wtW5:0%dz=5l]h`ΞRa#xcnihlS*--و褢Cײ\۔Gr(V7P^ I7&N!1[GɸV@H$,'['.%l4M$<:&#'D]llDrr*C-z?W:~-RYOSkR필)3Hh}HI qX ENr{_`$S`U1L{4ƙsCo5ڬ1^S!$4lQ_r!:7v)"5z˲HGLlR&"kMKg"{lTܖ#ѶsqDH hhkO?+ vD[ q%,ZiTU>ii^i^aX:U2a6\x3bLDXQ8uTp$\7,}ZAZz> ir=לL<<_ʞm`CouEXX3>QGT<ٝ34Z_IiwsR=F W#jaX?*Ԯ 1O1CUYIgXaq8sZw$iYřgVCiD 0r0f sg:*)뤳X,EqI@ia#x8 &斋bWtLa˘iMW,h=% ,񩭱Mؿt\"^W0FRd+aPD2Ȫ ຽV;}fo6RJ;I)NIؤȠX[Y; t vϳVң/>;z/;p9">1| B4d!Ps#-WISABVەaV0Ԝe?Ҝ#,IcJ*H 1y ZG$ 1.'rR荌^SAZea j"\}YJd.s>w*hnXl Ά.)7X,,_էx>j+¥\k^u,v3F8[fͯ hO;9{Ϝ6ngXz뺤ǚ9MMF?f\}D{D9e zȀ+);3ו,z|ǥ ܻ&Zy~xv|}ϵUFYy}dCIu*v382d I2JN43BI@ZJ6xmMN)ؔ%D=K` "yL6gI +Hifl zflPKib:yŶ ЊlMӋS?,ΟP`qq\#g@9`2!EaM4$#\2f]36a۠hפG4!aYi#ȄQ!aJȢ(lԴ>A8YhoB¾dHƜi]ֆd@"H!h5jn$Iy`7K.-dl&z'iuKIk@(^H6GKR4'8ALڻJK8t`eߝf2^L?sWz- kp3.0fZWʯ5*]w̸or_2_<!x=cKq8ᖌ9g\2qrNΌxc O' \56j嘣 G0x,4cA p'~<Ռ<-s~;|!~w^]ޕdKԳ[A՝T3E9v7?W+yA/^?Zq^Ny_ުRwfwmFޮn&,+g̥?̃oۥ"jn ޜs8[Ւ6+[pKq1pTs`ŲO߶ٿOy9V*_rUU.We4hstˣo_Wv_M}!_XyS.9}z`ǶE? MMK Vѷu&6+R|hZ[X[k}﷛~X<8nWME*t=L4 &1V{<5GE/SM!t)oQpxFM<,,}S(2:C =se_v}᧫cfH -L,Vhnf:]_/ۨ3̒hBrfFRkiѡ)k3l>I쌺}E1̋ym"QYmlHEcLA[Lflp99qBqrsu,x#_k2yJ\WԑZxW :`w0筴1L"^!uq?IϦ:*ͫ3T%'8|%2t|U93dwAB &DNŷs"{[mڐ{y^ZydJv{e@~M2ET[31ث;:!-cqD~V9t3H5 Fe H%7Vr6ur͂ ËpFS?Ȫw{C&Ǔv5~ w_KtT[Zhk kbiZҴ,|kMWٶJ/h>R]X*X0zB!f2m+@/ 4gOTlVMW.V4=l`(DlȜ LQ1xI00JB &Ny\\v*(:zt% `ZCWmvEh%cNW%g]@ZNk]AYYEk=twtUP|t8&U݊ Z# ʎ^$]inV;{8]5`A{ OCmGW/ vyUw}_\3rH">!S!z *;IU2ƬsA#tN-*oNW~OWefzRJj{/{mgi؞FDl<U ӏhVVA,]'/^,œgoL~݃}-ZGv >XcJ*d|L^*'7YF"yKt}Za 9@\}ga)9S,ީMc9{6^4|AO`'7WAً0M?nvý&5.Q^cUflxӛ,_641fE 7cլ\keDn%m:R׿mNO@R1^;D}<?{`hc=ۡpswqg!_|{=F^san8ힸ^܎W͕?qlc!>KCǐ/oj7@Rz5M` Z?z/2E%!߯{xm<'(T |TC8y=jB[6E-`?yǫa9,? ie߼=f[O\{0+M^Eyjz `4=0U=o i!HBpY4O) Ke:M!fb tJQ\̪X?E1S!*c;#gf!#i.1瓓e ==܁;Reg}UZ?uj5%:ikS\hU&D9Rț 5s-cg춌QWIơul >-h5$/gLl8mgǟ؝*ZY'dBjѤ6QSДpDb̺"er\X4^U`G;-a.4(m|Cp0B(2̢Yب8LD "}yԱ}C:,gLΊ$wE.!Q"Ff# #]X˂QӎvԴԴWaOVWV*Xa5V%Ws@%+_V4xZ RW4]쟌T~\2Z0N ڭFi$H/Kdf!B'*ǽwǽw{TPĤNd|%s^sYq< \rIĆ&]]U{ÂD JH*6 "U'I…“mb:2}aٳm|zZa{qpptF-ٞO{j߼ P)ml<)B{HiUb==%:bVLUƋPC`tP:D&x ǁ>L,CJ noJ (]9xZ@ufڪkUnڱlecS5}ժ{W%v7fil-TWJY.k:iIvR^Y}~6vj[ .?6f 3COΨY I+[jOtgE='L:3q&]M\+8?&ݲ`yyz6Olzibik3] ONqs*\y+-oE/Ր>sy16|=bL6t6hf}4h(przhxdv`ZglˇlYe鴮楎ԉ/IC_OaJ.ϗncuP÷ !;l0_ӏ7?9o^,G8VIp|1S|˩S+S/ϼ>r˼_.q/o[}_ߌi79F~5*zzi' j~9_L{˃r*'-U25&C?{1.xN7n1[Sq)=b4Vc,X?@ %X(gUdb qJZqO^߃.$ڢ)_[ymr.JcLL6XT@Ǡ#KoT! OsO%bVVq-1&b.栭QZ`K)/v:YK9382yFVcb]z%֒{]KF7`cyxI+v/*[=/޳(s\fJ6Lm[֋<[î[ܪ%^]ޚ/|:hgݽ-֤h y[#Ni_1D{>B7OnJWonWߛJNe{y\z;6J6]O ,370Pn*/Gq k-sVhä聼l}2kͺ/aúlF [lR eKf+D:8p2RUsK֚: ~,: \# P G<=89~RZF\Z#cctr &ÂJdg=SrOY8M8| ^{ 'r'&5b2$E'bX>:٧lC*63H-Ȅ%%]t&H-Xks*q|yy=jR.䛢PEexkU6T3c9$D a?]$$xE!̀EE(% E,[Wbi=!H<)G@ 17xv)I# $!!M"A:P`a4GAֆUJT5%b<g*2~l.U2ET +邯%嵚\gk&nVEnXH>\j=9J o,+ZǃGk#HuVw]*w:;!;,TCKR"m:DJ$dFbO\s٩_2Fi%-! Vx'ռ"Mr#¶KX`ѐ"}j%r3^\wdv4)û 2*VNkiG]K&0<8^vI(e :-` Xv携lenGV rLNʃ*tdF" y,l/푶~{wu2u1qׇT,K0 w@܆w^4s7Db6y җq#De{5d4e@{V nH*(S`b @p L>\;dy3qb_=>*N0 N&(.64I8K.F|d@RPH;H-2z`,:0"Y`h#܁? %شH>K4К}NoX^WH^Y\R1M`]#53Kok6H*gs]eVGae?h,WMH6 "}&J X}0sɺ6J6+v&1ꦏi+i<\S;e Q"kK^  G6T\1 0l܁Qfque r0Xp>4灣dj `|pH߼)}>sWpл!(^oRPe-B? -ls&ʦA*BwPzB #RGTi'O H7l]_۪Wa]IY!|⏞-uErs>@b9v%" ^q[0zaBO(mQBaT(>6#>z֘*t 2WhâHY>zV19EJM0RO 5zUrUKx޴l!WXӦφbUWR竉0K1:Fff%Ф1%$"0bE.hI56ץvkT}"!hTv"#DoaLn0@5a^u (n_"CDT¼;ǣOB@GΦ%wܔS]u Q>&=g'ٕX}}^9teS+ .Jc\Q\,~7#އ##.V n;DJy!]/ ȪNoΦW\::ûv|QttH_Vllxc+=uէ7Hwǿ>_ilJA^zYg82Ͼ"2eyާ [?7?ou/x~׍to>]]kZbAkqMR-Qvh p=cÅG'~ Xi3Åcp\mۧؠ7,V0N (E g_Lap9|]>pϻ攣KJ"z'u AY"seqm)[R 8XKd&"]6Zl"r͗YqmRn=!9?8SoD|nTY YeU24p]qjE@3SRZ;TKJ}I$x1Ů]8~m{.v;dbVPj|K1Qr\<)J)J)J)J)J)J)J)J)J)J)J)J)J)J)J)J)J)J)J)J)J)J)J;Tr܆7'_Lr7J.r}3/WPCD}"ZN;`_{QT 1_nby}198>^`rS4.T;?ܯ:n6]Cqe[q5{IPcs_ZOW\8Z647$W~1L~ ]2۾N 4k .Љ V=Z$˔ }R2}Bgw#Y K܋sn9󇏚Ao7g~xUd 5x%nZό?!6Sj抓j*uRmJH=Jk~%K\湸xtޥM7s]~ID'd|pSL˰gb`}2 ޥ,⍴eJY#ͧvS FYhYY|,gS*KU郏LN-a^{89[xLsѨX=d>+}$nq5sT8AAoƣ_v JHyw6,ddI&\ۅԋ{_K^hl˟o3f6&>[1}};ݵ``Mei@ %@dG Y`I|C^ZDCֈ'3H?H?H?H?H?H?H?H?H?H?H?H?H?H?H?H?H?H?H?H?H?H?H?:vJ 0ث_S}2 Vȣ_R{_E$‡1J(=j_˒>v|T:]g0l5O1v(S[ u~D0pyt̗8ތ]T _px_[̶l%k7_^siבT1.c%I7~r6"s},{WI_sVߍ4V+5W%;B/[q nc.˺W:u(FY䡒HNjJϋ>\}ϵBwFk}&w=`^׮a=z&21> ̬w+ {~ٖޯD>PgjS+Ǔi_T^'XZG\JS.Xɻ0:I7_nby@(;;}oW#u P^[]n Ś% ?W!( E+ !ۈ2ҭD-עۮwJ^PTǞcG󿟣}4M_f^7Fj̡Z8Ɲsa9Q+,P R:zlbHUkR"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U!R"U*| z+y:jD\?;');V3kNIj \OF ֞T$.E,oΌmwN}FO*xP*]CmRѝ2yOx=HM[Kmc|9st;J:cyhGYM-㛙2=~R킳}V˛MXfJ<̳k'o_0d!dmWo%NCoz>U] h:}41fxP]]RdJUO{ANlt&"@?d {BtCãdXzOv")RCvUt>W.+_~]\̝8=kݏi{ğeR3:`r QM uZ)@slxβL˞ VO7i]C Gl>M aj7 _jehgGL(.'."sMdR`R`.0tquqxx\x˫i_[ZZJ_2j6z0QmjLz~ Ms{rǔ,S[I[j6̂̎Fd6-E.^xgcdώƓιznF_!fܳIiNRlGlȾS<#UCo{E4DO? = @l8L(f"W92^uNTs絓2ISruGѦqo5%阀#8Vg3ՋlB]Ur>_Ή]y&}_mm`W[M)iEQ# 5}) INbFpmFp3s^8㻛~{|d㋴l6rC_A:KQ' 9' [UTdFlJRad+o=?gs2X:}t8]mea0j++품Q+E┮v) +l=tsf`]YJW'] ga; F=t%~-T5UIJW0]I?flN']Usɠ+X+VrJW1])/;fcOW\O%]Uk=tU4;LW׽N(]UBH_]޸܅d$4wW+ ~ۡϺWܿğ*Sk1Ae}Kן߽N2sW506"^)xXȎ9MG;MC5 e_߷Mo›XZD OL}Eb9̩Ze1pڭΖ:nmJp(.%sln9sJ- JP idp ^kfE4 |oD8$ pLL.C13R컉T3_(!B)2 *S)SWP]q RW`y8*L{uTҩרjU&X8u2;o}J0*!+$X1=!SWP]I¨RWH}0*+ԡL{uԦSWP])*Ht0.EqK B$-CY>IeVF:$n ET[,ba2BxIB T3f 9:N[\7wUMY[{nCTp . ~8L&pAvaOۃRPځdEn̔b &0$a%ݒwf5WonshJl{UcOHgtHI8\53 +[|ŵtcq,4  Ǚ-$%Lwшݖ'IUEK7pǥȒ悓y "8FRH*j#҅upBTA҇%ЬF43r:nhI 5A`騀^I&ZAx}PÎG#Jg]MZ-h>SZFff,U3a^ƄфR.[7a|eɵm mHAe,$ Zlbl~Cv[[e0\|4;V1H1zP6IxvBAg&vM80`3*h&DDz'H 2FiYm". pwh в5rִf|7~|)a8 8󓜙 fqWX4^=ꌖeYD$ H)&Mb"%!2Z 4=i<Z]zD`s(Dȵf lteǣ)˛e@I)%xBՠ,$D*(U(Ѣ DE񳲽 X12n'QpTFA' º{OLr\Qg"cGB l缶("@cRE?'uh1/ C[eLVǸ%`(L8ADm'H- RSXBbU 16g3KC Օ^C{?W SZsŗ="gqT1ūTs, S5e=U2*~G Д+̌povxv( ӔOȆ ݘwi+ERO*MxW`8EAyj]N߿jh2>)-Q *~ƽxuz&їSoaLuޤyv3⏕'T__̃H9HXsai?;-7ws;(BVv=<'v8ABz'{h5w3KyfpT(R=_'q9Z5 ؊C ^; ]]s w-H'|~[rCwYuloO߮w!oOF9Ұ:ɇNw&Z ^~F1$<=_TFR!1f@/\in\ lmE4|A˘-ébRH^YNO`Oi؟8Y[uH@;_5^Sʖ9/H:E,: 8b|prcZDDBTHŢpA;mgg=wdJԿW9>J!6(Tč:QC4f1>:<8L@HcHw!Tח[oTHk<!kE[0|<2a}$1{F6=s h;_q90KWWK7V`ܷ}[ay\-HFfC!=nᖕcT7pHVVK&皧.LD>FR@&@DF(X1p S\I'@$Er*NDӁ2]H1$CcØT&H`SG'DQÊ[#qX?Br4`W5.φ&5_G>:B!fqv6- ng5A, s܁w4epf3gZC2k4rr1f"U3iDBuH-K&YfEȦՎFWx㩎F&`o [y+`$S85rĜ]ή87%B>u: ̀}AgNALW_ j 憱éz(e`2B{L"]WXFW딊۾Z:{G66o{#+r3~ze20׋F~UJQ˼c!R-) Ǭ$$u,RO !%R󚷛zg)^''.6}r:͜SéhKþs<-eCɓ-}k>oZ E>R rJ=8oɭ)2eï%z^|U[<:3E;x"yKRL.)4k@PDh9a&*X:NyFWСFiMp+RUs.{HSèG!s1`_F X'(|oqT?B-!rB=b@,1 lP&@h3`QWoC}r_qO5S}GNmb7tr/BoíIzy↉)wLBA.zzÜk۱ٰ|䝋mv7)\^֢ wL NK4LPqr{e&'5r&^>pVx2 H& M|%7AM"*1F%5BXhovFɅI4D: N,`Sx-yZ8xg3ޥutt(Qr0j( ӜY1%cJ ^+Z/sOg^ek76|x}~w;l[{8uR~ JqtQ^֟-Y>廳F-I1:͸+ $ |HcW&0ʤ /iG6U Vi<#6"#! bƑO빝Gh3![3mZ8zȑ<$%Xh(ϕx Ld;Kz$${e ~]Z3^22[7lSȳ.SWBrHk޹" $F (̊hXѥXa_ғ,,$£۳ha 2Y˳ eZ ~1AxsJ#U!2kDfOu)(W@^^]}2z*bo0 ?n7vysAb7zO$Wz?7-dm~n &Ӿ܋k >ܨL򒅛36c9d][o#+yI6*؇ds6.d6K^e%\~)b[ڒ%ykdufW*=$bQd*!Վ ( pjt1?t64Tդ Dr5c? J]A'W8SSr)R!DezFDN)K/mι.됄*6nl-lPAz%z{n`)i4BVSȓK^g݁(iiIqأ#b;aE}(,#F;HU=w25Fbw>ρרqb\6JlcC \P;B{onOGܞr/=c P("8:UH r^;>.Q>W'?W>ϭ2K\H.U#욦܄ɇEU'wR}0_bM)؄E: YhYzYBM!2EAØcH348y 7!/>@)K$K Zt2XJc ;5x%zz|y~>)-qc_=3+@؎a⯜Pl.ϖcLTeIc%t \ QP6-sv'z|KQ#cM'{DK)Qhe)0d60Lx~}[>c qk˓A3?|I FYuTF$Yl2gFqɣpJ%W۠b7Qië1#f^wTKS=קUzo}ל|*no*G*𮦉[ճrmF:X֎>4|srj{>(5gzr̟z_R\xrt:DsqSקnRyUu~%Rfaaws}|s؞oGuxo+n>vO~[%<0O:I՘[po Z xCa#3Ҝ}@c䱻%KF.Hj??/f#:n52:W-`/=1]ЭqNm9? &kMޟC?\8JԑʘEBZe!%]\FL֐)¿9 m-7+MD80T0XCi*ŭ¥ku՝*J'Ӽ:^] d_q_s[xA%ujH*kEFUdֳiR6S.D:鞅lU )l6dkْsQT\. J.9msQ@ٕPwFkYw[ƇQNlagT[;[Kk쓫ҲgvpeqY=0'bwZ f!1O Ml϶'&/ =7%Qɺ.|;vz;0QV!(!eJ@)HFr0J`%B ȆЅ _:+1"Ckc%dTE9&dH Rb BS BZzM5״mjAkXN84"eYev(tyg3Kznrm&(Uc6ڒw7zsDؿ-Zt+%]/Iྕ<ĭ-ct%OiU QBI".H[q蓲ٛmv(vI䎫 h2ID'h2g%{EκaF˵kLړ#?o2=?&Mʐ&jЬnr)zW)Ѥv_Jr2A4Rw.ԐGG_)߼9],7xjYnCs}ku JK4?{0;i_#?Ox50mp3(9 _NGexza6oVEmU5䣓zp4hk֖Pn鬭ތfn ;VLXgb7mN9+[UV7j۽*Dy3@  k_z1ۯ2cy$e{v^.jX&=hvxN/>8_޽~{}?' %Z݅=^<߷>iUޢipi-:on}ݩC{֩>ˏq G_Y՟t3kW|ZAW5ף<ȋpYf[Oǃi~W^^<5!1Ӏ@[\y&11ֳw1[Sy)L/ZbqFx`}؟ FM,Wi __/=g)߷Sg%JġC3cK.Fjc% .D%#ީ{>7SUO ٬{Lrȉ3Ŭ]\rEz!3O@n37ދ~E1M.eORg2=I;ݓt}'mRԞG^oMs5kvoY3H\i8ig3ruBK+/`!{˜=QG\.l2J'Z)HzJ>Qg0ϪG}B-1pN$0/< dHKlːL:J&z|t{Y*q"vnr~w5Si0vq¸)@AH,#s8O4q~1⼜rt Q*tM1dQIm(RO2VzLr= r'4l//¼'ꎢxm{m"Q9쥷 o\`3fêVK0,^(Q%3%Y%aJ+o';GoLZJ>ʿkc s .7+mȿ+@mٰcr>Ic&D"eVR6>x@,|Cz+?3jI2; 3/TǛRx> Ul ̊dA* {á@P"7*$r YIY9nBqeo.h˙gR铍JMY]|J+c nN%pDJѧrrڒ8m?9Dޅّ 8 a:=[ԁ=uwH(~ ԁ!}a s&]wT:-E 6>Ԡhd(KF[џ|%r0 4;ֹ*}yQAE)`2R > 1\N׶{qwWWel[6(ukOjnw򓰰*+w6.+P.؅TT^%1azHjTg2qۥCqlZюbDVT;05᮵A Űt}\Y4cl" xIΩ+ _~{:(v^E]YIYyx`ۡZRR9cb1>N͙LGsʗXjߜL ae!nHbi0 D=C,]ɗTzLߊ- dj(N9hk1b XK d^H{ggu ĽĵLnq[_fG5zY:/Ho>lQ3|vth|<} 䰖ip\T>mR gzJ_]`ԗꛁyf7&@ /;tWWKiRjS͋DR"cزsԧج^v[e ֡}a C B-qZeD`CIy" '1#(dtRqVmj}m)bLrŅP@Te{Er ]Z7#gs5t@Cvj饮f 4|$lvsB? y~ >zJS[nrqYuE 6)j@] w ca&ө`ES߿Ǚ=$c:YRunvR6; bC ΍}A{,h[P_ }fІ<+AWK00G )S|-IMw5Dv*/?+^1Q1Zm}(JHC)S +$'+,}& G)^<wV׏ #yآ~|^[Wu߱~yn>Z#Z C lH# (1d:ԙ1j%+6g|u}{ճRiK^kS D(ePY2$@ f[ & iy*\8|U98^YsԻLI$+a D_[6 raJaK%qRLЯևbn?Z 6]e7H \UڞFx<%B7/o\1>e̤F@DBqRΎWrޚG1җo˵<$cM߱R>+2ft7Gv,Vl̄b$D5amɆ5޳w`˫osQ{o1I8ti0jwm7,wbOKT0kNni H%sPYW̳ɷ'R(dMZu6&j%cT?l yl rɁ5ۋk6r5PsK&O)'b&!1KIRLN1"WEk*6F Ɉ^5QzwtoDR!q2>,l|nz:`61-Yh :Fwt|ސӲGmtY{o0Q&Ej텨}EFdC%^(6sg" U\.m|w6_~Hݴѐnؽiey_mλTiQ6M̫sy?p썹W25ˏ^ϑMQ (;s1RaxTK`=<{H9y@=Ac ZI+!'!E jH-hZ 8ǯ/Vp|mӬ2#2[Чv!ś`-28ڣF/|5i0ؘ0vEx&(l`eȲ=#ڳ% %m&),)mQF=h %AV<*_(J2hud*(bc nFf ~g9 WCeJeZU&]UÑhbM|u{AK1NW%5yvogۛ=h(dfsV>[:'uNkޡ(`-IE`?Z X־y"WJ Q7C|6;Rjs !;6]DW')Tl}Q ]fw7߮<{n~Sqo]߬|s~\z͵s߂ǣ| ݶ7XzӹQXsnS[m A[̴4Jd+F[)z h*KRօC^TKzIޱ+1sg}Øe.ZIA+L0n6_ȶ+F*0R6HdaU׷1sN[rV7yv>}b_gteCW}:}OmoLǂIwhJ3ʔR2ڇc2gr In@DIZL7~)*HXnb=H[{8'<.(mEfiC=r{Q3>9NGp(vUPSU$nL\`CK'oHL\̲N jm 6k+97ڭ]*V=qrK(XXI֌٬ףWх8㹺04օׅ;~f Nn'8/{zg7tq8&o\cI Jfs*!6X2\F%P|m 6T/f%tJJ6uh!Y~ 6Lif R||VlijT/c=Bd@P:;P Ru<j`};N}/? -W|9&lq JDec`)X{`ᕨ\JI–~Y۹ÛIM7PF6ݾy-:k=K12d.VjV*a]1A8"uU p4ꪒF]1JCWWJzufN]1hcJ]UrgflCWWLWWP]9^ܫs`赮L 'ϟ?~@t)r5 DeIypA( ( Jʄd , K [Tv?Ǘw0&'bX|n3 "yŸ>9^G~[^:x,>bm{xL:RKEd]Av:J1!XE]_iQ:y|}1a_^7MԼ~ ΆfP73<j&8+T;Cl3\JʠT"LkY60܂$GAҶM]hF5y*xbLk0E;gнwJez;yajqYH?v3~Nt 8EJ'rʬ(ȸ`wt#gҷCg(SG<"=G];ؙJ"d&Mb4'-L`Sp+,%J@s2IDx@'R)҆l&ec$edAdhf݌-]=w|~0wYޞ%P)`G+ݤfK]p|~cybnmJHH(PR:gAQ"!PiSiC~ wb]z6[R]X5x]q^BH`NJA6UͅA㔡q)]ds{a} [5DJN@Fu))I XN@`q($: JBz܎}MRcugP :ttAh0jIJ[T %ZL/| =_A?e̖࣑!<.Na:~<|yRoQs>ŭG.4ػvE_n ~R43Ƀ4dv{Swͮ¿\__q=A 9b:풣Ym7ӀqtCBzu+'}^1V>fY~fA4Ou`zGFO9<^XT%?u&W=u{U~K'5J 6>6º̈́W93y*OƳ7z.E-fOppy/_wmm$WfGuɺ9c6Y6UW!(ip"*ܨ) hvW7*NVf~m_{󯿿&Oe'y%!|u/pirեmݗv]/c7\kvI\V8߬so/?Oܚ?H#]5%*h)կȰ-rC+ }}TɨQw * DH 1;?sSrJ;$+j$>DbS$-ꤔ fJykb M)1 % 5+ޟ$B9$IU|O3}j*tn2 Dn>УO[>SymQe^.?8Wtm;Ym;,/pnv<:}Z'^Kܡ9Ⴌ_i U9r)(Zm45Dqdt#,(6S䚊< 9 :LIjTm 8^s.( JvJjhs$ Wl"UU#fb`՟ XC 6X840kPzGv{r`{0Nxo9(d(IVmkaR&L;Ԕ̵W~4NxRYs1]}GZK*/ .V[jrwQ##{k35D_m~*Omi0ýO-]~zGzFC=^?=iS~ǑIsNXsﲍK?=dxik,y}y~5I lZ`^ӛG 6xDιez9hYsXd$ Rp٨Ic(ȗ >([{t8W?=m[;RfU##hoQ~ǶGTlPMULZ2C}-S$LLPY trL$}JXX1U*j&ւ:c\ŨjVق g?_]^^ T}'l+ƍ#m%u3ـy.jԽM:Y;* w\Eh,rйp P`k&W`)؄>뼏ZǐULmJyl.[zKC:0ttË}zBx$G鿩a"o\5@ɲL(=H ࢆVҝk:}-a6~aW?!vX+y%ێGw`~&Ƨg{Wg=C[C|ջsrZwO{</moe{s>SsuCw#'jdaL["m,ŝmͷ氿xÞcsEȈ( Lr|׃I9(DJ8iM.!.͋i^di~zSd5 qlj돣\H׍bmҎiC㬟X݇ OhIU?ZRq~7|:44sC]Kl*1:) r5s 1TbRźB2*y^\\74Y$vu6Fо颇b>k8M e -Wgm:=goE=zWZ9,bݞO՘BRt.\Kq\U0B[뫸 7uc̕H5#WQ0"f-j8ޞpv{((x/L}a}^}ὔ?!-|.,/;R~L.?M.χu`M磖*W٘jBK Wg޹ҷ1 Zl%24i}˥MP%S1HSk.U{;nMS\6;赭뵭`x c.dSC_ m̹LN@keȦjXVfH$BdAlK_Eeo p%/ @Sj1v݆΃ˏ1N/N#vOGG=bᨕ$l-Nm9SP|:,@Xek!v[vvXG{3VO4 T {!靲ЗϹ94u9e8ZZ=fw0eؔЎv =NӏweI~ؒ>36`.m CS"*,VQ!/)PXȬ/"hIoDaY*u=%뮇]V?H+EU^iގkg=Y?d*0|鍆-$yS hrσ[̀04AhR{羅jSLe,:uqVHibo.L.øk6̇!|T\1" Pa66+]mh;NJsdX Kt{pA+Zzb3~;dYPp.%̱b|w;[),6lr/%g"` .RDk O@ʠz\xk7K`RGTPJm |, gw1՛3kɛmӗ ؞cg-@ 1В)E 0÷8|$Xf+zZ; <RXk3j>LaͨLGQXmPXZ/}X! D0?q'#P +鿭JK|2*ԩ+V`|*QITWoP\L頫D"ZձDL,ގbQuB XJ䒓WZ]\%*9 +g,<q%TUJTlzJh)DF'#\ȩD-9zU,ޠR"yE>e |_F޷ˇfNJT<{VxMk6pZxxwJrEm<6:,#zdnF AzTxj96Wsr5\!Wsr5\!Wsr>\!Wsr5, s5\!Wsr5`s5\!WsD;߿G"Vu'[Sת<2ோ Cth]@ju1i{Xai}зWZв^eƫ;k+-A5GzgZm_pV;X?e:q…4 cr+lICz Z{1bCƑUmAsϘESӗU c+oCo0JgOU1KG8|zE ӣO˰%: 4lY.l:`|O| BV|ADT#*=.2yrQqJ|gFg:8WB1O碅4oN6oގiҗGֳL-!CM0j߮9;,Y9~#5)ޱWn@n+<ځFc,5&FW؈qRh\P&05{}miY~n_RI+ot r83β]twb? κ1='YRLR?I;ShyL,cHk"Qe)%Jh9NxQ >^,INvtӭ2@U.U FhF+R`3r񁋆oF-I)ǖRoR/~sA;4t:ń(͜wH2Y0-ya .fIY#;~1h/omk pW9'&&{"1ǤYaE=*i\(e鵰ZTAB\B(P؄ A2 ǨU1k(µzqGtFΚ)T3%}W +_pm/B$\'=Z/`tyZ:ve>qYPT1DBDFX {buhᘶY-C#] >4ڤuTx")9 (fE@*i%1#iH&V_qNJq/0Lnq7f]C:Z*Ձ#-)Zi@X#JKH-"N,Zkb/.su6VZEc)&:Fj$"*&Q3O O 2y#u^Rw8C8_`!{q~6 9;#Ye_g?1Im.H$eF_*[U.$}k5:5J5.|.`޿&ynriTW<"(b(\_&C; dKG#L9EgRPޝ 7vpVAޜ.gͿJ.! 'Ta!XpX1]_F3aR],_ޭIJ7UݻG#F`Ry3V+R=&S ᬸJ5H.T9LћSݺ7inE\ɓ?܏ςg˫R3"`Nw3Fd^8GiY06Q%`*|4<4 ]>ٿ鏗L0Z-dۨms%%-}hhHmtR%ȯKx F gr,u*N^;_>^|xO׏?\~ LpV`\ pl%AmM?"@-x6мafh9t3nSnHھޯ0?K\eu#<5A+'Е|#@6lP칺=JHQ!^՗c4&'%8Ub1NiDF_>_]}%AF=sRs'5Xgʫ_FۿV$ĩ&U &ei6B s!Ip SV=ę}[G\T[1YHp- /$}o \ 'Tܠ",G+z63k Bk/:*vKm=65tsYco'PƯ$WQGmz dQtKCCA$ݕpP u,Obم18..GS"` "+ʁo*os:iL´LLr`$ B($<3`OLj9+jh<%gK[^Ag*"';s,asϏ%A[l?kI,h~XW/$ 9gxY(t(90JU8.eklU<6y`^2o T csQZ8&ILU.\IBŜ~ i?%eBrp` |>2<T 9Ĭ$n1qR{Mg_Nu3I,QSU0%l beHCqfƙNJ3 r>Tj)VZ cҘՒ2,H2kc5E:=ȦQ ; ZDl ` :aFhNPƎqfg䬑"h"_zJ03'3|7DGke|_GڀWW1Ӣ*ޣÙL;Z<8CNXZ $pa"Թ(d#$V|ܠEx61%rtҝ1 q;-1T2Fv -BP!$hKvl: !f vyvܴ[o)0|6mk}aSI-\g ( #\2FyP.P@PNnb <*ÔUa'!Y0dG0IcH;CQ[Juy@aŒx#Uة`x,KjK4ŵ ?z9ʠ$;nzK-Ӣ`?NV`T`;F~9libe %8(oFW>Vѓ%k<8F%1((aMMf@"ЁiXn%Yt"y\7nIšE?4#8,9N~KRd8BDb ״h%=]ٮ?\G:-Z *6E]kv}@h";K]FϻE|_Ư\Gj"$!g$a ͸R{M#J)B4+31tt]r d?O[.|(xMKK6x- S3^현u$ZBQ}`δ^;4+o6 p{\oŨncܰ_IӲ@u=-3hۼu*.%5bc DD+˗z嚣4/ϗm!ڣ8߯zeՎ`EǪbiE"! c rd9HKM֨AA_pmza㹪G t<׉saNmebݚ\r}B+t/O(Uw9^v ̼q0S@l"ǞFu!XE4oCoeKf Qw{ó]M0r ABLC(p2֓1Dn{]zLg!mQƧTEOyx*qwz4mEPZDZKI8XyK+VyBi:„ vEF kNsfi6ʟULzYq/Yuw>ٍ7on0}W\䗷2N(\0s,^r" $cLK21)~m<2q@FE}wga|;k_8O/N^%3iaJ" LI#h`yZ^PC`#aMU`]l\Q v-Xhm 4wF-yQ9b0XJBWpmpzGy>k^j->k2`Wt,K%o欟2,s>>q~3KOၱbT2(Y8S-XHZNW%J'6JpfK*etPrҕ "Xt(s+ 6eB5UB5$VwtutbM+-pQ RBW1twĎS/.S*\~c,KRI$GiƍkAQ нa>x^â9-:|׶Q/\/EU4n#Wê#֋}jNE-2l gHi:VRNIN ,j㱰`qe٨ (CcL3OɌlfuO30oVNm=ǬjጼHبlE}ڡA7Q _pJpZV7 a%0i3ŕii0K O`SGWB[(R 4u{UO>26ЦI_O#ȈR4c05d(O+LrK?'5/jP> G=\uI6ocK 痼<2b/< ͠-uE }hac10kzBΫyHbAOCD.S!<$6u9?CuP+>ɍY~}ue= 1nQ*Pm 0VҡB 4ss;X{0ϗfwL2rw=K]5Jqn(rne=#2,"1rb Ăb-H^pܧy?4XfWGyM$3/V˯> `,L? (Jac1<܅MmWE.=vƼ]ɼWΊobIVx߾=qSHfɘѼ*˱XIw87Pge`F[O6@ˊY(7AEibrF_[3A6/4x9k4,}|9C^쮩Y׽0Px|_Ea"/d4|=FEh<(я=]Ǿpʑ>݂7=0=ž=,{ :exXH1'}3 kײvT}% U:o1ټG96pF?(̓ǴDWЩ9+ԜI͞Ç1Ίɮѩ𭞮^*w*,kDJz{nm_`B8XҞJL0KxA[ FRg(תLGj0;cƌZFL&Z iY v T&$v=g.ޮ.Z_~CsbJv3imvK]xqY'!~3ȍЊ} .7I\u*.b֫P:eSlRzs0bK2UAnͲB(Al*Wix_ռ(n+%7(77t}uuø]_,i[*:ygRaM4gݟ\ \]-^ƪܱڌwq  uxԤTuoδ0oo,o^Cҫ[goM6~X}wZޙH9oC݃1޿nMI-=wp-/A+ÔNHg!f6>ssdȡ19̱haއ5)QX^rK<hX] JQ</ڣt.9vPVl=rGHj @{xRku&H ,]QQ8n$*Ű[)r5;k}Q0Zv.\0;heV,~xضr1S^[})|{Q[VV~U1V[RYEH| Y΋LYAP*"ՉP2zER*&@=)Q[M*Ffd1g&I͌Όpamq(ꚹPv\\!'VC|x.xPI Ӡ_pƶ;"@FEmԆQn+:ļ!8;* C'ZK(IItYMes) Hu`ZD"rRXmMaڤ@֦^֞ױvxfθ`NڗfP\$^9Ka&9'\G jM|ƁXȠbVA ('@qƂ&հ(M|X|Xc>$bl #&2"iLj#ncʈZad`(QHLd$&:łKҲd0I `vV"4alg5} Qgl$4{Z E­xٲMS]܉BGxa9,PW;3ٔ?a/{ A(gzˇay~77C+Cg+.0ξzMq;u"YGh*ٙEǸnqsT"6]`f=rZaʇrC9y+iD,_tMK)\F$VL2JwmH_a&W;\WY'9.lu𔸢H(W_c8HCQPq 5W5@[)cS @A5B <M/ı|u˱[p&*eƦqkSKhSdSC Q6촜 ܳ"+NBakͬ;-H.HGI)<1E[EfNEˎh*&^+mP,}qcHiu xmROS{mr69Îp,WjX_ĘTD,4|IFtPl-l6v!iT D8.DxP2#@'ηaxY_&1%h77:b|TsYKH}z+EhJGpF1Ĥ5zьd+\,s L 1U!RD3q*e(#*GK/UΒE7c1rݎrE꾀l+Y:0%Z9{$O5Enx`%ˢR^lf-o vT79k-gI&`6Uw u_-J%#Kp r4{s h ox[7lVT7lΈ8+rsaGvTsB0LJa OZ:EX {Ʊ:4ǫ5 g#Sz s(F^)W |ȥF-)4*AW9Fr-ǸkŔlZh>xE @`(,c )T4Ԯ i?G!qKx@woJ_eKCp3٪9Υ!A$;L4(v(T0rDDB. )B31RHp;m<8-RE!2aZ%.(CS0$4`9h.AguӞ9)Vh)f9Ww&{~YPX.\, -[|gMn 1$evO:ŤILD&.JL)x`!!RhEB!J.(#qOǒqڿ%ڝ+ 뜰nWԙ(Q <&&#>F*EDyB(c @-}ϵ5m"HFa!2D&D+ Qc\RÒKcȎLRBVmR4 )}&3A^ ىK C <9l_n09ja}@k.H8 >gU1aDM&UGRэ۳|3s#򨫌0,{<|yΒ1G NFc F 儝ܐ %96Oh]d4F%x}{MԒDDp'ol̉nhJǕrʔpBJ_g.XtTݙoEqo'oݎ/̃H tF̵Odn۹"dϷͥ& ꞮafYޣ : g0btG}v5CV+#{׺2z?|2;>?eW SzB;Ʌ=UtMyƺ\";ޟ?;ǻߞs9;u307Z$@OZ457kxkAu5k}(>oA ?t~~;ٷnݯw&Z N~F1 uGiq2SCB4b@ݹpyBX1Hhi-|e̖T1H$NhVSG)}~5 ?&cQGE ZsaoHuuYuqJ+4"$E-$v:("֨{ӑ~.JQNHQlQh ,uٝИVBt2fp2#!%T#P89\ǍV4ZcV~^EyAtiQtlCJ nq&?jim GmԴHomO ճk57k ŀH߼3&H|2(b fvOn2 r e\Ė8rn{.vz\{y;B%9t\jAF_{?ܯA4YvwB"|^o3kuW+6څPm83F i-,OO??VL8B 3[HJ¿E#M=?69[i@"KSy1* ŗ&:"]\QB8|:P[i.4L9'O2usƏ՟qC $.Ed.kh.V#gzBh!0O? o#[c]h5&+ϗV5nPy߶́!@]GhHZcPn"1Z13_gN,MoscLsмiJ)Jbw,D%ed(XBJ.ѾM1ymüVnJvPy*[@Gȏ7{_tB/{ϙC\L=oN 26E&򼌔V5D2//KlCQ1L$@!xL.~BߌD @+%%AZjSsOx,6Rw뛩j(§EK̊T!$]A%aNEɾrFU<(p|ZSzZi6j.jyDBUգzë4b׃OWӺ$ڢuIxRz.4B_l#C֡E--ʺ5lb|4kdgsѫ5$[G! WzV7SoV4chCjw~wL1Iy.hZHoSv?UnYg|[n9fgYţGwf[]G6ѣlgݚvmaēy397фcl{1C2 9]ǗwIkhQ1,A\\@ 匀I@%&qL(JZYh#9gCj+P<ǝ1 cDV@i\/ʃ.%b4K$ *F^@.Jolq1k_ 6 ^Wm+/J9~ai6|aEm>lK`3K4}jmP_ӌ >,pAZ Ubl6P|uo,Eu%l@ H"|eqr6m/AoY~q%q^~KGo(r E -";st@'CĒ1yN `CxTbc$ϦSWR|v n*mw?/o.7ΤɯۋntO;7cvMvy˻D'\1ϔT#qłJ!~~NkGEznI/'<#[OCBA%û s&`-NX$u$)( fhQ8ř-XD YR*G&In729;@' ;nKl-jc`y;Cd M=փ㕤c^ :0a ;B.dj90e`a8 h<)azCk.1>B&k.ӏAYc}6:k;83 ~Bg$$b0RӁ.ƙTX ̕ AuJ 1ڬ+/j*Ns\nZ^c!9ߪ[U4Ou;v(^H4+m&:C (M,Z"di{CuF$B NCIpt\e;s%xNEoy6 scPQKOB#%9%Yh4kTV=k$m+:G8$f;& c;& cCG3t$@;{w߽w~;{w߽,,Iͫ7f_s1 Qr% F)騍aNkY]f3LIkJ*E#tx Dz#*- 'xb eHч8rAh89̉LPz6Aʸ9;ZS|Jhs|X0M]1n?oMښ+wp&x7M!6v6e.УI;f.Dn$m\' y2_RįOnL<<-pe=n\vy>jǛ_BqejǛ_y|]x,L" <]E.dm-9d|瘶M*BgvjT_:lsm5g}~nlqqMLA$<1אO,"N9ߙɾCge8<\!@ vuŒMF@^q0~^5m/X|:nt_ޫғ3 ICb) n=g}<%4 dNTu'Ӂk"l.o <"#APJ!ǵ.xyƜwOd~o/vocOL+o͉#INnlG V0 SFT)+fF( (QK3 ^D[S!O[TR֤HAC-= |&%U[3V#gf܌R^EV}u M  W&yw?,ş`05v(YbIg46 < 1"Cjٮ[~ *ee(;ič6QEkĐI8mc"{< .9o5' jDg9cƨ@r2虐B57X`'j7ϩEv估):qɾzT֋ӋWg}ty|gǷ)ʫ {>[*LQ"؝QBPk{?H ɮAZe<4\-xCwYрv*?ƀKFk\DT L DR#ٔSM̑%B#&.%~a.RA+ۦ@J1{BVռ 3t&Pd+ *z^#uWP|uD5/ThٿV7gs.ҼX&wΐI+(X;km̂%} 6Fv4T 1qui6TSz_wJ]$;JG} W饠Җa 3`pF5^ܰV\s`˷+kI-ؗqX+ S$vu-x=̷`c!?Я/Pҏ m]\Wo~yZOn`[``tfDz2w1.G)s$Qi1[YIK# 8"S{#ö$4 :,u"I(Q2j´,{ܐD'A@MIݔ< U["ƽ!NTA'+ºS݇\ #ed';&ʗ.HDLd<@ `s;Ǝۏ5"H')tZ,sXH sE ¦P p#]G }Ru6. Az΄83! Ndc*Q=B/'_4/1 *>`o,<.}B_4u#~'kxNJG'Cb4I|.7a~[|0_e~=eԒ3 #O2Ę,q? n()&8㒉[vqi& >_'o.pٵV3$c.p:Sϒ+0nAnL ҏf=|?.ᤐϹ}:{C|h=uosַdJ7ە6T3D%."tM9G+vi*ywT9`~2W~{+=V5|àx^^;so|݇DΟh&icXK}1 zw~Ԫ4|jiaԊU}ق2oG^3C;b֠ ~a[KyVt{SwMtu=L4 Mb~3M{-UJ쑪xZb C.qgvc&-kz-n(FP?ڬ*_"4UNYsI`A%N(3q뼦/uW;ӥNf)ؐو6sǘ2sFi.) !xs;ةr3]{]oj]x{xCk#VfվRX(b. <|s,ݱWlI:&MItA/57SZzK#(e$"^a9? o$rPӴ kJ+-rJ?2؞bW1:nL/ yf4lO/,9&dҒqF 7hUa`am*b{R +|z \>q9)+sKϣ⻕Z~4aK_JΆ A HNh-Svd6X*K2YZ&2\65əscy6;h;LMyؽYR4Dh&w\q(Ch@рdqFA^Ug|a|jkFޫ ~@eycgf+ +A*+sT:ƣT[bҬohcUk.ӵTF9"sMPRmԈ"<9,rT1yeL .AB霹,ǸV9kdNΪ5}k}{c!\oV қ-2a+A=Rlc wu Y\$}৭X<ًoղ%[eʖ<bbvUe\1h|w! hV>#Lv%ƹD,JS ޛ3OZR)<TUc xU/!E&ܓ3jрI)@V!)$E sE$C7稆y:ܼ[kG \LL:3E^oGeoÈ%ԔzfYuu6m~2DȈ͚<60p.(jAbV~8SřL@ )1\;9DAڦ`FDŠp&Yj dj oS 4r+lT$mVh29z-oqWP+:|"!x{A ( t@ u o- z=~^ͪd9"yRͲ>]Orbw7]IզTw;[>*/UH1I5O@oFfBVSLgP]O멞yF=g5֌$'PȜGH3֗&r#|['S=Oϓl[Iboĉ51|cq^1G/hN!!T[JJdVA("/YТg?J Գ@Qif7rOAK%x^]o <Ǿ,޹G#tĤ E!oÅ ZkJ ,:z*x+ƷޑseYfE=eG=8HJ͏s*ls1&i:0YξYSߓHq YfJd&YL??~K?6*= ?^TNYhqy>z.k6YƶYzme:yM٘p>/9Rei,NNxzM34 ]tjYqO ԁ}z}miz[h[72߈_1}==j(z|wOfHlV 1z徎$SR߁9~h2Z.p#(7j"lκ41V N8l>]4󸸽y+vZ66N_ާX$\ߴ%k}-X:-3M#io6,^^K`kqfp;AڴY}ii\c60'IAqd ]b">^-(N0́;b-N0́cO.4*pjƔcHSb] q  A_3xL8(f多vSO{8/)fu"C]I.etUUIW%]tUUIW%]uVUIW%]tt ?p/UjA 5FpFBcךZ V¹CziKqn#+a]`" \Y2#2DsQq1AXgI#uU Wo;.4K Fmѻ R)mM#Uqh9g+FњWQ*U5;TVDeQ0@|:+ƽS|J0E(HkaS=1 #yNRA{ |`)-16:z&|BID ȹ53!:%58Ajy|TF򧉛 m!#I醓p1GP̊3+O;O!B'=ϡ)}G%|Y,5=-ޥ2 I['4tqVF`Ύ{zW];ooO[Bg ߼9F_ʟ{Q 8Gף{{5kFfRȵ.u$[y~epJO;ϑvU\/u ^in' Ǵy0s_LwHރ%R:udH}ru"8R"FXLZɣc8A+tќQ18Hh3RD, OWqQߦִ7JRs"hv*abQB/)+")s"mksIG\XA ^9yHS1MH([Mskk 'T!+ M*G1zI!]uWLjT.ܛb,M1,|.WzZOn:RJ=íwcڪF`£B+mJ£ZDj  zE(g逫d#@6\x3bL ˨Ќyf:±@Wd#s ˊ O"} mb={\gJ<껛\/:SgJ㍴λȔF,$/q^Pbg042!ҽV7  4Ƃ[@>jMrHȒThA)JJ+p^wg"kgj]=;xEeo,G7s'VPY-q6$oqSY>lWv^ P1"C:H™iR)&Zy#e 7Wٞ"=^CΉwe'Ju{rot&X@M4){?wg__4`$>hWr6,lk:~ۋFְ*Q2E;2IuѣlYm [%?EyCGĭEB nx=ˇ|XJk[yy}S鿴e+Z V_`)Y&7Kzg4wlv!Aha8:f6C<t~ÿzRgj׷c<\\|N`e94 2O}F1.@edth)qOQlsKh$BbdWA`W|||OUhOQ<3qvːcZ 'LR"'Hy˄j"ZsD @fF UOkB2N 8 % TiXLq+e+XWºPU]q*!0J- 8Q^op/_v#E-'(IN  E\2Q!Ċ8@PNqb(l`7dg#9YV\*/Il= JKc"^QإYݠfhbcMiY.m}oł x&[_ĂiP*8%0IYST&!h#Q@!DўG&D%L{#(:h"IY: 8a 'Y1F,6>ՈFU#V(1^+ g8u.JwRXfL0VKsٱhJT8 $( Xz5D%UfKq,vlg2xr)F&te.* '1 3tF'Q/P` F6F 86%"0 ~H`H( \LtI>կϬ.K .Յ_09CM9fIT8Bf% AK&1K:Nu6:I)' CLhy:%5,4D`Q$mb0!V .gC0vUF7Ds٘="3n)c5Iis֟d[z}0F?].#) xOH۴w~CvԽ;7s*q/a~ ˳2 `EysЯt69C^E~kI&85ڑ[ggwFg7#mu=dYa:^dԷ팣ZǟNlڹ,jI-xEͰX`1Q؃J>n&=m{ֽ|F[*#:E.WM[歎8#a /}5D:#`rO*E%.w|їO~x_㗯GO/qe0.A?[>KC!&Ms ˛];|vvO\kk;8ꯏ;yfkMUQq7]Y]C'Lb4?;Vܕ0r+? hCg, ɻq (F& V%E: ͸q(:dVji&B*}iDhUkc\g3+J"ƨ50@(IdBQp2#!uc0^^J6_1V6Ez[&qk*/ks90m'IZ!+R]m8^6J4q'ADιkR[mBnM)TxUz1:։)VYLkX#ur3E8\Z Θ !{MQO׌6znK!,xv,-k=boj"gqڗ(tk"g)5`MdCsJƀO?/~|waL8&R7-ogAuF aõr^`7<|pȒ"gAEq %*zg$H"\$iuR^C\B$u|0䳗N[!\.EdơkhfvLrqdzBh!0O]|Pݴ fOy9v7 -Ysޘْ;iɦN̈́g\)[o\A-Ic mKFzVDejyOWue{+JR2levʲ9K[U t_LaF;Afh#a d6 O,e6ʜLIA28%Údχ󎏠~̠fKHohz6e#;l28p]-";}w#xxCD4:3^sTh] r=fen8Eދz 睿o_~x 7:J_??Xl`Nֈx39<\;9++M-x!#7^ IѬ̊]h !Plti*;`l~8XC!p =N"ky ynN·Zi6: g{ЙqW[Gxr\0B) 2YdBSu(yG(r"n8ƑesQ^n+TF3Vswv`:YP _閐6dN^n 8nj36F{oh̡ՒIjǽhT%b%ے|BLڷ'Zϙ!hcRͼKNq% 'pHR(Ecڹ\$@N%eL*$ <#:j>12%tL{1qGgn4kLA)G5,4j5ٻQB}'cgiVǖYjbndTS>ո͈fdmzy]"df|M2\7@E,wMccockM7̟۪Ko/O{R2XTKʂ1+ KEI8˒Kh ޵57$U=`^Omɾf+ h+#K^IPjٲƭKeK^7&dbjgU霜!ڔ.гF>&㰿тay\Zō;W^KB~ {VAV)"(0/yT!:br}F3xx !e" /Va7uS̔b.BJe"hIe(t By˕Oe g}T-:dXPC:I)mٍ Od%}n}O.40ZB [ \ ES#N *jE OxfD_CsP2,(z+)e$' \&Ot Mffv/|ϰ8)Wl4Qq9+q)ܔgH;`ZN샛9 ϔHm9{x(AVOn?Bj́ m췭܎w_§8Ntlk=|y葏J5ׁ bQkQ$}8z+-Z4MP1ZNWyeER׎89'us uuon_`?'1uyO"(v=nIxv-k/w2m_JcG בo+{ =EaY- f NhvGbّƲ|ޒa`2ݝy mQkm>LMhsna>~lEu۵x) i=u®18m)ŚM`"QԉFɼSee4I ^^묂!pJFBY&H3Ȥ)^RTEp2O:Q#hkCoF¢oET:pzf'䄀!)@$F#3b,E|)3s FaЫzcmY+Tq5x6fGӹXU=9T^xXuv R,ԡH;I0`M{ZC/gn~ڬYhrD dhaa OV ; |)`StiFR`KwxY)S\Bj!{$'RKIXNI wOܐ3wc?K/ u=p|qm xCQDI"YcY?2˒$b;1 A+ ExHi >tAX+inƽ2z' plO AR>6Z.kkT* noQZFK4~ԏ浇x>~/T+-* X19q&ؔR1]Ldrdڳu~LHW!3eDma, B[;}~[Pnv>ڑL}3ߓٴE ԣdmmc}r ԣLߒWHf5GC i?/hPR(fW6gQb3( e.$7$őWKݶkAl^z15n:"83_.I!6,PFdeEKj2!{;$%2\w&q&t<(tQr ,CskHOB#%tKzl󵫕KOb,ɲ 04aj:”Jv-alw)#FJf2ˣ,DG(oHyJȚҰ)ekpt }V/*D9г~MIR2&H'iK%0%d 2zji=Y: `;!EJ!d'g39qɆ`l\m:G:Ͽ=]k[ϋ/ğuq~Vmy4=nHgͶ«`Iںj^]v8iп?m#^0cSn֣+V80DgMvh^ sWMospՎyҶ3 l ܺZg.ųUYfa1!ǔ!j\oj\ʳe wr%j&?md],?&+i==kȀ7tpRHLd? tHFj9wj ]e,\x!;j٭/-_|[6Zdu2^,?8bԚLL:k,I)d-XH2L%h]=b$ AHQ`S n-\=fZzk1h2bWai2./qGOԖ.jK>xSG ˩%ĘE(n/Nr.TIJ*.[RkaH:Wd!fYQ%IȢ( R >E&$saM0M)/"ʈ(Dqov`& rE&0({= K΃X:5VED1cƨ.eJ3! n%T^(+=9QM0"$_g5.鋋2./U]A(*F7_)J)u;W)6VdɶRKQ%/j%nF %Rz9(r;SL:h$H'K1{oFV^*8f 1Lʖd+:z>TU~|զhwвܬ^̵T9D6G J6 ]R1b X9#֋LUuM,Jh]}l۟qUo ;ٮN)%"DOTU4@6ʳD2}k*ZJطn)a߬D@HQ 2LɠY@!ťt,pg rN9놖iͅG/&`,]Vd u,,HdUamn.I3Ը%]x۴ zwWgrx.o>ݫk EdØB5P<2l9O儾tKgP_挆yHp"OAxØȒ4is" ÑY?O+ׁ71.̵ֻi7LRR#D,1H_Bt^Bg ,|NC QwJǻ#n שFr=RD۾162f4_'1nP2Oщeq!xlKtBπ< yJiYQe.go%.FE3׍vkng[{(h] bG 5בV bQkQ$}z+-Z4MP1ZNWlˊ~m;Fx?{۞1qs_s 8 O y#F7,:(=).;Ⱥ-&Y%*REBj~)!++vRu{Z.pTkܶ|ڟt\yNݚqցvv;au7yx[Q{vmyJmMtT#<18m)ŚMޝLoGvܐ8w,2^ .7]f6 gd e ^ zJbetFx" ތ*NŅENjtU'nbdTxsRkQS!g@d)1 5 W |,pRVKi\~^]t=]Ų~c $У 3m{?M?j] 4zӋώve`G{Ӧ7`XJ*Œh#u&]H.ķfoѩG[\h fi9ByX:Z'B&Q1&(MIeY m,sx-:A3 `%<*V5˟ESӋ} yzA?Zb_?Jf5nCPw#z%5MWb@K|`_ndFFlҠLF\r(Ch@рdqFA^UױR;}J_ݮlK8H>ʪ}Vߵ rk5,Wd@c<ˀsI3dɩMB _=ony\Y!*PͿ_b.U{aQwބ+=-85-ron^cp=B'|u)Hü}E[`C-!wimQN[i^ RY!wiq?hl 9rE0Vy @d *8X K0շ2颫t 8ci],Fbhm&LG\* !d,8Ƶ̩YYk9@\Pߓ^e^~?+5dͣǂf˶v;:PO1Τ`}!(? ']{xVQڙp9nr7p:(SI[>J4Kq F}1@D*Eå{UEy%q`C)Y/r{1ptʃN8z<|%%j\zWa6o~G[#i ` lHQ51,t8Ù{8e/\٬6pUt[t: {AvJ0LZ89to7.'7s"I)Z(#٤\w^1߆Gp[SDoDGi}i(8J7[Ǐ,n:z%t$5Zr()>D.Ғ-9Jٵ.[r(B_%Ǔϓ'5h~PKniOɚ-yPXMΚbGGG GHL&g:PQb1p3W^)+e.hN1p&d\246۟֫o5$[zl~OK}r-mD}5^K1-q/qΈ:dBH!ci@5 L7N\;G ޢS[r@f!TғP|e Fv ^9 OWܳs#La«Iq.8j?PUI偶FX+,4ѰL&O"6h^:6h^,6Ц($)%) 4 ֢t^#֖sk"r5rĮ\JS-tLXK>M_9qsrxpM>}w,Di`!C45%#"3 ]"&3L\&ǔT 1y ZG$Mg9EFF/UuDU@P12Cpgf)9̉ީM2FΎNn>&dv[ ~f`oý̧2hv:<:lζ-»_zx;iktp'Dn$m\*(& }/泥g%G𖳛06xj~1w._v~q'BeF~(~כ>y;p3#]|˅gnρZzh k~.v@]XIE/j6Jdu)fOS b8n>)'2}5cgTw&iKwl&z.YUk >:85y.gYk5%5iwc5>f{UD![yH,g}<%4 dCj3N0$ETJ]4zyD$7FH%rc 5k]Db詛fUceezIEjyQx$֩4;Ȁ*&aʈ*eb%%j)u&)59+)K.H`,L.1ٜ$1sgRR5c5rk(Uta5WʺtQgrWISB@]qIo:HW2@hoiMS38OZtB c5\w[:lW4AfNfF!SQp6 ЍGmhu/9P]{֞U_ZzO|"_1X-2ƋL-5.6v|ܜTnMXͪt$hM:Z˔%@0W@@Fަ|Z8;9S7i`<.qQ|_dǿ6S#uS4aLΨ BB̀{yОsힳ]65 z2_#[kLR2b㳎AXd'$l`J  0L*YKSQ E{D.K(3u( uB{Ⱥ99jW#gGiG]׽̧%?f aFq]zņ~n 5mK2G@l3f032BDV.$NȔȰ- |EBXdgIFaTrQ IzR 0#q_+-C^yе #aݩNd.ixpH慑2DȸcRX{2Gs cA]I$.p^ke MlHARnOfe(#Hoәg&܉sL%X2>I֯iU"ڱ _o7 ?wg09{aįd_\7oqIzg>u9Gy{~2?2_j UI5.x8㖌b3.8egFK|<1ބxBJ o5CBp91 K{IE\Ʌq r-Uf `R\X }޷л"o뵚GT3D%,tr~UV&NG;+.KчP0=f_[wߦG~hnr928[A_ rr}^%Gw;`lݝY{'Pt6bm2Bp0lu`YŲB?j`:uw]h [u\ 8rcE_Q,ʽxo8@|E pQM<Ůýxu{I/O?|]Ys9+ Ɩ#quD?v{;bz) )H/a[T *QH| d~?ww?:O6z\^w0d ?[RMo5 EG>u[t}:}$Xm߯ g}~|-2ͽCU^]\Y~i~= o|X=0 cClӸᑏGG;.&2p;J : ͸͙?VjiyE IR!>ZIu"PDqQ5v\Gى > -vEtNDANj%$N',3RvͱӉ]wP6=9LS(GL pUa:L 2/>})g/{!'c\Q.m9V"; $p]ؐOlb vkFƂç]ly?H$FgeX^3 ʍ2 "ur3:8\Z ~1AxG[2]@fcP/xwWal]<5 }[lO"8 ?ylW?y7>;M/V+?\] w\w_3`m:Ղv.^b[ց!r؊IGh"uqfI S]4[sm<- %E+)3C &N5z6 &eF2 ,Z#ՁHO4ZNb 𨟷Lm9#VGm?(:%8ɣA0LJAQx?+6p/3PB[W#נǭ _: `ƅAe]/qN}ޠiz17@$ihP,I.MҚ+KnuFs(Y ^`%)FHu%8*Gm9>X\8 9y 1/Bc֏cm`NqqWبFL@m8ꏇSIzy*̐Ng(A +HTȝ "锨ăCx gI|q/qM1f3 ɻ}yC|~,H}l(oBA=*VdJeq>i(-p%ӲԒi`4*O xDWR/Ŀn glP4DxJ%sBv51BŎ1/X:d) @G1\)^iXD$+S\IkI9DIhX H©I@e$gSG'LIEx*H\La!h9 pqNXKM^rqs7t6du@?Q~ >G| XIIYq9̈́|~e6m~DȈ͕ 6pI 's ܁w4X++ά8Xq%>$6B9(b@E.7(:G()뤳Z"dẶ&YfEȦՎFWx㩎F&` }Zy+`$SaYL5R/'v X>ΌQau׫ UUjWWѵ4j>/I` ͂hI|X;AFxIsvGgP{RX jIY48f%!BH= R'\JES`kjNtmxֱ MU-gn>N-7{,YPg}aS Cg*N(.SyJnM-h:/#1:A Q-N,BY0TðGÐrr8KC$5xA\oJ:~D* 6_J@]`ϜwSWqH_2u+gہ-wXwHW>YC}T^b.pIX  D7 7Ѩ o/C5=Zfy8c)j9ID=$a$s!X_F X'Z"'u믲_BJɀexs{vF{Zw(b0v߼8pǽ}?s 1ϓH+4 27(M0~ʁWF.K OKcwijE 8Y9AhVg|5׃wEՆuzd_~{t -4-_7bDy{d ]G-=ʮ5c(Ɍ*93>#]I f^VWH͛>Y0AgBXB Cwqp}L s qeF_F:wąO4w KM;!EhݸekD[/G+oNfNžkp"~q~z%mZ(Vtmz1G&jG.ze,-no8ٺSvbiB% ̖n/_x:ף^r4u/)iނ,hIEBIk" u7ȦȢoˋ)tU0<]( c9%#Hy"Ng]Jh8IZ+!:."F/kOQU2*U2Oھ槊q۪N8~(Y50YnjJ>CZD}ǿhv3ɘo-Qn++"@nu@D㢶 @8iȥJlOw^MnSE751kB2jǕ[[cWS!\0VXOXR!YF\#BxWgVcJe ib.lWJvR+NJ8Q<`GYG6^# #V5ӀK~ V&£J-x$ԂQP "^J&8B1#DAFmsIy9(8&H7B`i(a%2EOhػi5DˣKiǙ:4H뼋\yH,VV 2(џr)8a-6"MP*FVG#&9tȒLhB($'UY=0mpKv;|į+s z\&mW H\أ9F\+H0nzr3u6͗po:섹kt"n8jkVa̮]zps;M\HW&>rh2,]sc3Z(P Sx@酹u3i]0t|Χ$q峝v}C ӞgZ.ׇdVA?Vvvl3 [O>"\6 Q['P-ql1Sϭ͍~8h0M@|vt~0L(UZkx5="qjWg;z{QdF_Ѧ l#0%3Uv% pR-S vgjw㺑=aȢ3 g!d{-ɲWW+nI}.oթsxY$}?Z}a7B}ʠb,8bW6϶8kSiuJ[{e~薗wKalx瞠m_׫^}'_ܨwOn\xs#WO';"+:gڟ{?w͏'Ɠ]B}{qOɵ?>zv϶zv7>:zTjSsi#Mk/Z{Nqao3gYX^K]VF]u5PG6".j4^&׳.#Ƶg({&[ݳtcu?g_QVM*!O=%.%{ϳcj:}w) BaWg_v)HYL(ƺu˵Fv'\ҹ0@C`q6JqE`NӫUU|˽cGD*yf<Όiոpx.0}|tn*Mvo~/uM>;;g2Ba)mxԼV4ea=OuPNZ,}C0 Ok>8% ƅQp5 ׂ8c7^-:L6ӑ V\{4WjY-$E{mMyM#B֬ʇCCʐ.r] {rфFRYpkoĸ/Z|<yeF##Vʽ4ES$(B"-B9hve U19w4B$nQ)ך_O{l_{P_=>7M:^ÇhQfY53#.128>c\&y꾁RWMc{طf9Fcr8nCi4h(Fc s8 `v`j¥t(t5Jw(;+q`EWC6}_?78>w胄-u}1]xyK'9D ܧh➚C%hrDԐ(k3e! 3p@ϋw O+ͷY'7@q?9~zbjfjS0nYB$pW:T`B̂%^y'E%JR?tuWr#Su.-IQ"pKt%eeߢ(e,;4;YoK.BJ$ !rsW1|mg^o$#r}½#[lRϸH%w`V{ě5Ѯq Ľh1)<5@Q-!uD.JC; r%i'ݹr'cZP$3n>sh cå1Z/UF-9[ πHDY7KМwJkMi:"4gu,M!:ϞJU`K~B fvhdrncX5%\sp>fx9J1D!8~GF݁F4$$~,%;++5PY(/`T+ER g4>J=鼹fVU!\sQsu#5m78!Fg@xlԑ)ٚ|Xz"Bdʝ%[OZe8: G4o!9ㆌg>&] PR"@k,R!J`+C^S$J-&`OWkeb6Cn]ɳypl,%6Hu=Ԡ +!wB&͐jPo]ki* e C(Ɛ@ %H C|x[TPAx`ja;R1JXkN ubBp nW(u+@e4(@@0Jo[UT gQ(Q@HseԸEp5( g Q^UuAT">4QPWF+ErI.JDsrd& Y*>$@H4=:aQCk?v6h!M^2G ^-kX4ğ$'曎1ND0BKS;v0.7̝E/x VA?&O`P'kwYJ6r V|A)V ae,#z<%\._qVU6 (T.%@_1~:k ) 42"uBqámDM*Z,U5 9Cb84EFa?=3y o k*dGou^Ip"XH$I'v> 44st62؈!|-n9@Mn1~˹a:59D)?Y[NU \a#r5wEsE;xC*y\SAVyR=V֭z dP >&x@: OE#'oSU'y-VDy'@ 2Px?mQHR(|4#ʈU/p`\sMi㊯Mpvm10WdnqHZAN%_{,b2ͳ'jC%K< u~9(mxK~N#6UG#.Q!@ /u1` v,`-4o50zs!}~Z)A(@.hU>ja՘-2d;]jiBD!X\2ԤR+ C8/S,]0pfd Jkb>]~FOW{'B#,8$K2ތKq՝:ō 7PyccTF|R=lzyl{~绕 >6)&F$ Jf7nWf-3?߹74O쩘 M-> pЯ4nГЪӯ4~'ѱ}@e= fЗ z9}@>cбt:{Wl K_n*`#@L?Y\;VXFig,RlWkJCi4m),hF7q@B8 !q@B8 !q@B8 !q@B8 }8 I #@\ʟ Rq@_ xauEϏr0;6iTKY |=ώzf҃qtuY붜kg6~)m42&讥\1[ޫc!9{*%luȩ/Dx! ^B"/Dx! ^B"/Dx! ^B"/Dx! ^B"/Dx! ^B"/Dx! ^Z^febXLy\Hy?}Odp+}G+`6}έ9*Z7E+g 0XkۓK \;I 3:Af'C8S,"^//U5^uBi* M Dѻ X>x~߯߾;z N|}s33p=ԍ| D:{6.Jk @,XCl;%D\5k^X%rboZ1)&;p ,AL;opf Z%<]2};7Gb &n] fL/|qoU}xiU 4:9V,>;Fm}˷o{Q)j3AՂB<T-H+٣GՂJ kD*NU ;TԀ\ՠ՚RC/V7i9/ ioy]XO@0K~3q P6I!hB]` fkaϗGuN/J(ߍqs]í\|vH-Jsl^j}Md^r䏝z aQ8A]ڵҜΫ-DY 1[R x=-@b~& hP%6{w~9w닰>j>bCR]]2s`++0>.S'ɵw%xw}36o>lWc֪nF/6m5,-Y﵇CpeEӚ mm;'_V(TQ.V\z<گSC;'/ ^x S"m"TWS(Y3sA3Y\x戄ӹ¿'o'wm>-Z[7}1 N-/v`<_@B09@9@9@9@9@9@9@9@9@9@9@9@9@9@9@9@9@9@9@9@9@9@9@fb=X;;M~/czx#~NۅM}R@Ӹy-C8OX<Z(-!J2| sK:e͇p{0:p0Ve]?t@`,BER[Pp6ZyA"] ʆau hIilv=^@V4.yYt'JO'ָ;Lv~y8yG.ޝ{u|_?| .e-|P T=MF;eǢ =: i*b.^3)/4,sEg*")bNz}zےs%?0+b2tVZg'HEl"S^Kϙd+]x]:$L=c?}=(|nj 0$UkM"]Ө^l*+>Umc/ƕ*=]9 f4NZe#jmk+j a9Gjb)د";=zLqÎBP :{Sؠln'n0u_^p⟯??}ë3{Xu=0 Vۃhe9PUt=κ[VsqwՋߧNR+Ԫ:ZO O_~_烩Ox_G٠BdJpW@l&m0MܢZKYKF6!99m51c#m{\,>Y\~偒"껼r#4fCdd.#!+|!i& .I͍</Ͻ㰴yi=}><]L!g{~Q@{]rrY1Մ0bF8{Im^hְHIN'va6?P}[SЙNlm6=x_m*dC?43s ƗiVnP?yrkӵFl~0}&ˍ|a崆֎0qe7 s2m*͝ө8MM.yЋxYy9_JE@k:Q xF^HM} [ŕ4;BYBf)0W]v'v&~"XtB >>-a|}+ ~x:ټ;eԺɗ|\̌S:w1k.9oәT!B<P0C2íp<(3D?ԤF.33e7{H) Yn ܐLvh_~fgnl\ўko+6 ӱowX?:z}\mf.W\tVeyy|=F⡚8$)!1FĔ;WHuC9q;ef9ɜ_tTQI sxXeQo0GꈗY&%R3rn;u?L:<ɖc8Sl f6xZKo>uܝ:gmY]@)u: Jn2SG -'QSa@ CU/DxDNL憇Tsw-hOhFL[P4V&1h4<j2hzZMZ ,+BgE2D!Ռ:dցKG__ t|b%!3a./aV\ rHA"kGat|HYIѤ'k2g?p3tPrn# #'⡛Ets_fs30ϭ˹ J `DO lLfG7VL$J47`T,7" *rF1A "30^Bdy39iΜz7FQ\~jkwx.Xz^0 f}_!3ҏ; yM6##$ߩy(=M^ͪaYJׇo}Cy<Wfs{ZzײkΫ7~{s uii"t c\ۍmzh~LULeŢȮKWͻ0_PfQMqό-zYq`zUfRjx(]SOf4^VEO9G߮H^#i[|3qh)1%ʶsWmRP Urc~]ՙrmIz6#nrt1 N tU̥B) Ω` 'E4 |}9!ܕ琜TrY[nų~ ccS%kFBL|(|8;[^$TJF\9bTeJK6Do@yvG bTdz%gx,*H(iJkG \1 Q ^ՁK[ui>AkYPV̦͍jҏibU\Ӟ{]_!K(M>O:$Քn1ݠZ1|<"{Dvsڱp·eҙ܌gm[pNkM6<;ʰLn5f4j'uyv`??qº'N,D-H`#f2//_xB_FJM"*1F%LHrF̓3[p9dfIEBIk" ~ɑZOB@s8JFZR#v֥Df󘤵8A:+zY}*o1)ûŦ]xXu7ReH. 0xk?"Dd#ڨPuӌ V@$hP!!$.i$ZX(˛H0d1eiF ,QmmD4 Aݸ 5Z/L3\Hn٦B'C©%tGy~x7a:YX2y#I" XhA,s;20MRĚt!wƘ] Mk|ZǼZ?ovEH!^ƟX s;+ȳOLVY( qLw\<)Դ]Cξn ~P袭mчi}fAt[9jly=cQk*>ip{^0M3' ^᷍`/a匚sjcv<S?n=COsJ{D5$W4ڔmQ!$HPM7T"@Pz"P%Z3+a`EuvXܦs)}.;73=K 9tD2Rg-D  4q0;c$:F8BdiH>.k\@ͻcݯ$~t;bo0 8M\Cy5w9ɺopt7\÷쫮oOm,Θݤu&%wZxo]9+lj [6YJ;B 37fR!tLF83pfUYiap%/ T"E*RHxb*'AK09ugSjbuUI`g٘=>XDM` ܙ#N H4h_NKlqeED>V rQIhEi;سNuCs %VཱིN[c#^HҩBSTHeHl7" |IϘjʦZkrF+ű^:RT lKGꏫẉڪJ`*p ͣ?I%̣.d2GɜE&s4Wg@ PbycJ 6Bsq$j~L#p 8gnfFzhPǙA'[}Йx#.r!ZY#$$y P{QCeQ" ] /FFC4R9X@>jMrHȒLhR /I|ɽoؚ^BF|BqP`ru*q{l,.6E>h.K"DV-'\U" Vq(-+S3#=C$6ᨀ HĈJ\tH -rdtH*FP D9J@mrpk1Yg3_sg5Wksqv!69&Q[s~nd9S)7ԙ ln]zt3t.ݘqIGEUfy\SWzrbNj9˕KWM邙2K޸7Ѹo}{plB?t}}3kzΣYL϶ I5g͟썿MmR[AxYCA!xu[/܆2A@4{귁E~x]B1x[Y9gc;6);7FفeI#Dk1FY-=wBL21%j!YxGKS5}DL801yYw!=㨚QCLqgߧm4dT[Op$rQ6je1Q3a! }o!Z橏p(̹_"ڗTJ5^k4Kxy,ӓc DzmXn'-S7_zF2$X(;-"S!v(&8D%2!lвRBrXR8e=kB#ʱ`c)U6_ A42g32Uaa1 Ma,T Loe~Kp o㦬@eݠ?rv9cZ:NQX0 T9r3)e5Oݐɞ ,_ceRy N⫙ *!#ڠ&DҥKh(池v19-ڜvYi<o[72,'wL2 y _CYr 5E0 G2d ) <*Ě,2%QuhC!DRtTblÚf`<D,""/CJN %p@ jIHYD4" 785I*ʃ2θ༆(<-2Pܢ>*ٌw#C\ O1IDRN|!phs)ua<Ի {I*7ǜ:Ǘ~9Y`EO"uWSHJJ"u"H-VТpd*+O&CVc,\AhbN d,QW(7WzH.O \eq5j^p5 O •fhzBp&t* T*K+ f)UWo K UXV;J+(;vR.0NؖC eraj:8(]_Y 0Mlߣ֮3ajum'_I*Y]ˠ[5q9a#*ٻ^s~P'{>< &e_:Iai%*TrW.Ul xr8+MO{j2tIa "U*T9LBTզ7}9-@ Lu(l#A= R{ !P^_8jԹ7[Bůx[+N}yhp!ft;.} _&]vg1dDy (Q6C Y[PIkQgh4VBoH]gyQ^y8di6y,= Q{)g~?S \NޟB`{Ć?8|uYq{ha.ץZj su6UX).Hy-ih̕yá2V+\maw ߎ4z۞ؠ%;}6oU8k{+=oV|3&¸ mcz$=d0{zoP]ySd$UuOuyˢW{H]GITGrK(MW!\Ŗ d_3TeB%/$38=?4qj䓤BOo{$'wz%~kz9&&voɄK_lLZEzhVl}{9D3'sU5X !Aj:}ڥ|ZۮgnӰu}` ["bSj;7<׽UF͆P;w_/7rMh]?3!t*$TF ̊hXqv긂 ~4.pun)sݴ[\XC+Q65n8 @|dL"U`đ bgj-1;5Qފᄊ4ז>{7v<Ë:@>o3F.BLѧBڠ^z}7/IAY>IeV]̾ue< vo,\u`U(pdּcsh,snG9[+Fik갣4an(mcUkޱzؖ ;wTNCimS-USLR>ߪᡡDJj$Nc43n|ѐF+6i wӏ}}g=P9;Es3í&T-lkc-֨\R(Z.US [ ]cBWx8]!Jk;:Br2+Ɇ5ʶ5uut%$$'g[ ] et(JJ%Ȉz#R++H.thU Q5c+}[HV2 * ]!ZNWҚҌ k]pm.thu@R=s 4A+k-teп*+9ψ]!\ t(5bW_]MZ59uU.=ynrzR̺5wtk]`++h.thi;]!ʶՓӆ S_© Q.~tem|[upi6yWhePBYymjסBZ1nhi-m؃\;I_VNluFVэlG%An))ݙ9B_kY5nxpC ,0jl2b(fb.\}͞4tU&Bƶh>]= ]qJ)c:\ vB;:W2]!`ó+{Zh١oduut%f<' |A{zht}8]J^YxJQ4Ɉ8djUttutn~2p9ͅҦ]0eN X ]!|B툖>Ԏ(2˳ZDZfCW\ NWb`wt;]^o>` փ+|`G=@iuP265JvtSLj:vBRBWV4#+S!]`I6tU5t(۶EGWOBW\Ղze> U"Bc+!83:#\l lbWʶc+)QBFfCWf ZIu QrȾ$#B:5ل%#J1ҕ*]iI ͇.Dܶe۲;z2Ru-͆J6d#++ "#BFdCWl vBJutЕqUuqQa Z|pCZ}q1W==3:֔rL[]fow5m甎ZE7UC(_mŮWxzho4-{SpUb=ۻr3)KP??Ǘ?C>^0YC<Oc}w949]goHf*%n1F- ?8|n.MR/,/ ~^~ n>?W?v1#?Kh +?)Uc9a__y鼖_x1p{) v;\/Aaǫᢞ7e76HpuXp2( Y(ijuv>>W-6# Z0N{ը8YJO } _ެ*톉$=Fe$p~eYX W`|z͡#n1}<GqB 4xsǤ:cyH 傩2,95@ 0'`^|H3ґ獮/J `qm:{[dըo<<??9Kw s|9E^Wfb.ZD i25Ἵ3p5>_Lyk,G793tBӊ}S $²;  }x DAAfy .K 8Ɒ3(p+DpWAEo3yI2-As<Og/~Y=C-;ʜP)]U5jA/΢r,v?U-c:sP|]8@?LR<-CO?%>je/jdF ˾eg__rzvzJ>!Y-)7tWo)bAZKl;LҳgIYR|rA/il)8ݥO>1}&^Q}虠=0۸ spXbФO}a;Ɗ>>*ow`ϼgrɬ==u3>G²lh̄oNmqm֫{; L1,R BB|$pM$% LG0#cxCIK2F":NpolL*&&E$Vρ3tU1%5VްYX/=+{ #"C4-v?&w;۫XwE֊[Vpt˰@鬊\:wU!ł&E$)RaH* "1#x o Dr4=2*uë)7xx*pKt NeFSA"AHK-W[< &4J0;f&5ae@lṦ&05E=<jq π^-DDB)yB5RMsm:At8XvA虉,A5s4X RTir1|w : Ӗ,,OϸV8T2%3)0}9&eyHI}azc2~^4N0C L70.zO"y*$\,b?ޅ ҅3=L{jX)c.e>e -xZM8oz4}.Ng2QB9a'V'n:M~d:8IC_[oI`SI%MS6)/!I޽YFB:ͪw~"9}һt%,z zoJ&:;;]\V/]t:NONOA75MRƅZ5m>\8&oMr֦ӽ^_0կ;?7]W `t}oJYy~~Fѷm5)&!7tNwN 1%GAb/&_=sp18Mݵ.mmeB.b`=-e+0}~Oǫ)WH)=<~ǰPM(T缰h~0꧗oݏgoՏo}-e}@ z`(̃! <5MU -F_$zh{[|@C-AXR8Qn0 crr5+хt= 1' (DMTkT"bؖP?Wcw↏qnM#I#4K?Bl^:ޒSFY0ƂMRiBr$7H8,)_Jm :Ѕ0=R=8ok5u4ŐQxؐ2Z*)lOTxU֒n]ui]]>pn͘khlNȎ~&oA]9+h՝d^TYqʇOY-*oM޹hL`Vu;IR?BNdz q݋2 ki}zDڕC}9n DI ?omTDQJO`_;x!fcP7YQoϡ!9ֿ \ |6hiŢAܑuy˾_u0N;6VbmqAѮIVu KKꮕM[Zkrjc+Rp!Er9s}E S{)ck]+b}*l+C&A+ Ű(|[ϨRiAN;xHij'Mрբ*$҈0RLj!9S1W0n<"Yu5j )ɸM)0 Şm6NjP88*-97U\/){lŻ[uQ +B3ŝ(#" D9B(<6" M_[\\;<ܳؑAC>н͚lo;@i6*M`CٶcMÞd05Ӽp (~9K]]^3Z Ay'DѾ`L UZ/m[{tڻR~3o#q!H8<EY 4z!{!XDeR)Ed,l$6:0ۄJMYKl9—iuGsٕ, ç 9-XN"1Y'9)yo4Qͭ:b a!IA 6)F &1'1}(Bp 6IQB.H`3w,]|eq |8m(nsi%wwX>؂;r\8ɍ PhCK [[/M,.&N4Pٻ6rWX-<.^rR3٩NNar)+H IIqEwS"%Q(PlJ*KTw} \b@,1 ,VF0HٸNzsj/mO <ކ?q5`{g7Vtxa>:rr'(֥6]ң,2 M`hPꡰnP>ī'ZR[dӲ +VGud^~FW x+w6Fk.{~z;v\__?Oc]vZGFv>>Ofxr&go??^ø~lXQ9NWf >]\wo*j爴UHVpJx3)J+cIw^>c|a28?Ng1sZQC.lkB~~z-Q5-{*!(4~4W&(ϓMNg,7%#H,g]Jh8IZ+!:."[O'U܋ܗ6?]}¼XY+n9JY:$C' m)zfgDѹY/am\52OU1:͸ a[$IEBb\ȱ;딵իҾUODݗ-S Y&xtZղ,?Mu%Ȼl@"H"eq||'3ݫy^$Ew9 zD, twOd_gZk}םkU]Z1jOglg c:9ڛJ <S/)VBS9 R,:NEsOn:TexY %&954rgBx`gxBϳdQ^ZqzLzU]&07uNbz'*A*YfwWBJDreEHyX rQIETKzoEGmsS5uL&sMH[M:mCz!IBy/B*jg~{6gv ^Ȍ}O*gLgY3:rAzk禂puVnr.cώŎ>j%`Hn 6:n o4{%d釋" 5'_d!z<}ħמfX`qL)#ؗr?ǻ%syo5zsp7j@hqHs.".6Px?|Z4:s7b80~͡hASq{}3gM~gv1M= AehŶ^{k Q_".O-17l_(fM`F_Rj*e%V&5IINIf"y{(d#Xm@x8ȨМ{n:`z]|tϥh<_W_H·вkCe/f7-^>˫yiw+EJ![/PbN;j8\p)؄ P&h(Xj#jk%NR^r*yb bЏ',9w_RtD ϊUěp(ּK\cI!r`zo s g[tQ/(9!0$qZGnT )5ZyFHY4`lB]Zr;|6A7`9ܞW^'x!>j~U'%  B+tU!*GWy0XAbOsI>Fr, 1GfCVt*Xb{EeVmR$vH5c (w[ @HF%S69k5I,y]"ۦ<&_kO[NtI2ڊXK7(pM&eJMavM(yY];|kU2AӦy,[WLo asZ櫫M7Kmɥ.S%'`{ޫ˛]m(oKݳ-|MBَgi~%M lҵ/[<DzE*ł{͑"m7q iODJh BVV1+vrƍ0w_M\v qٷe8Q:X̼WK^h^lT;SH&Gj8G.uiX+Ua\:{\\N#<9.!y%A&KLYӀ O1IDRN|.p3sֱ-x@XXQaя{~ DyQh!;KFxUoGR`B?|fH I*U>ytdL)B({O  \s" &d"`9p1ELH"ʚR.{Zo.ZD>_˃ս98%wc1F#:zA%& H>xDsĝS9ۘe.tRNL$.}b"FAPd33),>rD,Y֗)fN|ͺ&ٲ9NNWŇ,_frƼe5_t=.,㺙{]QTPflBZgJr(5Jv=FR>+̬]7Тp9U6C+V0uVr++$C[ 6` `*[KM*[ի+ ׇĮ`8 g0UֈZ)^\)@B3Y-W\yv}VWZK*(?沃qleW!T+48 s(pմp4Tˁ+aիS~U7?|}Xu!e1`exuxM7\)w}{qžg1|cHzݶ꩑F\QW\f^~:\e+p,p  \UU^!\``MU6Cl-Zi• ٻ涍$WPrrYH+JTe7W+ `R(( H.[gAO?3fwcrUR\}J`B< s \Ri5n h$2W D+UV=))ƒ ͕"CrU1WI\\U\}J4: sc~8`WC1WIZ`R[\vu"WB R?*IQK]٤Ol.Db ki> "d)-?.39S@y`j>"ɃHG(UNjk,*8 =1d%rݪqX4%Lڜ▱?oW.}Hlo ZmsT-P/Yu>LmNNJ0}yZ(ڜ-rm|SR#',[ٗM0}ι`r,M'$e+A>Ac4K1ĶbҹVD3C iڻw oRa),^俓OS)ѠOMhRcbt jjqNmON0 Nʮb9K`fE>ג)E #흭 ~2c&60alvei'uej4;E5O=~it)&Iw-*ZͩP閗hCJ [cR Y4 ^ wE%KvR'tt T[͟ط ٸL)}vulVÇ%(n~:s۹*:hapgkI P7 ̘U| 7%?9BEոgƽ%ERV X,-}$$T #V}Cm<'cM\SLթXzm!,uZGK!3jg>sWL#,8ZI:5 n5HM/yhv"# X^PMsZ f/W)8oE ^=F{M 3BX9u. ތ2Vߠu+XSsް7;tdA C%>`%0{K@8$,c$`Pr.B q^A:aR/1",DDꥦ8&ZBEA`cc"=yZ4iS(b!z(P>Nfr`Kw,AZ}m݄ כCvɔ1ʃr bupQ$`,`vb.j<0aȱ@"b iw(js R`6'#L #{5 XH>fhnm%0l\{g<`[n$4Ul4ۻ%y{U\votC vA:!} QXOSF6`#N˵DTc1࠼ O?`z:M`y^1*-AOE)D"RhKćY۬X&}d -:CB= ٍB'ԇO-t W)e͸pS!) 3#&!5R]%,s\{`oƻ7NP?W>RI$AHAq5+So`X'5ȾyPx17՗YٹލSV<^iuOuB3K,T:@M"dA6 qw|~u[^:/ay ߞ/\t4;#W^\kre>?um%>P!|l"TV)@FXW`nۨ*^ײVWzټ3Sq5ƅΘZCmqPpQ^qGӿEbf76yD޲#~AzgH k:jToY誱{ƮV(Ɋ*2HvWE?B Gv$UXo*p_ u8ɲ7ϲ7i)ÒIrrNk(l4>ɾOK_uuhˋqoY;nߜ?w5P eU*$dzm끯Y HJ %5mgLYԡw3/p [jg;gゟ R]?4K|i3O 5#; E Ƥ0?orr߿#g~Ʌ>}Ni-uVp\'r:dT؟e4y80ߕuo6 ,w 4r~e#0|] Z$LxθuD,=%Da/ZF0a}0 e*q9Q&SKCp@3,Dz]"0r/7׻jh{mhM8]Y>Ya)6xVϛC!\?wJZ1W w Jw}P|mW4eZ#HWnHقBڧ==nCK.{Q 8$ՔVeZjlipHq~v(VTyZPtt:lw,܎܏tP[a^R2(7m|U"َͧ{#nXWhw~e9 ǖJ`Wm-8:=i qAG$#!DJS0RtĞ["pAq)[,UPԞ1z) Oˏ-ET s8ʴ0y)[ccDZHiY׋Ki^|DqoqcmGz׋IO1>!6=½U}㖣oM0xLc?Ņqg`ZG}Î`ʼz~jFU\=> c/xLbȱ|u-a K-eNғTLmSZ©bFFD  fܻEg!mQHh;9@l8K»絯fRhE X騴ʃC#MǸQθHtiά@b=Mr{ӉDXtⲾmg?Z.^Vݭ.=8]tb\VI[_?xfqa~sx"e,iY}!2X"\j* 䵪1sɭ`)$!dOd[9Ϟ|8y XK9JB?F`gYc.('ϏH>cg/,9nWB$W଺zϘOz)<{Aq//vo]&\EylKi22(oWR(# '0xa^p2N"(aHXd"Ҍ;b(Ai(uT[T^[xGOoS{ߛ롛Xgg} xG}řz\\*$0%Ep" 1 Q4bc&PfTQ,*GT KIwT! MgZ ,gwke=~Yg%RJǐw\+\6y$S+\.\*CXoG; wm)g.96GL X4CEQ):BpMӖ( a9$$V8m0sImc}J!tE N&x}C6?ːQ 6<%B)`FVxHHIV XyTDd'v 4w::Qg(f}:ݼ)<4״zJ@o>ˋxdMah~o^m%s=|R =As f.2405 L~}h?Vo&dq_{9D6%kJ "?nGaψ^~z! %V)BX-{5Qcs~uՏϗf*ɈAY+$l҂۠6.%Z/ $ SB*.GeU#e8(w86V;:Neh (7Ζ[W#:ikJ:sG DZ)NKkЀ;h#iB3hUkBEkڠ^+Z#Biy,<*:4S$Nq1ʹJ&EI3bW/w֫Z0-1)e.Ixr*hh ^#r"БY-˘2$X(;-"S!.$QLp3'bIn$V'ƢG(j,paB<Q@PNq&eO\r4dg#9Y6\*/Ilb; D02W.֝ۍab.&;ZmNZmN;yI<o[72,'wL2@=gyx\`$nMQ{T֣D2zE{ښ,2%:1cU@4Ql5o Ifx,|jya;YVz< PéSpT]t4C+`ZRE4" 685I*ʃqy QFIY-2Pb=sblG|\:Xg1).vuvval:“Wd4ɔ9/(D (@;xx4ȹtjua{#LXbY9>e?>Q#D/ۂS0V LPWn2aE!D=0UQIБ1xyDHڝ].)@晜&`bLJY>ˁ):dB'頬;[+ BQI_hkrڕׯjwxrx8Jc`JLZ@HR{D#8Ρ eK¡TTĥTL$Ha8HIaꈣ:dZ +qlV5e96J#gX)3Z=^i5󆟫v?j]nw.YjnaW/ܿd%WvCq^|ּg|*gHmLV걫;G^0&sI ⴎXKv(ʓ55RFʢlq.vc c*kP UO5Zk)ObݵBܸCzp%j 8_{Ϋy"G uBR/-TYK"Q*0JTT*jyWYX{_Iֺz &teAmJBjIg* 9IK>P8'0GD$-3&~mtrkYzEx>D#"%(ŤILDkv' a+>*{fB9p'uL:—gEiV?_xVk~a^cq4%Կpa=|U8ߛUXxe%JF Օ(xZ$|Q ~$pASuA)ؕ('WK!+.(%pMnz8y=׊B,9~-izw=:LKs]}Ǜ7يŊ0 gr3ۅI.M6ó7v8ABֵ$ږ@nr]3l}36(,QCj3g1ඝeG[*#r]Yю_G>?fsQ ~IOhoԄٱPExDyX\9_~ӷo~z6o?ROӏ~U/8qa8Yw~٫с{Mд(465lnZN6v𢡊\v\ھq{?~Y{4ț-UQq?;]\C/_n_ڶ>^ W~Fb7GP`1H7$8,$E: ͸͙+4yE IR!>ZIu"Pt:U|=Ct\;/RiXDu1>:<8L@$ Yf1)3c9{\L&v:{<+JZENyA烞]vԝIQqзgC⾯+7՟5~+YTJp'AD*ιYRϳ؄>.yPԑb|!ق2}O{7zëٷ~_o~MfsX6zq`\ۿ Zu&?1PKwQß3&H|2([}ɒAZe̲_co6S 6`׀r ޗ/"X; 'qk+߽Я{ͻmx1}hrW=K}/?YB/wAyi0.̫2sf5mCvB[{Xz"OlZz)/W~Z}pbnWH]ArJR]}RL9;-Stᄋ8Mշc#SG?%x¿S#ArunZ`hX}e_ 븵)%J+2Њd7lhdgKX$̼J;${ 7w?/O?dQBBjjıS#[(Z`sd>Lm$ aR F)LO^\ xC 9^mIsq>r :|r\Y0lapCIМBaap?foaޣOzb3 LqԘi.$x&W׸dO>`4a/Ns25Qi|87_]oS8Z+~M~ >4*jzu- M=j)5ߙ\Ǎzo>5/j޻y#/OzPnQdf]~>6]= '?UEV'2J@fVD$}$)i Ѵ*HH…I(ZbkNG˻9Wmh.G: #$@Y1,#Q2q^v;Y'Iެ̑f?~/-sZߪ^h$ȩ)TeVfkQ,&Bp;9R׋I4ì/~s^˻w-^/k7b8\V(A6[UߒD hUזMFFa4I YLf{#KrAHE iqTƬEh+@)V}B'7liV}-@lwuԟخ<)uZ ̒N%c2$R6JpMVdl)I%eGz}9^_j͘\z/F׌R8* HTGTI B@Lµ4=/=dBilJY ol&ۀ)JZ̔E8_{ x3S /!keYCH{M)9Ф4IxPQɔ F wLp+&Qet,yrh.gkrґ T2.C#'F&@Lڧ}psy;ThXZ鐬C'dɐb*@ @H)?EZvϛ&UY:QS#J̱6gD$GY0ۢEI2{g`C.zV0iF7[(c} &G(( E?bB iY1kQ.)yVcOJ=i]:gICR [AioeM1jd䒳"a+]0)JcEҕRlT":˄wh\ zXoBXr E #1٭(Pd@>p `-ĩжlguG(bIR~ (jl%.dp/C̍Kk#4f,qY'qah P i\Gg*){"/3 +pA`+CQg-a)-%TC(Zex@WwAQttd-ymwwS0N[՛LBWJBƊHa6&b4XIaJchıX="z]C HMj3x9CH4[A N\ƭ"fGaB*?%QvdrI1 y*ED1nOY i+XCeҡI  m'+Tpo*>b8%ah_hґ%ᐼ 0|R৔dflB)vNeW'E]%I/k.6`-iD^>ѯT%@H/%i (D4pP(46je| $z aΡ ~,|:omq 37jy V֮G1/E5 (N3ƈ)T'rvZh `"}e;zn5]_iUurr2PlQܸƁ|ѤƂB'E&@pMPE"eLԴB*#@& .Ks4m=Gd%YZ{[f@܆`m]K V&ESU_ ~}sqgD%lLe5Dw <}X'2Ny\E4Χ,}XR חXk#p`W(cޣP.){^Ls@ QwK }30%LD7S@ʀv1Jp%7h';0]xdWAfRh" V N+]Kdyz\=EȒ,\\5vƍEQpIL$E}4" <@ {pw FD֨W 0$1d+29MJ6 M{}tґp֎ip`Dj3˱o R)BMVDEK,6j,e ԿIn$,B L`_jk٫&Ɛ)^`6Syym7>y~#F@bsA*c5wQzBJ@R9$TiOW,X7mdllAV(c (+TC(MxrUpՎ9B|{?"w^^0`\SL}BN28FR%HԜE`,r"[`pOK6"\D:csjg6im#i&z@JQZ^v H `2 Ip՜Z'k#kDv ʷҔ// ľf1HTBmE1ڀi#bǍ.L`<02ja3^\ULčr(K9)vcI}6FJ˥>t`eSFuL!jɮIBq&[JoXnG['f˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,b˭Xnr+[܊V,PnuWvx\S?qO^hꦂaZ;ZKqQt\F}#Hp-8>iP»>uo^s8zfo_]g+*8__~\/{!iX_+BE#_g;?>՘_]ypTĹ_GmU|x@K( ;m>#"rOqS=^ w{%ӳE}<;07 4x^ \aa7e@p=%vuZRo2JDZN-F9w2 qx} Os&X,b凳Ϯgv>u~0Ƥ[޼½]߯6^F^6釸(+}:xwt9[d;9VGZ맱ƀYx(}ys|׿:c\i8j{f7~E=?=rZ| ]vwCUvbdn|vyRs\yӻ:;:.X>f,=m " s9=8˹q~vi9n-awK/bybx|րuo}x~s >խ]~oW\n]imxk^{jWss)L999*,Sw*o6VWybpKy[SZȤ.7z"7=7J<Һ1):|_`ѣo"ƇsV]nF!y\2KZ̸K[zsKQՕ>MrN%4(]O^PTV@M Kg{W..9?z'smogۍnEJc'VxU8ER7ጕk"K9AEo3 :,I^899RAx?eTCeSu5 :㙦j.ZZ̮i0m< ôݫ0n3@c1hR-t)c,$qxx]^6e/fGs&n4բ$So]R+< qd葓'31VGSȏ$dtƫC ugRgXJӽ1T/NjÊlqđOe\cnmb]Yy:eZw,fWHR & IRo$B|"A1vʃD55[e9};ǝ{(?V{#W4̺¬ې&ϫ2m)y"Bm VjEсРZ+D󡎞ڟjiچ[n 3R9GŢg&Iǻ&-z\*~붩aqT"}{{ZMDS1d<@>9Es-22QB9arxSt ~p:C!xyC[6".g@x9M]NWUw;A8lLE0Շrt9ZwU%W*It||4^*Ct4NVE^-jjW׊ \BJiNM.kgeMu좹p 6t&U[v1[nv@6~>,'n!^ BZR%!7t koƊfV52bѸQ8yJ> -zzr0ݭ.^նwe9Ʋi6ÂGڧW|+|Di1-XT;*UH?Օ F" &H?~˓/2o/_QfNyrq9A'eZ`޻ ?W,yjдilo47%ߦN_+ԲJZ1?>/Ӑ}ߝ/'W7[W:m7E\@4GlZ8~}r4Cr΍N<`U<;\Llᾃ 6Rɨ+I&ҩJH!x^Լ'N<;p3O"lp#0#JBxNI[%ڃxҴ) @Np%*t2 G!ބjMYR:eA." S: !J!4qL&d':fκP Y `Ѹkc D L \p^`ɭ7zfy fڮ{=췭OH"B}\\{#(JR2_'/`+Q/3 ]1] wy7=a >aM 0_GRgց3뭋4!RpJ9-QL O}(f}YKqU2  Y0&ZqNf"Pks16)3PeD}&p $I[U*!d@zj ;y?JYwu(5 !D(cݧU Ocr>R[y$:~mCg*bV4h3Ϸ`:#|d;d"x8Kz}ř,cEdK1#/_x09] K:T1)B6#(_l[u]>>pJF *:ƙug`[\Xrn}7y]ȦX^&P :F g«f-dWxu\ٮ[ 7e'iےEK( !$5>h Zwe&y?^_HAz:%PıYje1S!^4R*ŕ4EƂ&zܝGv}B;\p1Ez7h54Bcrq" pi: oʱ%>Ӓ (ߧvAZ͵UK]MGxlx$j/V:($#WPILO4˼p=r㌂<"UC<g ~e )&e!X_YVp^zMm/y:<0t"HZG,9^<" A &yl Fd1H5:=w{x$qbuh³.CCuAkn Nv♁B4>p1np DkĵH eLڢkA*6(hz1p1Si ) <w SB)1A;N}Z4 L_0 ߡ8_f׎!Oo8>x1P*/o*"$7WеVW1{/M{v*WyaY-S\/o"T(A,^*`'ʕ_BP:-Yr3 1D8!g__瓗?MίAKL,Ot䞅@H̥`˹lgl|uQ7vr}eqzFׂqՇ.эes6: hZ#7zXdIo.9J9N '&j*sp]nw̼y6nz^]wGDCC7K/ B8|Gyc:<ޛ.`r5}Uݲ4K~0ZKjhuأ 혷1 ӚL̴"+CNƟCDޞ=ʙGw/%b KCR)K%^'B/S`H^eo6v ~08?1.{YnPI9 kK.P5f͓2(% 0K;HI=$'y gZ6гc_>3));N0|dq2磉Q\EYOr8Kc(O>,S?8|훅o+'Kr@O<$D$;K`;bR;S ԞPwt=҅=ҵ8p#A.v{sozn~ߎZL&Wt0?@~m/s ;%N^>em%V[ 2'ImXFx,=z&H% x 8TN4 LJX/]uJ/g9QeNj Qp6¤DH<a Ug&[gR䵯IO)/[Ga6?t3/ kmmPΡލsOeSQj:u1BrtBIRBBg; Qx-XIri* O6}d-c dymljE0ě~ɴ72gYt25]'G#f7 ,N&7bx1]3Q'>tnkca|s*3cà"0L\w=(XW}X6{GR(Ī<͑)+ToVIl*RB\wRfC89]3pݰ_i'\d VMߨĢQQ+ O%~nLȽ!eӃg"1|Xrk~o-ݛ7;ir']R@J5'lΏ!?И&*~_hӪo&YS'M,oeP~v+V&ƽݭqU/ef jh*#u3w9'=:ք-./#7A<~`Kw E%%O{0 ijGdH֋I\5~㳍CQEJ[[g˚E <)1Z_ f*}Dz}5#BV`Y"U$%@F` x ft֢-P`Xu5K%-LkKYD)m*5 $BN}H$NoUro'0axutLc26`lrZ"1&ۈ>*#6gZ:lcut슉)CћcLvuhkUH?{1kZ>`0b]~w£x^B&?Ig gn!A)?m \WA0䵑ð| ijaRՌ(Twu0GgzW 1%ſ >cM̻K,ɋsg`J>3@t/n #gѳ:yyblCx OnP%_lo\QMB=eih BaD[\ B Ou\c4&.HROp)Vߘ|`+2[F>֋i' ;$8n~xo>=_rb&:C lĢ22&yy 5%g|F:@̄B!g֓I]&ޙ4y34QJ`BV(**Kkрg*Q֞ٮ⬯@n|uկ?lY 6b& 瞧E,C%YV eQcCh@C%Cd˜(+e-tny n);|ㅗ=:hd *r.YcXGL)ef)&Ezi矝[~ g#\n?S|X=7s^t[Jt|FKFd-xP]6CAJg^o0hU~~ep ?P٪ۙ~6]_jc4=L L kPwǤ8ʾ,Ga)b{x-V&2 6@TRqRdO>cMkL1yؿ`mcxfz`ǧT xjʐ멒NhaD$yddJٚle+AMÁ:% ҰYȺZK٫(cJbh BiNZAuVlWj2F Dfe[FeE0p̵1M&լӬ4g=I2cǾ[D[čd54k V"/IPbU :F{ ).9OFm-#)8k銱EU )DlrUK -x9["j&v}:]-=i,.vuv%N>(BMR)>j!VIcLna9Cjlxн879uǖޏvGG߼bֺ~wufeyc~ V!&+P)^uRo$<|jgo˼XOɟ/}7Պ6Y1O߾}Y߾"^ }2 ռRTы5فTJ~s/ C`,(5"e:\ BjLy]w~՛W?\fnrIWk?y$;FyQq˃ak^I`8@ kd˘YW+)dph#u)XI]>6;ϼ벖f*Bc-I-ZMf2;;>"}2YҧUdz) vl@5yϻNW%w6eO=TZKc z,A%=XKc z,A%k=XKc z,A%=[Kc z,A%XKc ]=XKc z,AX{Xv4~\ݹgGȨRP! b"QVZ莐rRDGHFNe\{#u\rgSdp@En/ i2E*gG=A/?#Uqy=5߯y :" k?X/7p3cG9%3dk4K.V19xjU(PibSBH2)S0'}XP ^i=Q@l]W㬺I.PWPجjLYNoůk\:,ϣf_?}2xK/]}އ_?^[MV{ 3 VaPh6br,dJQ+s.#(KtRR&Θ,Uv0uZ*#7'Q0Zƍ4 [e2HGD(RX5k&v4 3jjn9~Ѹ[ ͰZ p`>/}lOZ0lsjK!`LY @EtVZ!r9}S ؔ> | mJ:FAc"t+$P%/-fo!zɴ 5r{~8pY|sJ;wVȍͦ,_WK_kg4;a1Kxskazs<*lahrˁ[.rwv~m/2󝖛t ; =ZNsw ܕ|:ޕ晔ˢV%1bEq^l dcL6I5%G4MXXUMߧwemˉo 8ksعyHTY%s3Aڱ]WZhPd8)צ?˘d!gGjEqyϟ=tp[>CsL{Qq¾o.F0}W(cwor0IQN)93ZL#n?MKۆA?%sqb}y^3L-쯝YE^\r݂>KGp_Pmi[c(x!fH x6HFZ{1Άb XeqUi@.n?חa6dqZ|2KK{X|(Im~&%WћqPo߮`=z{P7to/Ģpoiˌ&K 5V=%-%j^l;@HpC9y47ֶB䌂*gRbr)h;(r}h{[CDE9E! Sm Y:1Yvd܃MSh1~F4[v{ZA|GA53/@u)Œ ]SK5 $9ƹy0<8-G;mEykvLK!SM~cQc| @ C_.͂r޴|l1m8+sm4ΩZzܫ=P6qiGjS/|ˬ_ynҚv9p=lâs  > (B!"d#r,=xqpS3vSc$7R :X2o- "eHMHiӠv2JI)w;$4"P (O Q XMEǣ.gFC:Z*`4Fka(A.yjPv`Z3C'"]&e)T{%!,Cs/ks8(oRaXRNNU}r*Npt8# OKD >C"qXФ ;K1z9\&˘dP ) I=K0-WZ^t6)1S**[.ukJM9 JYqOujkg:Ky|vYxv F8_\VsKjn$ j2Ҍ;BM= {6uChvv.,@}K3f1bUavFk애{]dSMJ,a^X IsS.=/~,EO UK ?G' z/~?}_c_|/a,)8CP\sWow蚷5*vszߦVo߫+ԲҡZ*gobXo=˖ժ ^q?dG>A͇t~&h5*7 Q!M~Lonζ|k}M4q'?t)_$ }"W&bM*mcӁGBA2{ ;0yy=)7-tF~ƻzk *bar]BI.*j(-F{Va6 Lk2ql<3{`YtijVu[wf뎐C1EB̩?Bu~߈`bҹVpb#V%%<)Nlڹǰ$*Frp)BSϘ8P=t:FY^1T'@ݷwL6;]_'@bHH]9\AO$%6 37 yN=mowP~T$ Ώ8Ϭgx+t(987JUp\W볫y>ET$ 8J4給Rby)KPVR)Eia{'qUB)P ~J1ky;'#Or#.Ub]!ϫࣃ15նV!fiqC6?8&m g!TN2C-sG:gv8XqAGMbӋY-)S[^+ .ZN)1@6ō8`iU"UlRa[0 ^2lm8o@.{|*m^E"kJ'vPutP&YU%Ssߍ֊/Y*gxJᖪ\9#xS碐!h=h ZeKПUGws>HwN2YVﰷ eJ[΅BH=!Nviù3W`(b3u`y7kxN8xjyPךCpsed\O1v: Bn Sd^LCVII7;{4 +QD!EmbNdP*լ) 33V [bd,n igS^]H]ގ .װ/hqeZ #yx >(߁6`#N˵DT|.c(Ayv x:i4=yjǨ$?pS!) 3#&!jXc&BWw%,s\{O-Hx] Sr%q9J93 PѬ?e?̧IFW#Dݨ?X|ً0mh8tCoЯ-J7ޣD(,E!$z+,=ԧKż,K~ƣr1)N>QrTCx<~fq9{A^W{ៗazO2W#;> Z$L>[INR/O QLXၜ<'Yc3αQ>g4rjs:gY~,{5O&is`u zV^ ɾ'Y(3x?MgAܸk$?]НT8 B:^B6j0FtgKl%IE5yuݸ HZC[{\6oƫ0 ඞ} `o,<s:Y.zW.𛙔6VP^|t,X9L­ZHy( q!0+ ۡtnWjb2P,~[S%X&)_ {}VjeU^=j$o/maZ { |p-@߃mV!Q/tlycCE2}S4>+ϲ̚I5DP'hd"Kn-J5*Nx0d8[Ql7Wzi wI*7,6~Y*RC /%d;2|BDbX\Y_aۓ[_8}ߖA5 -~T|5wCU&jb`\3:FCQq57N0߭kݕ@pC&тyaV774L}]_y1ho**1QEyt1_@Ar^)JjY :" pZMsrDC0ҩ %8FW-l+zz[Y[žw) 3piaL%GsbiE"! c rd9^+F/IQhM6* %!ߢ2. c/xLbhf8Ų]9]ЍKҶXWZU]52NFkaIʘxbyϽy*< hRk+SfzTQͲ[^"E XUDtTZ4pD3.5Ys3+%<*5x5菸L`^ ܁;끜.FfbSwPY.̹`1X>N p/9v3|%gS4aw|OhV[*y q9H |rVk/;ӷKv@= ߷/|kzR=Lqvoqބx|S8RZ̥:gٻ8n,W NT˗|`wadG~U ꮢX*ssjCed`fT9*2ȨBGHL&g:PQ3W^)+e.h3:13/ʩ[g 25^DP={̷Moݎg ŷ_{ H| 0k߯ E,=, `lYc_E$msO^qKέNƅdLJ: wx/\;?-c.A G7cV,Np) OwמMzKw)u?K)Tŷy}}ӽdAKW[ӫn@*.tD}i&b{E$SH ˛6hM5,.t)+RaގINMz9HVI* ڧlRq-\f%c\w`2$l%y# G.B " '99``8Xs\Bx>CdKQl %YvQE{rjAZɁKl#[]phrJڅ 4[[Xȼ2x1ȹ _YaM>]$Uvq`;F[Ԫ 5Te2rrh [7|G ŝEr1K_ml욈)AѻcLxil&MhkeH]LKcV϶em@lhqUf ,x6mPBKH.j@XZ;P .v9Y@O^i%;GBքR*[! MfX}}%9{dX}䐎)" o}/oS8|up ڹ3A!ё#lތ/fLOR|O~L.7*>[Np=O3[Y" Nfn&ɉ1ھ6!WgIkU A,O.jѯ嚄f+zbˌ T懦;U``te/(kX[)fvo]lYt~/=/67fHʺS@_9ngfTRdLyԫqZL`ӐT~{KJח/YYIָZ)eh;M7H0Oyov7{]eE\TpdJ"O{ZշYN96d[ VKFx+Cv>OO)%^ѫ3zuF:Wg=|3z+êL"3UdT*2SEfL"3UdTfy0T*2SEfL"3UdT*2SEfL"3UdMUd `T*2SEfL*2SEf1ΤfË7~ge!X41J&M Y V05TY*$RQP:J摉y$ !jcdI2s!z#*agD E$r),rLч8rYJd.s>w*hxFd*gԁ]yz0-_mLJ q-Zߣ6uwA.؝ Ď͇UG^uSHn֦ux=n|B۫.nD͠ ۼJTR^==IeA~iRɖJ޺G5]yxG+%Iޏ2鎊a}ϸ]o!ˮE} j[˼dXM*B k}֙wXW[柉s*<ܧe1. zz$<1J)+ecRv鎍;~P[AA@2cSFz":UZGbY0Aes.;:zʇ}hʾe /\/;'{UQBR!1SSL #zIh4 &CR)K L&nǬ:s̢{c$Pcڃzmɫ׮v[D1HWЗ䥍1dccD`ٹ˕ P.5ì0bȄ 9 ,jFM6Va("D(29aOK8aqzD9G#VuoeT%1$E΃@:0A^+BK[a=1m1*emLD.~&-F͍bͱd#֜;W/vȥc~uVIdR lFΞgP31k2 1kd1~90zǡx6`x U9n я)Q2 /0#箦c^$<}zVS2JLOV/`U`QH[Z+j% 8g'z"Ftܬ^9T}n~=77bϨ<6-)/ژ#K0F֨gufݨĨO&,J˃17)O%2M })ZyA6 "{>x6% ;u`ʢюW ؐ8ya}X|hӭ󇜥gJ?Q/;/z|uSk=Hb QyNZٖ MmA,p[Wڼt*kb*I)N8f bZk.qnMdZF.Ё8lLR\H|bbcI2p!WVQwژKFC=Ba0n-m񶣞y/]W+y!#,q%N| @C]oлo7oV+lt~/ԛM_YmfW%SS\_T^y5cU{-Q3=WOO@tT˶Lp@Y,)2%Z/DnY I+{Y ־U}GPmZEc!)g:bh*be E@!zDKBഋLI VvۏE]:k(5Xa Uxڝ?m7/N''Ճgl1b1;~[cA&E>8~vbPSKB6N!h"Aa0,t`Ŵ@mN{5*A[;jS_ %l8ZFRȥ8}aybpIJ%^SxWV3;;vۗo~}_7OGO}s:z˳7( D.ƶ[*)Mڋ(8>zT!$o#FyJAQn5rn* [n KpE"YЁ"Ђ^ !`B>gg,2j\e2 zyvB{}0o]E p J!@)喈1E*pEeV߹Mr0B( d#َ E\53=mi9Z43e0?yMZٛIQN)93Z8=&zP -e[R1^͜)qvޛdATpf:lue<,YVrbk ;ĬW逰+|rVk/;[,l#7 vR/NۓF >M6daZ|v~ʥV e߆A. |4ܼ  #Ѿg=ۭzsv c*=u>cπ܌`[9oWD}XXh ռvH~ ֐Ҥ|}`[[!rFAo `hn)1H4z]Gw;oc ?@=Ֆ`. 15PHHS U.i+LA'w+yڋ5Z)y怨§wjLwjDg^2.S;*u5NYl-(LJ(C\?턷..@Mq>zx!Pyv=7Xjݾ)S. :7t8ȿ/-X4Rcbt%q9QKXo3E 4[Y~)hZf^Bn7o)ރۆZ֜V׬zFgKQbr-! j lVR&R`"k"Qc)%J&9NxQ >^I8;8mߓTmgܓ Cvgsc̱|2: %<@rFϵ,wJÃ&ΰy~1"u<\G:yt/Wz\vu`. Ki R11C (壼Ll+nqLc`bҹV:F T2c=x7vp IUp;oB!1qV{tHbaJ;Վe{9ytTK d]Azհ-kԻ')ubB~5p%je9@(AwTUqwzv̿5W}o+=6y`^KPVR)Eia{'qq|Ǡ% \?cqs:q@LJsh >.eA{B}CBem:9;IXȣ* ́P˜ő.6A3;8 Ni扦\Zi1`LyJ¬t<+qDl-1@6ō8`iU"E}lRrUø)ae9p+#@c.,BнNx7ֻ~U.9jALW Ji?ȏb>[ UF8g\s\2ҁ-V۠u#Xj.?ҝ1 q;-1.!/ A8zBm{&wEi9k[cdzI]n=p?{KI/5^.+>+c۽2lLY.<(@)(ƎY 71aJϋi*@Cta+!{b* $"Ɛv{N1'2(b P 33TA`c\BmVkNC"x>ɻE5e뫩b5Po󪁻 `vA:!} QXO{m4'#,N@}09jzF)LO{Nju&ړoHvsI_}vpZz[qϙћB1g N߇~?KXԞj||qAolO߇o>,Fz Xm09 pyh:ɓ Z$L[INR@FO Q LX'Le*q9Q&SKC/egWe>4r7Yr,7 ӴN~^d 5Qf&~4??̝a B~\4ۥF ]I6pKnYԲb\,TJte M\Ὑ!&ǟ_Bբy@XkAT5*e*~Txu6[?ӷA,oI<[zY\9+9+WPT|>8kpY\1~KQ~ \9R"f`ѶdnUjh=P0eJ,S3΅3PQgg 0@—'0(g/_Lл^!._M0ޜ~$X0?+ '̩2Nf=;5)H~f?efӃ?e5}3pʁPK_|qC/ bWM70R~`h#A@CY9#Z"q:Va2ѭ\Ҟ};+0i_,G Ƨpɪ,OGo`5 ̸o'[&];jtյC:l5zFpUp$Y њJ]P0~EeP_OxlEiϷo>ic,1ﴻhc˽;RHa1] /ߊ ~=|3ט-eh"z;*qL*M$-ԫLIV!K̎|-8l$gR~ShLa祦FedʣU6>'.)ZDJװh`x#3#N̓\fxz5cM{CPbݪ*ɪ_V9?%z,;y-Sx>5?sP: 䥣fGRD,QpQemЙ<}#O'Oe۴`k|b6hޡmt<;ߞCNG5[[2_N6˲%}H;<8huoNpS˿7zjvn ƻ-bȕώXk>Z(}#^}+'>"U-I^1'B&4AEsqwܜ4ARTE8USI{k%pN4dhc9\2'%d >ÄzoaN<,QڎMѵ(%|'ԴIE[RHqu$~J!u€"ҔU>9.RS%ݐ M#xH,&զlR\=H@IcOI=k^,>YЇuMNԵHT!% >{/.Ic :`֞KERBvDחjCvYviF*ಬvJ1u2Da(4VH(@ Lh=E"3|5gx"f#nC7!vD `L ӌ/%RC%;&|t X_ `I. pqT B2~ TtgBQ"N\ 0+x3k^ 4#}n`6d^i;VWދ0 ]Vkx5=+ElDjzFӈBU3_Mr^KY&)(ZZ,  ^ B9TE5p4MdPgm| 9_֭]b_*"͘u48U ULl^v')Q"/}0|g7֝ձ˃ֵtjA#սu mz@-Dxs:p4olK7#+J$W{HV*U4(y ~XQ|YE * <@E&rZ5ȼ P>xmBV58 e8myq{'*`P#P܎m@}KdAX? y E*ΊfmZA]֊NO"cpۜ7w. O&!z*`YeDM&ЈG!߁]Jrmڃ^Z{]4zD :$`YwQ׀R"Pc*2{(%g6ܖSBEk@.In<P9.Az-| blwXC-D;+g brogcQ0%MʀGص6Fuf EDiv%&PTT Q8TDYUQõƢPaVB]$1dlDH!(rĪM`k>Zl?fHgYtg#NGhBFK"-Ei(5.--Uр{9O j ߬n%t{$lئQ}%e Sa![ b4\6cs;<6g0\ 폥ji/Nm*dif@0uvtrhѓХE8I0=Mw'QEŨն[SQ֚Bd8N֞s7gSMh`Uf. LDC1 `$+* iPb鹓:hI_PFU~:C+<!GUQ~ՋAgrālHM >`2}sjfsxGMZ Ɛe AL^_j8\uί]@]CSiet@0ع\Fku@ߡ): u@bX: u@bX: u@bX: u@bX: u@bX: u@bw2"/xA: 2X5JX=ꀔھu@bX: u@bX: u@bX: u@bX: u@bX: u@bX:L@aJ8Fk|: X뀾GMu@bX: u@bX: u@bX: u@bX: u@bX: u@bXGCᏋŸ F?du@ߣiլbX: u@bX: u@bX: u@bX: u@bX: u@bX: }/:1nѫYVSjmo_m]Z]Gjqy\D Bv#<9%GiQw}8#Ye\-Gٖc6gdwp2.QHdtWIpS iv4*JLJK!yuxt̎ɭ @_"3ޫ1(S|l1P^.xytx/ߌ[¾~fqS,ӾC"r Qn˞ݫ1QG%cč`HNeֹx/5p~kt(ky+c x($iv2Wü/ث1AWdArlATA;K hpJ.kda&{0VY2Y8J"M<oګɫ+61{ʏ}uxN6;Nf|F7:C{0-!k'#x=?z39z?.s?~ϛ'<13p^#w;Gؾ◉w}/SGѓyG{|ݛw~ q-qG9?Rs( AM^MnzƷ>伌- kӡ捀d%>S$ÎcoA$J_UWWmf0Mn-B/FU5}G VZ5c eL9 A}`6rXjnD3ꄷݚ䑳sP%۲ ֤a"z/*[mhT}FOAS fL2a4k;m0YR?׀Q9$ʑd2gZs\rـ;xO||o~O3ҦClk͘ڏL۽xLYaE>ܫ=R9SM/"VK m6 Hn Kmz0^lXH {2o~0u8Ħ)}u} %gG esgpβ˛ q,\G1%`W(DlDB:fc)B;ĩ1x=t-+% < {F Z 뚰ބC:ZZ)]8Zi@Xڕ.Y`jv`Z3C)hZn"H{J5^h6BTsѡ m@* K#1i RT`0L~C`LQ#5g*cMRN5fgʬa{W֓΋q0ί#ya]F.=  xV3 }b )}b,̳"|70 IcƽI³~=̲j>HJAϗȓ""va'(IeTA#LMP,l.Co6%x CW@\-6`MH8,hRRL^? i&. "NI9x96> N&|&P$,>+4Q4_)1St4zF`Ⱥ]*2|ix0=]As<6ީj͓{ ~-Tw~*nXϫ l1ג `rn^횃Xmb^xh90zjʻs$wt4 iFa֙q`,|<~'zsp1_r5Jգ/4jӻKrYrZF'iq`1 >=K%=:NegQy;s9^|uw?}~ϟ_tWoO_V ̃Hl~,49{piмazhCs[W~q[}q/%0hnDI#`R d]ٲZ5E!m| b~Q2WuQP6UU*D4E7Ǐ}WlƸFD؝ƚRIw*rEh(h`-0tG$`E)v8*//Cr:Ja՞-u$go \ 'Ơ"!Gڥ ꭊڃ%ʬ5D Qʾ,dƈZ _/ww ࠵M)*hzrS1xgSXՅ{Fڥ=>eO$Z^>}mߏGv"ЧEX%]mm_py!-$.^^+N?ƙCZUڤl,QS6>%A912gAKW=k~(ZOmQ4l=i%`'=)[kv`{օ%}Ox>W(>&):Ui?(_۵e譕 ,w0s!,W:ZIB:zG6h nZ~V_^fAs2 !x%F jgXHy\*mΝ~`rnX yޱ^]b87 h=Krccɧ#\kjF2ed\;f&7@xT) a 6b4\KDu L57iGsߩFiy(qJKbP𷢔Gx:#$$R ΋ȌVbiV4O ߟmŤ= kG/C@V GG7ةhG(CPD)l4ZI5tRߚɼBX7E[-@hDPIF,ž̟W@eh3ҁ[T$¨=RM%Cc.<] %cSW{kpS!) 3#&!jXc&Bw%,s\{}vGKr&˕r,g9(O'`Ͼ[LGZ Dh0"v4} 2p?~_+WVZާ'H.$,*i+%$=ԧ٢,K?Q*u9ԿfySzYn G|7U*]Pugxꆃ*:*eM ؐ m3_^r|6pfxP 3IiևӢ<*~,3/txVO)3toJUCŬx]d5|?4>t Z$Lqr#9IV)! C2 k{x1y2XLsFM !8^,e>Zd$כzt nZUIT72bf~4Nmzk~4(;TAzӭ?ZR1eJl*1RlxE՘ rɑKsy_^^^ 0>}(PKe0Oͪ_̤⛬\B Y^ k[Tsyˆc3k&e(5MZ}@~Yf2t Jk&|Jږ7jOώd__ZMɢҥGWaU~Zq[HU tP),u+za՗5 H\.m,i{r+³2 #HvPJ~ _pȪl/FTj'~v *Z{f5}n7`2[KfMn:q\Ugohf~EgLxlfE;41W]ܚwymܻ-}->\GW/^U@lR܆ghVrCˎri/ ,wȠvਗaXWtnʩW?#PUqV3\ws^ىp*[\BPkXNrʱ +J*=eBnOZeb _GFD [穰;o³N [;S\uf\E }W3]ktv6q,(AI"kQTăH#MQθHxiά@x׉?Z4 {U#zlJ;r(HS]\0s,O'Fu/9v3|%gS4aw(FmOÏM /U\[}fϺr3[ 3߷pfR= "cw@uP)zt[Ƣc\g8kS3tn8`ZTTˌKJgRj.PHbh>e{a*FY^14ew8*9 k=M^ww/Li[Gy7X;@PC=?AҩD3oY6^93 `=΍DpGU; `}A>-`(ךJy1, BYI0;e" TJ9DŽRh<[NHDdQ0A[]io$9r+,%mc  xJIUe5CG)R*3d2 k {kȵ㙫Ug:+zhC ~*k6e t淳] c|4J7:|C8^+09o:yHeB Z,TcYXE`i&19堲O("M7O}`%MJ)pF&5u@K@rrQ4W1%cdSJh$9*jթ*mBFM* Nd|BrNJ>2ELJ 74܅c帔[FfVON\t~4>?He*BGB"GGK1;O.니aG:!.8.;V>A{%;{&w\KT~oCko)v< wqGe?]{.㪂bn60 z/Ҏ>]jI7|7ԳUM:$~gŹQyAكt5OE !fNI72 g]nR$`.'7S ֶ(cM !x2&'WlsQ/_;z$j ?O B:؍ vU-.{DÁeztgȎ]=K}tnvR俔RvtINdFGto)rtn.-=[tDwgOVr&IJ(40ɠ5Hee®2=[Ēؠqm^ްVV$; /+B3#t`G3 13('4?? [׏NVOC)0#ɲŭ/nRl)9.v9؇cw==\Jzcc ѿiqe9>sȓ]k(Bi2ӷ(}tEԲt9.j$F  2pp轌LqiɀXue@,Zg[8ㅐ:3-P7VΖhOծ"K Al45, eyJfƃ)}FQ°dPfUhE,%\x,\ٚŖE5 Cfso.&!SevʍtXٮAhC1J ^G1dF9t8ɬICS꫷捋r #mi앖hTy|H!Ȃ@<ش( C$>G F&-oYwZ{]n#\eA|?:XpaXg- Y!JȮHD<&%^WI9$hʠ~_ge䏣?]yKH]śzRc:l:z%t.]L׳> (jqh-em\  F;%d$ybU9%|YAA&s)麑Ɍ7qV%˘18D>lHȲ ֜ˠCʁڑقY__M&Ӽj|yWv`^ј+1KKAϻ.ńzα]<&7Fp<'4oʃmlML뗞l/6A"PBb&*gr;Ե{_fJO (?~}g3`ek Dȉ8V`N4ob/*B]cD6ԅ(Ф%cc"8! զdP3 2Dhuf"GwV8+ҞZu4Cf{Onnн f#[[ - ܐb"1nv( 0(}ѓPW6֕\[~K.]B!?!K%D/dv9BlN5Sub5%]̔NJem(ޫ[]u8o~r O-YS_v/.6}*k^(eXyask93B3·bۓσJ%M" #NTVZ( {hP겇j}9jq,b* - /MfMθeh6e_ɋJVƎ綗IXd|hY*ZxǙqx?b1&f?$dݭL@*Z$$M-brJp5#S><tg hQ =a* ȍlVg|@RI1-R1KppXPWժ#ޜ&|b\:jdW+6=v2D< )5>h*;U.\FKYYPeL) XÜk]wjV1M,UaZ+E? jD{6z7NDy2rS^%2HQ!AGB4Ĺ֢B)m3!@6b%u`)'"ٲR.Mk{Rh[h.Ve.ɿy+8gsrVڤg9KH\9A*CQ Ar$Ǟ+>C3 Y#idӒ[C .ѐWqϴgC$sMl6%'9,poF.qS̵+E?ؘ7LQ6ݶ/o a}Yoiܵ"wet;QQw c I׻q_:q~ǣS5,q,dV b'iguV'2 3uO a~*D= L>[+IKkgK6'.^?tˁCG5u.4;4h7I0?_3LKyn$%wոAOw3g|}冼CkqGݟNUETyyS;]dZR6Te^'k螻5(v?~gY[?ݵnyL.3c #/>[O3;${!Gd> 8ISK<MO?^/忺p9=~$X||Aء,giFxAyA(U8rĔ&BVZ9w.|gt=#h"m!&BDHi!m"M6&BDHڥLlӽ-eЦ6DHi!m6&BDHi!mi!m"M6&BDHiM6&B*en&y;w~wɯ*M`W}g`vS+ns9e TfQFR/hJF:Z孎%HK=XyNb 4\j\BiQ;FB&V#Q[vJl9ߖ~e){&@pHi=k8fӷXUNmq>99^ZqYd*YB$ MRi(yU୲Jٻ涍dWP@EUyxU9U!Q*qM4AYO&hPd8)YƠF_a+x+v fZ .A- 3"i%1  Hdb u=TKƽ 0-X whVՒ4 Tiv`XΚ:? ?tA"SEKBX"Y_~Yu -Q&°:f +HkUHpk1jF"⌀<-k538Bx ILZ=#_ͪ:C633OC6D?87$ y'2Lm$0IniPbd\xy|E3ϥlT;IFבG;ϦGiE ; A#L9K)(ޝ׶夘> }W꿕@\tiC(XR#Ta-g/ =wQ:`2J"gnh|U7CN?/ȔZ%XU+5ztx(c,Q9qS_h W)t|1R%0y-=S5:O5gmzīUu"8[C ψ6߻-]p ~I|p fBM3 8gn4OY4чQiN> ]]&0Zm~ȦY0fɣjIs`q >'8N]wXn*'5Yo0= ޽O~u~/0Qo^x+u+0 6F777pgS󶦆ꩩbۧէ> A[bilo-כ?_ O¥/ujP#}l`Jz3__d_OۊU(zJ&hN Q!vM)Ne/7c&d'߹$ }"WyM L0B `s!Ip SV=IpU^^Jcr:ԕmw+!$T!>H`\PoU,Qf!Zh:|)ft,ǭNZkqv\u#gxV:fiU]aBw8!J'tƧM'Z \"N1gʳGCԀ.)Gij#-E5XdcLJ "!0v3f)QUdG3ӦݗhU*d =OxQ$1 ,%2\&Smi *)MEL;;>zT!$o#FyJԤQB5r* ]Xq KpE"YЩ{m(DOVȢ#Yg pr}8L]}}0o]M p J!@1E*pEeV߭Mr02l4\rm1ȱ9;nn36g:܁sBUL{)'{rKseK2pezMZY(ڜ-rm|SKQ(]U &uq56s7n.SYZF]uެp2[v+,<_nZ牭1^i $#Z{1Ɔb XeqC0u(`I]}B6[)KoC=gWaM`\*oaG'Y|5,r-OͿAHvkzX/%kzuqg@n_ m޶-XhKռvH@HKڗrhyml+$T΂#"ᷠ֓(Nyc((O4zfT[B2dp !N Tyk8~>M;iXf]eq:+z_?v|.ՈJϼe\ lTj:Z"Q1ΕPh? o+[c]i# s<^Zlɶb̛5odheimtfM]ޛp=_,c11FsВ8Fڨ%7΢W ymVˢtدyO /6`zFzNjJK-~`}k^+k)c05l:-R\QmQ'($On{M3ac$:D0Lɹ`1X>F{ɱ6$Qs-9˝R 33lG\Jƒt/]ȗmP uiJ;Hu HER. b5*8NUs-c1KQ\x LqY̙C:jXXLjAuYe'f#=e`k.30cF@AYE NVFv"r3@\GhϓP@f:/?W-"ʻAA=Jq8-_Hʌ:^3iY6^93 `=΍DpGU;b1~o-TP-re0Q<5K5e%\(wOuMzI/h|IC?PoO3aO#=}t9|4= ?[O]AALF CRӂ/SLT`MB5Uxa<'Z,4t :cř9upJ3O4JczSfLϳLXaMqr/a dS(˝vX-RG&5W5{9A [ƙs^~ru̅Y)pz0D'>6TĿ;PuQk b:z ʵ|~ }+RPkZ4GHrE<ϩsQe ;Mg^K~z5h;'c ,+w[bBKMO>}*)c?MhK,Q P)ƎY 71aJϋi*@Cta/!j^d:!"1ޡSI JZ<TGaŒx_ 0±+.Jq5\PΦ yw]TSGH]C_0bmE ټq%=YŃ]NHCS^  OX{Z"Ssc(Ay)vpQAe٘Wޏ_d58s_ͨR1Cf o` c<-coSwQ =դ8MZ/|gO0Nγ7-3cVt..tLڋR0@uTaC47Y޽o"q^!/o'ޝ5d$X0 _´{tvڤ g/_°pm}O _f5}M o.]PO+ B9?^zlHэ15E P(*Ss Yx%Z`P|CM1ƘTdL!^]˘ɻMQ/1&jlw5P1+ltQ?L#+JQ|ybߧcAo)^홄T] @Qe.~w6(syfѸyQݱϫ{& z roXjvi7a71~[1 ŔMTQ"}eD_MXӮ[f1hyzv,ϼ-Qm' d$[+hzGL9l@0{nť0:ylZqGa_QKZ О1#L2eZ<Ӕ)m,V$R0F`)GN鵢xaR:RE5uKqtobڝ|QK{e/-£UO8zPJIwI]f2%J@] ǃPkCZHv!XE5Ü!j7r%!J ͉KCgkHӐ"|,sv9'%ڍRw©b CFD  f;o³(LZ&z4Lq^0 E XQDtTZ_._ w}tEW,#e ߬{cbs'Zs!۲ _">Rf[;+.wdçh?zukup{ G)WnJRBfh1RjeUUFSBM SqIצUoD,\܌He c|b_qٞf` z&ٻR<mC騒u=A#Md6Iv'IɫÂ@89cI â޾@v:,Nw$jXu>]mٰ:[鬏_V/-OX/kW|٣z_% m/\w>Ζ/O :1WDru^>Y*ԅ,/+4ؘz՞\}]v%2$ͱQCK>z@t5 cq ߿W9jpùE[]VòGa3=:mt^Vʡኴ:{yx /^cd鷟7V? ^9gîemOocC$e*$"e[3;-^w &SicRL>@ea^Uyyyz ]__J7 P/{qI{ OF[3]NOɻ ":<-ߎ| .y,3t9dѽd<[֏(%z8v?꫶.+#o"B~:yYxXx=zj07lo~V1jԀGԱ@UᚧoNe`mya?TdwGVv#gĨm-M߿vQ17u;yq9&su@/IIФt3ob (%񵼞úhdPdT N 0dŤ{Kyls , -ZMjcHE4dTR9K-+tt*) z.78BhvPoUeQ\L!e+2uR"C-wt%;Ӳq5:B cR;C^bKڨZ%a X]>:kbTS,⩱L~#L6Sc~ǩq;l8\ shx`3}u|ܵ4w]DX->f'kc_g]TV~zCi?J瓱"{?;mʽeh﮾Ykh >ę?H7;Q i3?!?CA:37:b]7iJ~#XT?>X )_ȟ껋8/?` zo>9!_~~]k/8+SK~z:AYX&w WО17$sv+6ζgo&fq`|a=C O=T׍k6|*->mes _>`m v8阋՚|tSA {z*>9%T7'm;~F;A=]B|Y.AHBd"5k* !ZG8f8d4myq@{2|dPVZUvdnn3JY֏oCW2ૈ;ؙ2dAqup%Vƚb|dm#Ŭ:"JŚt+B20b50 lcY/ 5ŜhkjVJc[ tPU_c+`S:!mg$/;`;V^2NAV\\"ά 1 c< JÈ*+wЬgũB F9QSzXz0ͷet֖Esp"TFdoo⭻-̭L5"ƂZ&7eP¦lҰ B}YZk0LĐo,o˅hP½Gwv'K]W' e{1NFkaneE5 QMA|栁f=+[9Eo\a ~)vVxY:35W#s6f󬑲zhh햡@ƴ-z7||YiF٤IpȀB尾T2FBr v֠D=Vd*q@:t35di3MƁ)n1C͛D {AP=P|?y k)AߵNf aR97yYy)VxyC*ahar 1"*4Fv:JĈj&z1@U4hUp?{mBmT19cRL@csـܢ4SH5k+HULJrT *_વ@%'-9 V#;EuCȐhk7߼@7 E<8Ci*f T09 Xv}2A'I!y@B< !y@B< !y@B< !y@B< !y@B< !y@B>Q$$x>ҳ9(SVr33wCPyoA?3vV8v`T6$b [:$>Fk B^7sɆT$0+EK0_ǁ` $OJu, bUh*A/wK碹TId2|$ 3;9IwT_Z43{)"ú}: g|cAO|s򐧷n*ͳ\Kʬ5OO1bB"bz  *F߼?nwmK?_Lm,EJ4wY,YȌ$<R~8zmnuT%QvxX LÇ3igu=/}*Reǒ&ҳTĚkk}% ʦşGKvq$TǷ1-|pVpk&W]Iܮ5]V`KV;_-X_= _8Z~\*%wg%R Gt^awtqFͲ_` v0 a{n9*5U$MT_n?GVD_Bg20 *j]-._{3ra|n)Ic9eVih[uH}>k}̇$z 0tGvuReGA1yTeT?T^UZ<ʞO*%Mr|7tݩnk&^:gx;6.ƗMe1ᲿJwHgKum#xCqKvy'/)ל*AIK+?qA.:|'^381XJMsC`OJTѵ!\ޛ*pPԻ1>y}^\oҰncc/:Nw,Oxwe5Pejn+4򢭡ɱ?bPGRNvYW*^ɺPbP]^# &{)a^|7clْRwh:%Iu3pmAHNYlͥݎ[ρ_}7$:e] ( @4vxS v\H؝yb{33S}Hk6ָNz觾Ex(O 4Iy#+AFYe1OȴK}DQl" MJ joem} J(+9B;内(c5cQHTѓpҵFSH.w5u,mՖ)kc J{DRUuȤ*ZDHD^cn]E*JY$Л/uz#jj~Q*;.ɐ"GdˎWv!<3 /g/!+i`䴷K7uCӏ)Bt8~W.iSj9 PĉAppgWfz V~͓m8"c3Ҙ;Ob]r-%mj~\b|A@Έ=~V¼v) .MXT!Ɠ$:9eb30H 8҄Ex MMʬaIxm0{-ײQSf<:.vIk[^ˋn#uuTFŗffGNőΗ{?ou"? 圫Mֻp櫴4yqV&q?&`pp5&|+Zv7p:)pcWNǯKtFPnYkn.ްnX㣻Y+)!Gv'ч&SzOͦcf}pV:@OۥYSs}*+[:QqsVx:^k`] E:ˮVzM9xzTފzv"4B(Vscm3ߴj4y7=#ЙT'TKKJ篪J-f-4(J -jtrQ(u<O18wYzܸ9B|Z8>5S1'y Eqќq&Lę'3 )Ͼ*|0LYQ3N ' erM&é׆)x83zę5g7άG.g٥ؾJdq6ֿ*"T2Y ^}"I_ߏV4RFB/zM˳cB$y83ks/cIH6h=>xc?c%H`Ơnm^,q*Ig# I,牧8t&Lʔ.IbY ?{5g7Haw~/'SB0pMq2km+@'*T4e 80)LhQ, ݸtX\`gw mm Jl&Xݸ~N7gtOt|;j}jer1I!N JpYdBij7%+gxٍC׳4oyQ/,[w]i- F+=#̜gOcd r!1ocxo4N:/)kLr-mC OW1z%P IȬ~f5#4t[coՁ*~L%,5Øf7IbRC.o&Z̮Is`HCS;l*ө p32yH#^$8bqN8 wL:s36T53!/O^`'ϽAoy}u= ez|, |@, "JrNjY)Y`D9P;c$ךY-x&DO*@M' ab}9S`C?(k>o6 2ڄ1z+9aN~Y;\}g9UGn2>z9if %B/M'Мe> f3$1g jp]6=ܦ=NkwH܇jRkm}*EHD鄰m{zsf=)U&St̳\Y?BiLF'==Iltؔ$"="3\Pb2/x.s*KPY3kU 掱2~JlE֤jN^4BKZkm#Ye>.6#`A\ ɇ{"@~TK)RHzS=|Dٳ YLgTw){s\=zBj}qr1UIib@i"d`!DBKK3&;ɤ%*d 8CSM*E8yQTֻ6 scPQKOJ#5pkF l:[6ϽZLRï<,4֪4v㵂oH%CkU6+c3 ARnw[O:p,EST9jGCY8l`!“+V!#dr~䁳Ȍ4"|Ҟ|.8g9!Z8HVyLSsF9$%C !XrQ z3=D#j#7 7M9oݾSdy՜nx9u_C$}GʊWi?,| 2xCڕ(Y,vg]5?,5ػYanOI nN$i72=ޝ,GMzu3Κ"N^uhnKcqr|w٦Ǡԫ0 F@A(WE6h01܎i섳vH*וd d\H?fg*w^3vqn[(\$ LlYƁ;\ <}"1zorbV?$$z ~?hz-+"}$-Y_-Vg뻠na*&1{Wi6CoT!5ӿoh?/M7 "k2gm?q+b~twGG6l6L8nv6~ądQEk˪b+nZuHO{% eU'+Ԁdt\ Yצr;AHV(c >/5=$Qv8u)'$cF8c *jy  hJdk(Ie.ekC V3L+qN|=M8ګ (D==ko= N *>3Q;L`rX$36l̩":UZǜ8xLAٜˎsz$Ʋ,U`CL ),%g24:wJӆ`oeXm8O\iOB)W*㪱TzҿɫB$!1SSL #zIh4L?)0uK"0s:pM@fUR>)G@JbTvׅzfcmϮp+*^z@W\+*|N~cɩNnfz>-J#owɛHȢ>..{ɉho8'f/PrP K~ RAVz5K1cIg&])Wӿ%N3G=-2e{JNVZs 6 ϶K& ]UJk0KΑ $W/BJ dcJ`0ē`%_8NHH ґM}$QTVjٯ+Ե>85RդܤkmF{h&|lQݫV݁{gק-B+w뛢D]\Ȑx.RӭV%ǧ >e7n oU&30FfЦPIJtJsi2(E&R DNk8uBJ#]0$j E,^ҏ00kjُ[Sq0vxۑd^1|un, lBЇlG&-EE傒V2iZHZIRDp\p9CSXRk -S~]r1.D2>Bݘk:JgHFodR%]7yq_Xӫ1hyXM3|Ǐ,afRokFwKǮ_Sfyo/f>o&M2 =H:+%2ЫQlc_ !ickElm":BȩJ]}Ħw>?biz;o_ץ)fMiXD٫?F'Ѽ ]rh9~xCYUp9gVCYWo4*`>ypٛKlVf͌w ~yTƸm}[V] ΡuS҄u%靌'Un_-￴nwwOstfbxv?;CjRe"ӻx?΋z^k[Jq-2=/jo@q]`;_MʪetA:z9[ۯ /FI}bjvDR-[ Zh|օTڡq1өy|XKJ:Ȝs)F  r]>CqSCuؾu+kzB%~ilanP!{ e$U  T4S FA.IC%4[`Q=cSN:J eM&[rM ^6/U=+ Y@Wz3n |{22(@mtFrEq9m*kP"'4BDE zF 3A'BtY +&-m&w=}^|W UJ<k%+ >CcyfX{9Me(#nU|,nI.=9Qem.h0.VJZNΧܖ5Sf૔8`39 5w8NTyG3B 5oC }_''ӻQ5(7ݗEɧcE(Κ{N,[@)a(d; BZuGCZ^ƣBY$Zm eLNl %_u)Pd+d2'(tApe6` i]: M! U8>&ScIܰ2Ώ67zеNC=AtLI2%[j]K $3FPj IDs堼UwNIw2G7y>2O~b7߼cpY-Th|ې9oJy[34 y\YfiY,3?T+L-I*kYz.s!K[[#'^ǀLX+[!!$hMƒ eʎF99KY`R2)Qࢷo$\O\V7+m\rasGآY}rY)7:)2gT-K!ieq/ܴאނ> MS1YǠC ,JI%[6'EaT jG*ׁsVtx́݌Ӌ--M:_2G@l3f03]io7+ cY?N$w{ \YQ-13u%6i!٬S$)cP;꘶Y-C#.mLX vC=L*")9`NF03)qXFI$mp*n$I4N0˕&9[VKJcxֈW na'YIG}$E)T{%!,Cs/9(oRaXRN^U~"@Du GǨ3 IS6SF")."0sܾ0 .ߖ+ {@3{2jHrLVZ 8 lzNC ${kܧT()C13."RCs\sQē^&E젢{ėTEZ)`j+pҢ "a`GRP LQ  t^k~+ P& p8LhRKťHןA a&99:L(LI|wO>3$+T{Y63!{ zuW]_ir#E)ݨ4٣ L,m<0?\Ar0y{tVOkgLkuƛq}:8D!^ 36 'q4<:U߮-Pr(WɲsGa!֒-1~yKfH{36>X>"06*ĥbp<\\pZ[%VlkY )ljC#a_ӹKU`bVgjKJ|EU ' g0w>pû!?two~ :;RLE+~EӼiMsS~v;}v/PvKvsk@dfE^z͖լ ^0d'>W@lPq-WWQJM4A+ؗ&DmWHĭ$vgM$;"4XnR2-#Ɯ<(0 (0ecq K{ynm%.p%X#DiP ꭊڃ'ʬ5D QɾN'mlLluwy0ڬ:VGw:<#oxtGŊ/ -D:e33/R2ɋtjX'ȼyZp?&֒UEZ>WZnp6qY~>py+y/o ūfJϲ7`(MI^렴 0rNBpF5z\0s,_g.%<@rFϵ,wJÃ&>@ym8OШzk_A) OGOe*UZQuo LG1Nե|6:9Nr-c1#Iux f1grbqb#Vcgyq`=yn),\(wp)B0cF@ஂNLj9+j$Y0E^qq5/xw/aYfm3Ww:ʻ CGz K:z~R/$mrr,:qhrN{sQU=(yDwn#wKg%KL"*Wrzl%ZS)1} SPVR)Eia{'qt-wә[AmGs+:2,"1rbrX+"D4vMי8oY@ )!"tOMt|:O84 y^=|$ v 1KWjgt&tT{&]GMUJ-Aω9# }g8s[qAGJZ4JczS$ZRā+6VXA\tL١!cleqE $ø)aLgƎqfg܀].\X{sk^S)T+և?ߣ>T=`V1=^}$ v$_ iw'xĈU6!sF4L\{ޭWE p9G!ɔ1ʃr 14 rM Ge0E4kD`!zp/!M~0 CcH;CQ[J9yQFaŒxO 0±S&Xo{ pd7b$oF յKJ\]MkE" ĠৢGs, 5:D*a}yJ{ӉIpݸQGQm5ñȖk!E6!+ ##E à7ةhG(CPD8^;-L_EJ_*yhѢ7]Vknh .xR*F^/gn@10Y.Zƌt`"0*zOASQK,(OWBp]:{OoHlH/x6}9v>h2oKY./TU>%U,0<:=.<dvR߿- 4WA R$feabIK9}XLgtU<{Yd zfIp:#JUjhlphJdk,jf ΅Kf?H)k9-7?T]-/@E= Z彩_/07|wQ/x BG,eI۱4`i؜R|cFP/+M9ů/ ]x2˛xm=57q| ;-ZJfl?o?R l׿[ξWe7[:Xx|"Ԯ+$]׵&5u/sO |yݤ], s;4-T~xCWaA3kLsCQ UU$W|dήhq*n=v]Y[HBtүuf.^Mݦ$8/`Y[+"XUDtTZiO`}?wGmsɱGG([޸MO'ߟ g;G^J4^FG~xøDYz-[Y qF Cb&adfѐcFJsE:"~> 8MwjZdրz7'f!PG}'~K0@c4N^d0za@1PL_ݯ+z<2PFd^WrNXŴ#>g\qr)TR!$;uvv.OŶ]w896GL X4CEQKYdN!8 VR$eoFI*N/ \R[ERcib ஁dWt@ug4j*7#3kyNR H.?A8!U;wY=ftdϚ8_'0 g!M Yc߳|dE?-d:k.2+f"{ 3ry}wZv?f(YI).[qF];ul$Ns\ MGCwNP35zWxԗiaO^c3m7AկC][8Ij[}VY5`Hi.EC@|HJ,1>>Vˎi>%Rx PTdod>eŧp:(kiOO'%ZLADT m:8z4g'̸?ONɘeCDepY7pXu 솩;o\'"CSa-k'Ñ I^6 [ƒ)0Aj<#߄kXc9` ̝ ٻ6r$Wmp1k\vvg0` m!Hr2ɯbŲlؑV7*>U,>p'\mxLl-Z*Bmpv>A#`0O-?y KW2-˟/|y/+ٴXV/OA *.cGHszbD24x/%Р?`:<ӳY;jp_cyqfy)-K![k;s]~QI' Q%%p7B ?`be5>re}Ⱥ,>^rQΛ޽D+թd3eU^T\%AJ^eV3TN(ʘBϋ<H# G.୷@pBrfXKK2n֜{.!ܝT~ZoRba#5h'\D˼*f2xmd˸ N8)_Ft]ٚeRo9bb:"X( "XaVd-01+gJN:VmWh \ocdcBkvYf$Gq99ZR [(}4F ;c؄Sxil~.ڏya_{1a.,D/#``5N&qgӚa_TU^kGg[PT) 1Ee5JeMs9NgY@:*)Ka ;$),%g2;NjЕ3"՚Sa%ydl1^BZJcZj&0~2)B(=hu\`h8v-d-4=/4u`8\M''I\NP#vr]+SkLOpǖw.O'>;m&ZH$Dc3 mx^?hZ>6 ؿ ),e^4v~8 S)TvMo:C-8<_p™7&wxއќP[H2Y6clEۈ]tG)[  g^vUӰ/[FPjv㲳V`>\ ͗v7Fd)Ilr<*z\0I6|N7];x-qΐsHvr\=Of]V#6m)UZ\f@-  f%w9]~❽y9|g߮v[954 6dlʹ|Ow/=<9x;N(ҽq`ޒvZ /ozk;y}n/R"whk0}2mδ4 .ϯNi<ڋ '7^ӟTp )wl%;]{|8R36l{h.ѩb:ĥg]v6gGY1O=<=dpGAg⫌TMO1 Cbd;zIh4ξq hF0pF3o'!0R(PbWЗɃ1dcV(UdB$! !XSf!EG@C&!ZQpIJ)dE&dɪ5g?liq}="T{#n)UFy0\ğ9 RĠ^+BK[YΘ6bcTpHF\LH!Z57,Ҁ6Gb՚#~:CzԁtYtLjd_h*>9:{u. BEΌǬ,vNbG9C[IsVXՏTo9Y/ӣ#ltP,c>fJ!dTR$BBS4]#K0BB~'3)=W[B*d`PM!&׼p:/=H3t&)l2ޕr=]uyG^/Kn 2ݜ6~k s?]x\XX4[H^2YhTZ娴If)8h*E¡\ՈtLI,$X!0   V^rEe#֜FB] c#X{/dkBwA;>vl}lQӯt}y[֛1K/ncg]r^}x9+F4m9 9Ry>RgUHFX MD2ۂt}Uhj7OЦP Nb.$łk.qnMdZF.I'c2AP>u1LcшvYre&`1 #6 rm4W9m/tL0nON[z͔F=x07LϾ0nL'<1Jv引?+YR4aFF:/DnX I+{u;uru :"̓Js`# ΤRYSNVhiOO:ˍ3+t꠸2GDe E((r:J Nxo JKo$ :k( oreZEgMQO@ 5E:dF!Q48G)Jqj8Ob9v|0rR÷TӘm~h /h8__~~ûw?ݫ{O?Ь'F^t6> >=w.W<Ҫ֥zK/Xջ^t}[;{|vyjnm ?~ s֠vգ,\Ϛ *zij6_J)(zrEoKb? PICފ1H7ă$yX.PIҏ6++SKaV;cѡ)k3`E'IOpX/ǔtX29mK>ETV@fC\ĠQR%De^8uLqvuy,{$+=_1;߆M$~oªtyJwS3]T LN>T0%o$˹6(̄!t | E<#)ٳ-EO 9a-XdJf蒔hYrC{WgMlYTgO6,;9K a30L p8nSPf/ɢ=cSN:25o3(oS"5B%t5җ`2 #!0H󨐱eB}A1 \{c9LpW+Ϟh;'ȉ L!"Mj"r0J=ZfEFIgR'b&[=|C!x"yNrZH::fhg$$cR ?Qc+޾7_PvD\ҸN7ΧS,U(=@%V[頿wvmOըeiyyg>PKGp_P1(T>#!Y2)h7QTK)HFr*pC:qCo`P*z9yj}|i GWO4#$l‹淩[/H߽=w=m~44@} PzpWPh^#@7~8?T&<^Zm綃`mmVd#gBZCZ7w/BYI Xm#C@Fs& BR 'N(RLF)lA  \\d` ҆"LM!U9'NiozdKŕsff>itl$&ӕ64t/ʤAhFfR)L"2Ƙx 2iG:sy:dbo7mw*jTȾ͊糄ǛD?B(ԑTo ,<'i*bb|qHOF?gڡ`W7ם05MnkNNܠ>;ͩ+G+8'<6)mbl%dEFsDȻdz9Ob8&5?0f1-#[/_,xj:m(γ+S bǡΫrj$!1SSL #zyJh47RUTf(e& *[I?+^~JE!!%ya6Vj<.x>)7!X~cHaWOJawBS=numa=٧; ?HZy2#IN%CuG V(\2J43BI@ZJImTU3jkrJ9Vo)K.H`*C *1ٜ$1sgRR5c5r׌J5]X3Յ.4.<.w3.xýӛi 'ޠp05v(2Z$hm x#΃TMLE2 F 8]Z{> &CR)K&mǬ#J9bf9k0;&hjܱ֖֖=ݚv@ƞK& ]UJk0KƑ 8W/B}*UVkx=-a TwL¬OUA<Tt{y[ 6˼NvKoʙJ}2))*NW"Di^Hp.*X(VX+rǧ >e7n zk(gy{mJb\ kQ: Hǹ5ek@b0ARHbb FCe:C&`1 Wm̵\5rGs1=slցS̫`l0̸I=ϵ;(~ Atp9 P%*K[B̀iz!rRHZtoq΁-_ɿ'}ɯ&f` px?zF`|1RgT6"ʊ(Lr:8J&ydORY $ $8BEzH* rt]HKUO}M^.;AliKpjxaؒ<-c8>?ݰ,]_Ӧz..puĔ9$ʘ-X, IfIv\Hv)齑18m^V~ƭԉ$]0FLddX2*:k87$IPC`@?nBq*C ^a/u!j#ed&;&". G"b"b@ƞ_K۟kkEOR:钋9X*QM!F,<zA+H;l&H]R;A11 qfBAH͝C A\IF4ছՠԣd *z Mb~9/66q񊎪oR*^NXJ%Nr;0Bfe%TՎXtrʚ+mFLR'IpXߏthKSDe! $m1d*"m2]"KBiXS\gc{\'{,_A\zӝ#d^*o y0V;/#0;EtӃJA\/Yt=L eH͇ 53jW#$I6MrMK?_FG о.g;ۭz#'wzU޳65=Xlv [$^?S"iE ݷz.Q(P1 itWUJKBVY ;*4v4w|gqW*[TyOAcն &rOfbwkF9\=`ߛp>翑O:Pr1a6F-a *0bKtbJJZwMӝbjNkܣ}ܧ}Arq*>"Eڭ>{L$DD; f*T2!FMsFV;|t|$y{Gnkh{ N)|jY1l:(6Ky$"kY"Mb2܊V<_?^w6:m *=3O5[iRŨQLFSeR]>W2:?gLc`bҹV:F T/U&a={/n^9,\(wp)B0cF@1gz4‚S8jGFcx֯A:+^3t?v_^=m(;|Aw¾~\ G) ژ;qhrNysQU=$3<ݟj,;ȋ\ qhkMRJaw.J D;\w}8 _/SS:&,֗#.rӓto@l>MC' }RUe69IXȣ*^́P˜ő>C3{8 Ni扦\Zi1`Lyìt<+qDlb1@6ō8`iU"%lRmZø)acpVW.ۢm.,J}wĝY zPۨU+ڀ>^ec&RWOY/NUITs!,W:L'cuG7h o@T?W{|O=݃tdA C%>`%0{K@(eWBi˹T')mES=p>;ܔ oYrޝԠzΛvw0h>X[{76dN]2ed\;f"7@xT) 4/!X$ KHD˫q`#  $"Ɛv{N1'2(R1 33TA`cLcm/n5 n&dbqxa z4|2Eĥh%}s}g\Kf-}G!M,PԵf7  ;K]u]H/.@1Y.Zƌt "0*zOASAQK,(OWBpIm9vNY͑r֔^ >^pwهNqqo~tn]{3-xX Pv1>`dSx?gtտWn(w_ 0{|Wr %5jwrg?'ieodiayf`fW.I@16>Tm&Ξ89ξ\ GE@ZԹp9晙Of?H /K/ ȇhaR}I|xiv`-+=HV`~ۻ ]-BhHmM |7K}O/F3kF&W/MZ}߫~p(M|jfSZvǃ͇zR?e}srQTիu]Բ$Uλ t~YIf!UxF6gi.Au3< Z"q^?n vl5MT/y(hS^]/wfꕖ6͈nk'PU<aKɥuzow*,wdi ҰiYIAEV?)-G7]L΋}߮d%b#S u)j%(.AFMwq_31VU]zQ.;4fwot6,nuɍs?1AuVc޺Ǵ7"#syǃտ~`gNf%@DҳgrSgx:mwZHFB5pƨ4Ĕ3̦L #xP\ Ŧ{gFyutց@>W[F`DTL 8`"#ҊDJC,r N^@O"KCCZsM58i|9dF<}eۓӎq˪gT(>"RLt_mҸS`ưXA *BmA Qʱ "1]2={;g0 F%wΖcYѨڎքoS$7 c=Xw@"w0Og!mQJ4q; L2uoVlWoyM7o+DZ騴ʃC#My(dg\$Qk4gV 3`G W bg4RAlۦ3jY톿$툻T֡^z"KubNz!2Wy14<0ڥrmP&qx VqJC|+Sb.gV|7 g,St9jZ> <<vV.`kE%:q}s_%!燯>)bxjջ8]KI侔 fy`ƣnspe+>>lj/?v{ kYwdf꿖?˰whXi7JZk.&fGE@o>j i4#7Qqj";Œ0>se)Az) :T0C#,dAX¥L)q*A$*5(K)@I `Kȳqcgù]z\u}1QIp B8{IQhީuThq2[Ƣc\'L ofEeB^ͻ8)&yq--*?oJ\Rե '*a}_% 6Ƴ\U,ˋٸd!< ud%-vumv o@]7C&9ӿ+%<@rFϵ,wJÃ&-2~wDJ4R3Eyew1+n c`bҹV;sb#VUfZrKA.aXXm4ΩZzb3~u9=QH@*HZVI(|Z)zbWϣWoVjr.}_ε#I pX}x̎1$_v'jI(9vC.W6)]UoO 6|vx9[=<\ˇYb]>J˷fY0٫\mSh"o\2gs?N竝`ӧQxO-va+9xZW)7>a3ϮRj7b _qPT;)BsV?tNHG` Wy" ~Ig| n-)=;ۿN7nQNbl[-NwP[#iRh9l{KqqmwΦp_N /[ksWwh=t\]ݻ:r'L{⅟nN7ԡ}"ys`P>9B"XH#[+?=1k=e`>A:_Լ^snn>_xjug^z'^-\R/A(W] R j۞#sm 7ѣIÓոMA^g<xOl䷡ݵ8Zm<`D <1yUqe-ENM<0-^03:(mĎ.GE QqW$>Z)!IBfҰ3҆HY &߲lv\߷C͵hO ќ<_Uq&2cZߗ/%K2gxTV.ke:SRi]{"ǡXU딣V8gV LZF()]p c{ID\KѦ'7V! G4CbVȘD9BA-&Xq36[^{ D4G0+ Yyh`-Ǝ6%x4JHDȈ&1V aeȷ`s18mٚ|AY1/+k;7B$bnu4Ҷu{&^2kg$.iM[$kYǠQdğ k7MmU%nYUC Z9+(%) o*c$y ~ :־ܻ1MztZX4vݱ7 $SNl sf^C6a$2fD3! ]0,g p"isAuɑWm S,Xˈwhl92ieBXe#j c5 (_x@>0h-S!MK28n{̥5  *DT "1QTcmʤ7)KQAS\ZPW_gD͘Ϙcr,K0] Ac(JƬA5̹-2 -" ;T%cT2j) ERΎilX\Lt~Xeq@SмLm($AJu+Sx[Q -xSvՏ/XA(uYQ" l ɜB_VBv9"}X!oݏy \ZȋQ4 QMޠB/e}tP9B(k ) `22OZ%tMކVAhGcM;fCd`:Ո )N6G!cUmQ5(SK4"hQiM.JTr A(v]== jrq89s@<~=c88  "8  "8  "8  "8  "8  "8  "8  "8  "8  "8  "8 ‾V1jp@0Ƀ`Ҍr@Dq@Dq@Dq@Dq@Dq@Dq@Dq@Dq@Dq@Dq@Dq@DY8${w8r vs@RUr@1Oq@Dq@Dq@Dq@Dq@Dq@Dq@Dq@Dq@Dq@Dq@Dq@_-b8 p@0P;`%}cyEq@Dq@Dq@Dq@Dq@Dq@Dq@Dq@Dq@Dq@Dq@Dq@_ tmc=ڀ[Wy2i޶Qvo`ܩ\쪹xjߘ{~6;nIeGɢ|r>&++,'a2/M \G_L",#@fiytZz[ү>M9>d$ 5 |8.Jx[mp휎JՄѽɿ7{i8?MDӓ0If3wX6ZCywǿ1﯍@-B} @8O-gNNN2غNdPj:iL`}eyD] RAU;!yGxYW4:p(( |WEѶyY;C^wpl<)rV]rN@AN6VTэ;CDWkbh}8Vw2VoE|՝ϸі]#כ+]2xi9;.VP^Z>>1g{͘* )ǝtryBi/'[ɇW\'d Cl״Ԏfk\~GYbqkcdIs,pMY^NM&qNʹ#w˳ܐgU[L ?}\n,s:eM0=*G\&/-Rd-7RWJI6roMa|| sN-(fi;pz0˝m[n_w_;WHǘA&Bum^'SzGw3|C+7IP_7|}DZ^ =>Ћ\1}r|zaON\ V g\fB2Dm۲*ꂱI#m3ReKcfnt̻x }PKFS4XԱ|;hefw6cBoo[ʕ if}ovbwa ^v4NT(糼=/H]9w݈RgI0ߞy=_7OmNyڠ| وם}oySǶH=m"lYS?@~~ztxNIޮYx~t5W.+ܨZl."%_0Ӆ9okh9-(_v[>+fs/A?i_Pg!/&˶Wxtj1b=-b0aD:M/ֵ$5yMֻ'(G1,Ne`vRZQti䷷MG亂Ooo,ga~L'ӳ T:H~{k}q)'>/O;ロ|o\'޵q$2|P~rd [YD)HC._ _ݲlSVwuOU4kWRUb]-j~v4臻 o:9r{Xs4g]w 7ua0ApT odށ>^'`2 Im^hLWccV+~;Eٍ߽ۛuˀgZ,w][ BH{Q֮7[ۣaE~1k.p*0+  ڻ%+AQ1sMnLֳg^4 â*tS,AݙZ-/wfp3_"`CtXGqxg`x!ȯ_Wc󽤤I _dw){(B{]GB~~Ǫ_ NZAvco }4_Av#sr?;>thx 62faV퐔IUʃe';&cJ$T6̳F55lNBRk; S#}uQ3o6Բ ]VnY߷׮Yq._Atw6z!HD~߲c$Q0POaИ|iU/ji|ԆZʫI.S\[gL-OaN ]L3ܲgc}o`<+xd!\&Qϝg*38fIE3Âw`>0'{EV2OD H@_+PB:(JPH }ޜo8x4mؖf#۱n~6٧u62];PH.ۀmViqp, č!.ڹ9 #,Waw3 Zhk3 DXYshԁ=(yt: #$LFO((21SB%cw8vwLa>ψxiiGeTc,ˀӜYMƪUf|jᔴ)jE MI 1fJ } {KIVgs.{g+,]#]axCG>%.DRTo2J ,: > ˨˻tϫrǠ9 ? ۻ:aMgڌϚqoPg| S N|9jbT&JCԎ`&ň*KN%mB(YX?qg3~Ԥj,oZf[\D0[C[w@}ŒD]{ nҥ(); Zɾ, ҹ~w?\8ϩhr !e6$Jd=顇WaonR#e\BPn:7:U=řo u9 V4nGM صovomlu^?ߣ߬++ze),!TPY{v"tlo"֌@t3Cf{;6ߐblpCZuz{fG6 Ao_ͶChYCzƭwmylY=z^i0yP*YWwt< ,y0-7]97jw;?pl)]O[8!?Emqs򴽐3{\DsSHl-RpScaISwr1p۩UAA<}'!8&s^̂(*`MyGK@o)Ii~r=dߠ((@^;l#;IX֤>SF-D1ͽc8ϸhIip:IDAe*OH$YRIx͸^kbB鋉-jCGg}>S_Oyɞwu8_nni6/eqa9aZyT%n3,Sg{#1ŜYCTD3EIq2 Z62:ǘ'X@3 TO &Dy]>1TQ*ˠJkb֌J1]X3Յ.']^4g2Ed| YkP9f;3Hd[3 F#c))j% TΒ"3,,{.2½gJ%:mGJmV2 ڔ>[ۏq<];ڜڜ nxEeEt܄Y43#%`Js&ueY)THT Ȑ OtML$( cJp5G|>y^P#>ZF댢HYh %DiqSh%Tv4J h jͨR*r'zh,Rq>45jAʆ:ų0ZبN~'V d+FM-5C$ĘtBMR?)pñUJѦS['\ДQeD,Fep)}=}G}}oSx3oαj|.qûOgc!pMœ^ԙ)gJ怹ĀqĽΕ1CS-2W!FS&FFЇL qc #jOۓ(e' J#g/P׬tױt|x{ 9SZj [s'RWA'TJ? (Tj:lLΫ׋Rȸ촞dKO|{d6r uQ;w, Z*_Ãy˃l/m/v?bU"Vl >xXQ&h2x̖y􁅇-&y( L,5gGAoLm'Li,oe?b%"ĠIMеZ֎\裒:Y궄ڞsy9y3r#ȡ 3a aReG B$% LG[*G Pj/K2F"a=olqTLLZ ۛTOs$>uV=m6=oB|dI3(78lX7_Clɳ;5]mx-ztSt]ʒT3MI)2!> 1uʃ35V@$wNs+[1 "^`۞WkX\Et֒gP K"*mKk4$"9D BlqOǖq)IX ^IsMM59{"OZ hՂYkADt 4hZ+DN'ojkG[n 3R^L'>'$EL(b:ߏ\Lr0}h=Tg]uZE2[%X/?k5lKJ`DxjC_75w'bjk=\Bsiֺꚣ{ ~lTo'߶Ng_^gkQJ170Wr\ K>GCOZƛ͍=$Eg=[}-: >ءm?,V8~0!YiDpQCr@У(ӡA=qq-U!\f=l%ej$] ȩ9^;ne3T!{3cs8ST ACw1.GbC%hY'9s_ gUi]Le&r)3>>= L|[|p]PG"|4X ԡA ?omTDhlD+=Ոj7ތm@c ؅=˞ŸG$3*0(ͧ+-n/mCƓđ:bw?Zih1q-k -ƭfXh7 n.=z"ehHkcPv^-?M}-HRpۥ!ʔ/=gT0?=[W`=G'>Gfǎ!z 2) M>+6(©+&*Fѵ4gLqgP2CReDd('!_ON&oNi7dT#-EVn5n6Py7"ҏ^3ىx]h)y8&J[k4c=}&B:B:]v-ѱʧ$}нiN q,xEY 4z! {!@2KWhߦX0[ӶOTbW)A6c4 jޔttRH0z,Rx4`R qJ;0F+{ j^#E#dY&CQ1,7Ac A#d0&Ph yi562;]xQ<}UO5K<g i8İTCD&c}eAX3d (,[14n*hMq=x^GxEILBh0ɣeEL5"3D2 qr蘉=Y٬?'4ɺ -Z]G!*u/PԵe7W G%;K](R2_)-[y.u%}6.FجׯVg!fx,|.=;񂿺o_*+S\p,>.ܼ?ڥԷ)釪@s+l[9 JQQQe¢}Wƒ+~Or,S}Xn^8l̒exCx fѦ3(dySXK馨$cm^~0ϋvחwPy-knWX^ի"X0~MAV%b KCR)K%^'J/K`HIzL6b|<pw%;ƅ'N syT1a8ڟZ~Sn秖/]x6t!&e 0K;IF-zHNO|iS R9# `5.eN^!:ngu ЎroZ79O ZD,I4wtx>F(v.$%(kJQhؙL<\8@ ^݁{뉌.A5OtN#JD9^R֗Z}RpK(1IL8SH Yq Q?߄40g5)ni]ˏRe)!A* RMF)Ry5K#)iˋA]tda%PJ-!dZEt)X>S7Mz8A撲!:2D!Y ۘOh "h%TIvy=&M{}Mz~㷾,dSK5r QV=SVh=ggՀ ԣO-<)fnrH-S,ac\LpLeDa&:o=Z-De _uAE(6Za7ðsCUM6PTӔ[+R6|fy}o6@3YfbK̯Bŏ?fElXvÊlPzfV,81E^A/|sR O:ZZgi}:2Ň FY owW\hYxl2b{|lsbs~&H9pVTcb zr1m ~HԮb4qb1fV%AY:CQKtzMvq :n>h{>9]瘶MFvĖf1hR-t)z/c$G%5XxϖPIqP] 4.2ey4zHe<' Vya)$vA\PHBc ;F\ո霢%QtYuh3,I;5A4U,iv䶘&eݟ4E6I mn8[e+>@R & II DO:ADV@$wNsnsd/Mfe!dZ3qq+`ZKP54NPDZk4g83@ u{IT`cX8O1)Lֻ&5rhpN=3Uf)A /V?M^kiDb4rn6$ϬSJyAS# ELy RTs~"u;!ƴƚ:(!,Cwq&DqE-KV8 R* e\K^&1n_Ӑ0¢_Gqb`[yJYL!PR.|ȅ)4_g*CMWCsT"[=iwU_>-Ԡ12QB9a+2ЊfpG_9*lYk6"R@x' 62mՍ_C៕Or|Q%% 6ݲw~:|-p%ٷo*Mt~ Ʈ/\RJ.:fpRɖUaLMk\+ӭl8f6եE^_+7w\}}up2J ̥2o]횃Emyy/;w㷂/)%!w33V~f]X~uhxpLHT8gQƐ~]u*љN<"7^y96nzY!${sCY$gsMCHR΀Vx@)&8s! *6hKYz~tfriτ_b,ڄD{'$8є(՚Xswj Ej"qo 1XÁS/.@:s O+OcދĴ&r^0q<1,z I$Ϥpz870Y\̫g`4='$K!GHsuUG NPA;bZ*}B[f= l%eT*8י#[ y:~b~&(!hDP<ԗ]gyT+[[՗˸]n"?F>H랐B[CZ^֣٧?\`+&)8RDKǙ-%LhdOغњWoCG,y?OAXcTj BG,6Rr!ru:{DH\rr`"j֜I-$gJZfAp1xΒˈu&ճGgGR=la;FezL_Y,t{5X]f MWHMT\]3齐FPy EP˵UVwLv2yX. QX%Dic@33D/Mߍ"engl1I3>^;"EY48f! ;!XIWߦmh{k,gb׃>Amc4t# q:o]4}?̄;{_ [>oGw/[-a'#: *6E&Xu^Ect4A{[Y7I{;;t ͔K!x FfdN$P&!*Jh':HzwvL0P!mmnٲL >XC}^bR.pE1X DtoFnr"3߅t=yl3/v,EL2ʼKyjX4GH+>$NS{Nϓ۷knGR^>z򟍦Wd:zd<=Y QfE;lZWG:N.j.3p,~#<tv%mhJ_~[7h.D)7 f&4Mj>Ns.SsXv^X-|2KUJj ;xWEۇAi/${S|{:Z1M;CoõK2viuPP)uK`󘔵JD%qN5ЧrJt7b|<o*1.{_L˽h;w1oJ5]{ڵ,)ߡktk0J[HI=$tNs: e*XL9Di#(Xw sdU1YU;v[sS]8/`[k"G")aDĒpG0oA@>d A]F #ܫ3x@?p{Ut { @SZGJ%‹r>Rї ZR\Xd$FKQzfe4LXSH= ]0 ?܂H_a0pe =K \ /;RaRiWup3T!Z~Nkp-ԖPj+uԜp܎ EڳŴ>ߙBrw䔟"H1H lYpw8(~‡Z<󊣢̊9 R`D"ś&آQTPxH%H'DB8W@RiiHĔN:U2s:BIUIausKژ DϥVjBH*54rgB. %*E!TI y =LE{u71p?k9Ȩ`|y汭k=RGLu¸:_H.=SJy81 M)89C{w d)7v*ݗ,ݳ9/õrIv8͵6xa"):]RRƹDI$)N]$*O:Js24|Ĺz=tyC?dr%22Ou xlٮӋQ"!@LFdlW|v@ޖ>T"G\ Ȼ,_>8#4Q.!R%RL^% 'Hbq2Sd$ VwV J<\3AYm6N%)@$8$ c`cٙ8PJ9AOD8$#DQEᜌk€0oaR^ss8;!䛢d ~R]Q?OqNclߋrd'UU?x2mIQM'yG:ǟ>mV"Lj㴠8|Sx`M,Aq=~I:lmk0fͤlC |^γmO?%Z38'aW8Hz;\}ī,iI_ sGǁ0R%ZW|5!)oũ=W#4 s gChM>. 'U1IF0)0ȌOPLAjPs2C8c:q:\8ܘl~yӿl{ 1LB34(c:{mG`et8 d DS*+йeI(:L記ت'nt6@` LٜkeX'AWh7H>Itw*a laE\? j@1k ~>!.|koMuѵ?׿Xw .tX~ 7[M!~ 󯯋jxaճzw:&`¥Y๻al0;*olO&!k;vig[Y_1|j QkNOlD={|jJ=0mJ&B663a yN DPL"HiN{"`:!$*ߚ΂ǹDљi`-/ާ`HĹ&Qz(nX\'hh ʰIy *wphBgZgkօiwX (0Scuq1` "$y헔7Z1L>SJE2I$rt j/g#:F+$5moc{dc3 Lc)Gok_<%֒dH4vOKc3fnmcr@) 3a}) B D.-wmm$IWn;3#@c;=i h`_h&,QRr{0'Q2K>XXṺ'aXŒT ^#=j7?G$#LU)8аCt{|jrnY560yfqt2*n*>ǻ[wK>.y| {T߶93#mTO=9YE#Уlm.Ά;GF V>ѣ]E ɶDɶy։JcU`#JcJ(?F>DRԲ{ٚ\/宭sP2o|at\ i7!X&P 2jj)$!IJ<6NB/uK7Ki2׈GYuIZV) UM"p,q[NMo}u:{rd~{SdLUBǟw^^5Zt RQISQA.:Q3WEhlO=nl=QD$Le#ѹB>kIt6giT. $eZMӰ!Oݏ0~w;y&~s~}.*5&ێs]>ёJ쭀DsN/aҼU~)GbJsAV㼪`~1XRg󦒩*E`/ qܤ1mSm)FW儋ȒbN!F/<*fj *32MPj8lƯ9)$@o&+NA ` ɻB]OI_4,]\~eTX N]\ovP]9QzއBO/C.)us*J8!M$쏞#Ň[r>H-M~d-˹ Щd.7~OzB?ؓ{ ҇n,o<]Zȥ)B,|F{J2_1]K hdzPnj~Hqxz?{~^U]@宯njP~<~%es^́Cꑡc_i`lAC_X1xSo_?TqqU?Ku:[?إk yh{ߣ]6_**pwoww˻/hQ<@>>w< rt6յ2%EPM*FYAI>9k)؊9Bz܃,W›( ,PtuߵlTpXEciZg~;=>4~Lhzrk^8S'$/R5/epiϟ6RMnu zD$.k):W%lUP1D6三܌5hůsTZ U%Z934*@jldM8W GӌcЏVs3vHˋN@/>.h_9ϦO_; 8IFW+* L:x$8hI)[[F7`/d%(FE& GW+mEl'75*ucK/fy+|)=v$Em'>14sB*|c_s RU٥[/jعv:sR5f}5aƆ$2dDщ"QEd@K˃/BBF9ُ}P7Χ,lRq48idD"ޣI'g55,\F+?Kk%JQG bu (H Bˈf+)[VL19.i4RI4vrُ+\\w.ѴX\#i{N)DA5VBaS %J')l(,*i$Rfy9Q6/bx&g~}b[CfE̩{(!fuSrvv&]QZ:m|t2?,o:jU*%RjF2>p & e|[)))}מy/O 9֛v~K k7֕+Ye[ >6K*y6X ¡:K˕Lʙm"ˬD?4 AFQ;QH2J&~#ްU鮗V ┭oݴهnMt.v-uؕa+LzE$,,Njl:$y=˯'oʲzY dy:]}2nUUOk N]/e~v7봷т?}h "B)}I |3Y^^LVJsna mtnzKgv'˩VW~g\\LVlW')|z>[Ӌ.&|dM硅U<*-)̝7g_O #곥EaDkO`JcWUQȝ xA?X<oUUz:N=R6"!(UVv|GzU,]S&:wqqV _mHn;#sM}67oOz S6*KѰUtD;@esZjN>hdid#,r3ʺnt"vZPݖUeu 0BhM7->km%]sUվ+nz]Aߧ@vz^&Oؿ?äGvmR8EZ;\>Sh[lg2ƫoVk 6%=}U. M|dk_)LӝWJ~7MK=o&Lh1Efem\J}]j'Rom|:!o\C[\]̷!oZee7\, 2YN2/wؑꩻksw;|ˠSTooFy{wҮtZl`{l$|"i׎\C˥=^/TG HyGtˀL5?-׎>ov3w' Z#1->~i1R R"*J=|hdt5.Hd %‰Ք>P~p WwGyy;].Uo)=򏵸A&!lW-f+TQۨendiW1DzEJt H,|چJTkQ2B+ nVFj)vw:vo،)vJu=*r8?Q RZ[RXΪ&eP&ZuKDFØ1U)#͠U'75>'N8_S Ai=@=-Y}~/!Kh M> %fiv)YWK-nbE&B%c* ="X\L r l ltjUԔp6pNHoD0|$hruyRd I -H Z>m\)SkPYUVX٨R(j-svJWf䝨䪗1(!QaZI1#z=TK@D&d'E"]L.J{+قd/2 {FҨ~1QJXdLCV S ^3J6$aZrh5"aBVFY)dUp! Q@CNSX*,eU0@< 4@1e[]Np Fkdxd*>v0pZdGU&ʁc,M4u%^2h #T16l &!aa=e 8`SX-Ҹ:;F]EVp :$@]4U2+-YpdAMBP p'K?+s-MꋐAV` i˔ACUY@t4.3h'-a;&͠j`o\I1h0c)P)5_w Q Ks$.lS!biv UVQ r\VWiB+ ؂A bo.!LEwF-%\f 7(,Q{vt;%Vݣ.P˒ mSX`ˋE/[* Xp 2 RD AEjLah,)$qۭV^b-|IZ W])4a tjG Ҽ_ҭb^,|1E۲9b؈@@wlә`{1m"ǟXp *`|>T$MuKK/P\aY{qwmI eX~Y ˞ؕQGI(Kvi T@7*^VBG**Z@Ln\ph)bd<@n0'jKmUXvaBxP0IĄ sZVyE(C`rQίVYh~^\'6Xt_ t<&`f06W nN@'WNeiV+a`5ɜB\V'#a K_-n9yθs'S͂.&"b%V. m^l5A]%ྫྷTMnR"de@ ou6y"Ze a 9@g@RJXU@zQLMl`6QisWpaP" X{SK9\nx:\#r\ p 3PPyD@kO[5?N( z{Ո;`XolWGȊyE`1\sZfe 0ƔE^g<"WÈ PQ¢CHb4k00Y:@m oe0?yڰ(RVУg10)Fndj5R+;1Vg=_s$/-JLRf\K#L5%i\,k{Hy~kPP2(Kl5ʾ.G%z5"m@ja[`93#ae%3B Z̀ +(B?pܘ#<_#Jxs%? kslfC1Iبv""l٥0;Fff%daS$܄$‡Ȱu!Ē۞q9Y&.Q~UvW$,Xx3{Te3`PF"jxJaW@yMxbe'$ qr駅.,n*QEYe 3ǤǼ).:Ø;gw] aky8< kʆ{x@"pxx@"D< x@"D< x@"D< x@"D< x@"D< x@">Uz鲤Vbj$;u{ips9-\"0xb!n[D`pgpps~:yxG)L\\?/? U\*B=pj8Mx[ݧ)\R\͹7w0\8xz  2?}[a+F/@_{FZm,z&YxuAG:Wz1Qth;nnⴼcSCeS- |9§8vSL2)-hC]c":8"0[F@TWMUrLPndL.e_찪jt}uldpZ*͌gyjD-*%YժJ A_H7 Gx/n1~ QP_ .62Sc5&ff"BIUl88界 y#t~$oia ӯR*RMB\o̫K[q|uSWJW=k vl4_Sf|4L3wӒ07m :͚^53.yG%hy!d"-#fN)Rsޓw)zk*Uy.iZ߂=˺ 7q4 ˶3\iX oXahUeFT:֬'>SB1f2ԫ_!GGaG㏸A`KT4lf=pTHYJ| Sp:!O9V]hGF[)4z8Xcݡv/5ROȮ#qލw)Һ AyܩsRcѭ>A/56yMEXYj-_Bxkϻq3m1L6;w;G*ng[ل#p7`l,8L)΀qޓKiZmo`Z]P΋\jjXDB o{ hDͪ.uw5yO.y`pj\l`:AlkMثh,l-!`emD`Ҟ/QuNZhh֓KֱMȲݮi5+JqB #ҩ㄄8ilw2Yq*^#! ]ǐJw@"ފ* Sz%$Z$mRj Y_L 3b2eטw#8fjdN_fZgpJ=$D8F药qXyRת?+Ppe&Wx&!TUU2X-,3S6\+y !D!rH{op3{gͣŕ(̊AR{;6fî5ȍ[ CB)1;iVfCN‘:qh}PC[(;?ip뷥줿gexOD{ Gq/n,o`7Xs<;(FJm&ŽNͤs>GW2Lh7)>oL:?B?e?D|8|Qu"\S\YuMwV;d?NuIO#u<\ZqTyz:[XZwtr16T:vs ;ް 孚9^6θ{O^ݔ-6?'O[ &\EoV[?w\ Oh#E0b֗yft b?FCr*JCj^Uލ@tB*nMg~TolBqp ~4 .Ic|U9Mi?OgkJ[?7a};e.~ﳟ.~Z1]r{$`=guMe}ru2Iz=G+Wϊ뼬a'擏`젻s?%y7T!%)8:;OԨ4xjxv|PQuFSgB*s,ݜQ1W ]rK>J#5˅N'i-DޟyMgDe`Çe[(0_U 'RQ5ztv\ɮ2oB|qt|S?"A9HU)v3pvuFb'~"4AфF[ӔbsMqK%oA3O^ԣzÛTb$QPi{z"_s%h|V5d][Rb-dnGEnFWKu]NŹ? K)*S')1?/ 1.D's*\^yǩI8-&qZLt$&hKZXn50gShLeSlC <M6'alēC|_5/,cqoל]?M_t{1>[15DAqǨ}go^xg׽oYǚgv+[ΝқvW>霷+ar1~Ϗ'jE&iz?^8NxRWQ\݈?,G`v&YmA[*8fq"\ɓҵ^.xt]ߠkЌPEh?bEIqp3jYeS>Q^aj?%Rmy*É:5`Ib{yp7;劸YH<8'{ŖX,vلi wU{u:5F=,Щ !4Ls%Xl59+D f : 7΃;iYirLJ)5-բBvkSѽp=,Ŷ [> Gldwz:;OMB!=oذ"gclk>"I_6z'q.3!׎KxrؤjS{ב4u1"ǻRoFnoo[=;[۸'ب* ba۲'LSڤW5%s059Bpiy"sm|SԛvR?#wnfi8/䬥/]\R a wMSr"4MO!fv+/h{'w%qՀlKRwmH /\ڪnjܗ5o4P$v3C# 5UeEA7nc0U› NmPʇڂ!ė 5.Dhe` G S -Ҷ jP2DpnHRh ܭԺfؼKys_UZpӜCgaj*ulМAg`Q 7 2D1- 9i.ZpV.)7 0[񑳂Wx[Dָ8ƾ0( "L9//̫-TXvb3WtX# &jYzQDh^r3A7A#H<g8ee/d,iݥV Y*aAŪՇs#r q>qgSJp?2rP{9c&`6Fx،!,Ga8ƚLxCƉlʆX ް7dɐ ٗHY~>rtŀΘs &"k}h>~ʍrDQ,Gf(IHa o'? %Zo83.u'- }mIr%p2u…_"Rԕ$6 Ke+_2y \ Ag™eͩ1X;`kSo;@/\<HD7< ɲP~F4{gfg#ϼ|rp(A (o DM9dx4D7E$r!ƜR!US= 0` @`030B.9+i0~F )P; p!HȆk!DKKM GFӇ-kIqs7'."abUXm[⠫A>ի]Օvi4^Һm_"ʩm f>r c.EgQ=VA |b#jSpr6'PP3 "ҸSVFYB1~PY9XΔ ǕIC6kujbE +CӓXё*Y6ёU&H1ZJ*қU!Z'U>zmc\mݣfi!I_jo€ˡT _:%dsC8tAʃ> 4Fdзzt^>NB/{Bjgtxղ  ) glj* )#/ t؁ =C L@)TMo8ϲ9*a\Kl EqhIIV Y1, oni/ a;H}%0=~$AOZL%u3\NɅZ6Rmp ̩dB PW;-ECKf!&&Ƒĝ? }dΟ'.AԐV0C ,87T|TF _Z2< Idp0id {^& G`YAjJDPSo98%pL ,*M~V 01ɀ`$ե@ EՅ1VIYKs(r삇1(7MIte%p% 77S恼yA:1' R&X<섉e)O/CYVЦ~h00Ba7LJ^,cY ʬMNaA qB?#8C+3?#8>ޏ)Yo|eL7 t[ Ta&M-B9x?Sx@A[~qioB#j*oq roqr@y,B*t[fY؀!1ʄ Y#[^F )s_& 8L`svt&f=[yv#8qEͬߊ4Ffy+{y`QIo3966ެ\ݔѯiY./1F]/Z]Nx^N_̬GeG}<-~4`On 5Œ(Ysd^,O5K&<䘇1*2*3d 4;'rZ`6 d͞,,p]?E 1o0N:HguC,}E7eqjT/a7͆i ޻5H֣Z#Z_#,;44  Im7Q}>4az%VIX}!]J;fbkm~a\|v(ub8}`đC:͛' 7ѯyz.,UBIOt6pbrHc¨K2WJO"} 'a 4P.je,rZNo~d$Tf]'nKŒPg]C))Ͳ]+4 iպe[[SVx1#ӔwH9!eUhREtw@V}/㬈l' N?P(^)F_֭抴wj46mg'L +9 +X D:r\S*zh'&WaszUNF4^-׎dkq2Ƶ:!*!N{a<-~|g?-%}a}`#$%=Q! s9 [: }sI|3/[SN >2Gw;Ҕ&ueV= ٧OlY}y_gusd~N!shnFh@`Bm/0SHtăy`DYU6}}]'PLȳ[6V7 huGMjs%R&,R\)\ szJ1s 8{3NQ-u:3ҜRw^،$hy2t ~Z\{oܲmn+ƻOX͛65np9ޡBaѿGP|3 Hr&5D>;wHdlµ;0 0{=T8Ȼ"E"&"IE42IZwgtrur+eVQ 7H!EG6L2ų$}!nÔJFY\,#agqQ,sY2TjbDami}1MS'XbG]l90B}+ϗg.NP|*e4YQKNXHשSTDU]٢UGd8'Hv<MkҔ8L x+ŒA}˼VY!.fygd zf]w߾B]<ȓX,T UwgzxYpV$QBCĤGN@D\>>D9jVnC!xv>c j^V)kdqCd^6*a9L뎓VÙpΏ깻ny(b* *ޞ?Š_O% P d[ֳC=x<,,hsk1跏AZ]a_e>.euE*Oc>@ F$((I!!iu %S{sk5Րgi񁭈jcCEGUjYXQCFP8VYEaXCdFtv\ԸnIqYc,5gzPDJ $,M@F MjLԈ9[*uQCj]2ZޙOrfvߓ+&j{ib6 4j828yb̓xHJ&4BK2cQVu,"qT3DgJE 㙔jюui0W#n跫1TgMsFѬJ\ofFsYR<3qV@)Pno(Si( Jy X  (IVQ4Wy5 j(L{ޝmo^6,K$q!P6xFѩz)P*r,{Kx9I 4'"%Bfo"&`2WP9O~0L&5M&lfpd[U_˯rSN2T=ِi߀A^ɽs. ɲL9{)#hXM:ö~e2S`[I}^HH Y!e{L+G=Wh]at'gdvl3H#|F D$UXD8G BH&9 `dI9ˑ*_yd5U]1X~LQgͅ}# eP~~q#/UMK2c7f\Ie;&Ɨ$fpc6ޢ*<4&ea.]̖9zyG{{0 Ie;șm) NDӇ|`2 Fu1F9F9n,"m6:j6F4HK@ 6=U?z3*YatTSp \4h6QWk>(NQӲ2 xD˕M灈+d<uN=3yqߩSG!UmeUmVN$duSsxDkx$ j /K)it6E#06T=rRj̞H2"8kiB-U3nɧ5;\P:J}NUFepM.Ke[rΪY-t$~2:{}d.+T!TSZ "2l>MX_8<=}UYlRd~ hl(SEVMӥ㢝A{DOJxU&ل2gYGU#UD0H^<6=iޅ L"L6YnCCeB28ynASq K1U "M3p0`k̋AKISŦQcŌحVCZ=QWI (:#.MѯҲSRnE;l!.~J:8@Ǩ)UM[HOxΓLv A"5 J2>|VAIR@#-(¡JJ(x ΎWn8 )w]'Ti`f;ࢠװyz;mw။"\cvRE['<*LBωkx$XҪ7^oby7<ט{ԊfU׍*z_i @ٹVgY#tZBR]HlkJ@ӂvH-\}Bö[mGƂii#9@! B+p5_N5Iu &9XM0A"_Χ`0Tռ7ӠŒԌ5*wiA8Ϧ}wŽ{v-:M qoK`Hb;tK41,q AWi6 aFRZbN6NbW.fKB/m2o14)DvEjVGc/TOhk64!VZ]]9zyT@[,If}VsrfE-2*SuUiC>vH1 j#TC Ŝ;3q&AE0QZ'yj>/b )7r bZk'vP *x{oo}j'ȵxo~xUb71.vM" Z1SdRuңM@Jqe GHoi]+#@5 @ͤ}b_OaNsJJG#H{p:~=Vc@.,L7J@ F ${r=8\ECN#f{cW jP ثqOmt <RP$t?~[9}-cXuG: |@J%2De2OX$egr4ޣA8ط^1W`">.\:e/QE 淼7pq?{8BګΫнg8"D-CD~ӵcAp9q;7]}5,JҔU-Q ,S>$r@KP98v{9+A>.1T;޵[a Y%!Cd^V0EC .-m D _'~UJXzT#+SR,e7ߥsBK.XT)n *S{nwV܋@kZ95[ŨPъI1!b:o~XXK2RV$(<[xFdZ .5SU)NpC$sx ZISHPisr)@!3$?4'ƭBI Q_~kX5O\YYPC]~ !+%ݪ ެu>-2dc*^Ɠo@:~WͿͧzdodt†"N<#zI6Ƿ]Ub)e*_[WJt/hU?(IħXgg k7klR>Q6=ؼh5+>T`JD32.7V5];@OC-cJ=ñ|b:Ttc{kCDaWEqWީoqO@qwokd>;}BO8*roSZepx%D <л) |p7tA_w`-?PA*sjR(8I,z١ǪEmA\,)P酆^t[ʲ|HL Ij*;]ӥveSSə$1*% JR:]eG6m;U@נ<I~5" CV%ôk]l3{bń .m~ gE,#kxIfk3ҷJ.N!31Ԙ"c騵1.,Hb4#Q*>-Q6'r>LˇgzgxGHjG|NԅgϋկtU^ɸ'p^w[PV="0eUf~n wV l ߏl `YH-%ٖʏv$Y.U˥*%Ma%됗s%gӻ!3 S ]ṵ9 0~[xzUeKm{hu4 _a}|"y1 (3anZ톰O99ko6R<yF:׳FV$×Y"ͬ<αli<zq0ɲL&N7aL`d8 ~zן4Ȳ٧~q3- @dn[MVtI|c>οd 0<ߋgoJr~bIjSy M7lжCesY2! .WzsE{m7^MD=#a Dd~7K5k\SԝCs˰ʰktn!ЖjƦihtЮ~BeB& E7jTS޿&fDnHəU]4gd KsטIAo6).sak57_ZsKZ9Z2C[8/c*)̷Q|J^>IpX 4VmHNt &.fm6܇W]xyKS.Z^?Sm)6AB2BQm*B9&ˑޖSOY3;uhf`*-EWRiKv7a宦B)Yxc4*1s۟wGZc֚kA-bwM+`x".N2~N5/U`eNq{fH' SafyEyAΪTcJgRkJS` ScQUmdV 9wHP(cdY*nkZmMӫ,~_E< w"1X"1h1*$< Y&*m͊0`itꊚ4fלA3GfAFy Ox2#*E;<yv5$Z !E8Z)FI0QqԵb8A3V +cG(?Vs2:5 P ‰]:z%VȠ*vaxMbl -ai>vHFg\2N 6Nq+¨s"ԎJ1p^qB-8[d (P%}`3-BV7ںM.ϩr0L<7 B٥tATĖt9MHŎèe9'Two3O^kÞZKvl0 <U8eLټ}ׯ]2(Rvټ5'iwK?<ާ96Y#Ft1ːT;\<濵PN[ÍWMnZj[ _I5ôo| znל|ŀՓd%9@9"K\8P2j-e^M>ca0-1b#^C3`T;S*-*^!,3fetWK$Ð: 8*Y!$#FvYW?"UFdNѥPXw^(J{Je*A:T{lI7^W\U,'u.7ɫ:$S!A(0,voݏ+h$'cnyFVl< DQE2N\QRd1h a !: B+0s r7TFsWE}#y9&V8"S]ۂ2OJ)&9`MǷw8LIvL#.|kϢcZL?./TBJo{I/et G0sgS<>-[T.r+E7PNW8Mmis(BY8tW| uT:iŁ92j<$QbKi݃zyP0KcQ]bRX *JN\wd n (.JN;i?̲dDN/^I^ eTd?-h{oR:YTqIdO1ѻ$նBƅzPrTQ_ǹ|q".!ߙ1&ZY{{αWNdAJ,9}H.pʡG;ϛ ] usVt͑vsZt^ ̦r=Gp)I0t$.4-Zy0՞%p'l8ONp`U×g-+ ENC8Z6(d$24έ1&mwTePA, yЁ}p]0|V%5S+ǩ妒'&PThIM)ÙNpfeJt2v&qj-abF."Bl9N^=gg`D-y,yn+E$'I(Aa6|dpAp[Kb$E֪|+[%3chWMnz39>4b7;K>Y+s O"Q#iis`*T'ftZc_A>tX<)`&Z[ᒀbAOGұڑ=Z(LFq05Ae%< C>x*O7vy#Ԁ_Sc_:7kF0PH!G6 2Pv4: b;#@ H_F"CD]|^q/dd۳`4(awE\+DH~7g1eBK|܋ /X1d?]xt^y_ԹbxC^\.n_ɻyLS櫽g޵ɽ-I8Ma#^N,eSo7ě%RK1)а/LJf2 _,i$MueLܸKf:Hj3RG;L5Q!X=2fHg Z$A D4*qY,8tgi 1M2rJDDJ?i'.}'t$ bXӰP* H%6)[IS'RE "#nbPrgWK{&IL@ Bp|@\ǯ}yH.uےk*6m޹|vr4mMa(گ)է+[LpD}EݧK cxz%K>Ժp4Iߖw & *0@5Q @w4BEdC.$jwwz~T5.Vy.(U8 Ș饗Sknsy)N='{r H}ܔsB9X]˰6-ፅ{e 'Bnhg5Ado>/iN mMQ$mBo!Es8$˭ n<`'/ { \dQd?>r >HcIp;(~W!拄!B Dhpbbi8Mu$"2`wEKAm" yl%EF^`d;5%t}΃>f=3 jQ0|AsoMqx13#)N {8rA4LuQ4ђ٩tFokHK[̰+Y0w8iDI`֬$;A({>bzo[+uN6k^ٮj,gi/q%Sf@gM"WR_wt܋eŽn店4o &?j{wPTnIЛa΢ U$$ pkN\)̀m5>臉׸N^|#67IרΕXHd„`ce#&MB:H d#ɓDdGnj%!±.[׽,{i7CEu5*̓=tb^t6uʫfmzgȺWk p6˭I4GL]Lp68E 9 Dqv_P\q~BMdEw}jMY՚ P*Ou5 6$!%LjJ!X܏ d-R}3LX,MYp -]! )r,> nVKUAUu˔(H3I۳Ez{sAp*O8U/n5>%GR .J dlH$z  bQp|Ђ@tN'U8˓4|R3Ju-KE; Nb ,hQF̉?Uvc}\XD!Vd\fF$)װ] 9ȉ܆T:\阣!+C מƦDx|?5².V8C;SY]rp%3!U>eˢPV:N敷W^t6%ȍop+/^|k#_J o4]Z;y[I+g_{ Y5!~Շ v?!뢛 |ވ..^ '@h9 Bqs8} ږE#vʼ᫃avh߻AjFh-QnV񽛋fbol{A *FG} %F`vo Z琓՘A}cMUy/ED6?,h#Dk!K-1) OFJ8(@]jrp ľ~S:(:6lRcQ%,&(foS% IBQkPc ۵D( qnBrCD=][o\G+^v#׍,ҀL0 YebYeAZ}Z[8TXU,u9٤>(ݩ\8 \j4" mgzǎ Z8PYG]0rP(J 0:b}5gS<ט 4֋f*hwo~w[8v/ KKXDե9uFQf|sqUQax23u\3֋0d- wSܠVNy!7WXcNL֌&]h.v NG)wǵ'vHwzF6N QQ Dǎ]@ՔQݚ1V&TޕW#ɊtWtcguIw֭Vi5׵dkLr]#H ʙ]iA #\\O6*ssU1<o4#҈z.0Ws3Cc5u^|i[ck9lp3Z3ƷXȒ]Wk|MרWU5c|k(ek{N~+ zE5x۔cQqlxS0uJ۹!VƝG> tUkp/'dvToeڀ:gfoHMwaK}r`#oN@5j5v\1V1HqصW`6oUNDa}z u*kv:U1!r2c]3zpKl5 ☇0!o4]Ã'\vO#,6n:]x 0P:kxJ'1o۫d1t y@J}(;IZw<hٹ7N nfg;<ڵ1}x1ƝOaJvu78]}qnL^1Kk{ͧ%3HNp;Oc13 Jޟ~>׌:ck;/԰DYc]l=~7j~fsUu-c~ xyRsA:6A6NbN@.pUڍKe#,9@W؝+U+QhUکvީ?r\~9n;k>!b|b/TԮtL[t`ɞ'^pcXSt-.3u{`!`*+>2ИXH,9I]D*LZ8Z+HFRR1 !_{Fi kꜲ1*Jt*L]kzػTdHLͬ6 Q|,A)b#Zk^m QFY>zW_@`VL?v>[A aPp+|1>;"gcX[IUdb:+JlE6@IUs5R#djF:i,ns2@)3i2IXd.?^}6I7+WONj9#mkMhVfb82*f'/$]'':,+Uu  Q6ׇ!"wVzJl( |$Kg;cH"H7 w3Z!G>;=ɸ%8G@j:HM΅-D|Cd|'ihu3ǂءe2GEKD-6GxSǭSMrR@M] :gvRBҐJJTR*? WKFF(h4جK堋WN3;=ZٞLWN܇Br1Rڮv<}$l+Ub+:+|#4;QSEPײW!TƤP=&4P¬y +j:ٯYPMM$ײ@ǂ#mUH Vy{5\QnźBaĕSEbJ⬈DƩ5ܑwsU!)˖B!9%CBXA7UǪqFktcB3dr2ۨs8 C۬M.Zo{mV HUiV*٘9Xe(U:PD >Xd"Hl=Wr>D X}Yu:9`$ R<SA!'!}7N'(֊ YojB<['[oAE4{o"_K???^rԆs>&BiYA'IJ˱;r:i>QZD@,]ҌJ.,7p By=Xk^Y:f TB́hмKi@/d/Mrlb6 x$QۅFa9 XK8[^UTL@+qy=)DBʨMF$#LH0wPX%Q-*ѓjj:m"qVcprXhx5$2 [K}f1{3m,zN3]d>[Qn1oP.؛%l1xHVk#'9֎FڇFS-$+T֤e` TULP9fjЙѤ:ĕ`ZW?,.O*<Z"Tگ+k˿#Z^l3p ,3N JMē^Gr.^cUYSdK%iP8vhHoɢ>WsR泶]l;*pNfXȢ8HLJYo%95UtJְwo%VD `>SGyڝ7zK9h6sN: |4YZ( Ac+{Aq,:?R'NWf@3{o/8b+9RW}z&6{ϕ=f5 RB{&p rh#<^~,ty$7ZֲI9+ښS#^H4oOMMM^}+M/'0DU7ۼ="qnkYr/=ArO&ӋxNOE˧S=߯E~ϛkALo/naG8gWՅzO\v+@kWnok=Nk>R :%ˑ#G!r=vuEe%79f CʠR&l iVՔ8obUZֽohЍc]G㡐֣.>"g]p+oȏCJ1{cdqrb@81: -<9Sс+FhSh_Zeͤ"/k9Z(;"g7hx#;YF49!;/3nqYyieDz#ZiTة(;]/jux/LJlgP5P ]CƚHu r:4_r>ɞʅ$;ݕdW!"oF2~g 3#cyBv,0"ς'RMdl\3Toh۳lI?:jַXtw 8۔r|\&Smq*bх'DH:(TCZ bdĔPj =)b)=@,*3HX-\g$_V}@CFw-ܘQN? j_|E0sw 8Ўl[;̲euLu WsOI?+5tfVhx+r`J =Qqcg.^s9US H#x,Dlѫ7z>mI=>_JNܥXKRJ]l $vWٻ޸rW<. GHJ b}^a{`$n_K.N|җUNQ}Ert_G+*q2NGWY-(i9x /Rp[o{{M=6_M;,MTK⍈p\$L 8Z)Ʌ\mk)5֊AghOV-s)ǖN(EZ+LsU}jҩȕF;אָIj2;+ez Sﰸ2 C <,ƇM'GYlJgU̜6z8)Z')IO$HV+rt "nV@[9H<7VAs7ǝ7h67NO>~vvV0Ven{ZZN5jg/_P: $w'&̡z25)c¦#yL _Yzz/>#6@"EU'ٺrT%"~ȲGڴ27E.J+ҥVHHd]\P1T!0EhBص\OIHubZj-dp`]|YVdCR!9{8[ej3Pхڌ.f Nq}'XW{LsS %l\ m F\FN}A?@rzR w)M@idUup^euIe[^73 cV #>KhZ(i*oTuןK41jO. xs\V50[fT50U vLY ^s+?g@Bޫk.Ȇ=v3Q~(BF'1n{ȵZRJ "!zԊݩɐLgcia)tR#HԦv*-N*&a҄)"CO$ln.LPj$ 63tz@9ܻվE䷞*öStĠ@vsDA|reNVq<^}.GGSYy-X$o* 792zcƵUɇTʇ ;@|_̇yg1Bp7DW2)nN1E69 K[+;5%ȵcm[ޡRә'_vl'/ܑ%$1O ՞dCC4.#4f I],9>ruE.NGdCI^tQ_TRx&~<ſ+n髗/=WQ˟_9NN[9a`?.F4?~$̕ k1:T/tjuu4>?{q.:}rz1"g:\@U.x@^cOT32w#^ 1-^r <@7D^1yHE_IU+m @^*Nu_ H2a̵lL/:bsz$ޣ somo[y?ޘa!Fnn7?2}(t%RPJH˰c3ϻ7g-: ],.>xx5x?gg?8ūZIG#)h$դZzѩ3fUh#^\Ө(GiF &WpQT1F\VT F+"ZV:ck>*Ńu^g"~RQ% WSЍ0w +XNGYs(2CRV'}*h׽^1[dĤQP Xp᝭;bO(x[;vs!:[w|"{8b{1A{#:O2?YEtzCmYD^Y=È=L_bvٕ9jelݨ[U\_>Tzv>֪ LIz7>\OR"z3֋m{͈ҕ}nc)nS۾{܏:85u5w= v1ߎh&%*r>q3ٕ|)5 > -$Ͻvp0)RnB$9Dzzt-6>+Gn/KҶt·Hw;%>\Թ,CQA{ yOn)_zoxYϻ+x/7eA2:V<~Ƒ)5wn.`rDy݁yw`9$x;T!x Kn7z.aQi[ÕIG:G&\I6)n1V}̵$SI Ei]h.7̊Z2E?mF;W*oiJPkuFA߅΁>4 #[[s˴oq:bmy8eJ_˫֪$`粘oCgHSu(e)(ܓ ރ~EF'Gb?YlQj ~5ํ8&T+ǟ?zEjݵ'Ν~ѤNSNޭdD=_LYQ.>8>:Mbq:yBn3]bߢH{D#ƧՔG#B)T w|8j!>ls!)wQ%H+T86&l҅F80D^Z1$y:s #Gqg1&_pm a5iġZ?k3|$^\_ԑN>Z[OOUe^uc4Q,?.l 7hͭOtyި/?sakL{c~vID2qvR|-& &"e-S(sbLAm#źsGewM_֧o:FRX 0,SD^gZzԘh#G dcR_F]XFt:.bk^уCG[BNh'~4B;eqndp"ۀulzD!>S vO8.9f.TH4s#m< ƹ ڷq?ʩ1[9.vRهwNqX qT$˂CX.B@/#H'g9RnV@6yaT ޹8nЅ" +*qz w%Z(S5C v"E uHκ*0B{')){(|1n~Oףkg\-č#poo$pnw/6+QE5\@u,`u{d0Ɍ&y!z*!:uמ2+ldBX[]r<9<j5~^P LlC W:NDSELҥzIm-gdukp&]v(edQi\y[gu:ңgu9th9Xͬ'+=k؂{^sY,-ь*Rqk.1 !WAF=#]\X#Uptݶ Hpg*@" .>OAC O=Fuo㥘<5*.\$E}rO@'j!wXy(T1y){"+QG>: J/1Q)oCioRAU~G 7A>%1fu$0jPs^[ܩ!eҨV̑HSSVfܣz-aY'} &8B*v7|YyfT0Y<՟$h~hV"GM&E.ucpǾS+]ŜAiK=o ]3q2ٙ_Tn '.UD~IEluh &U IdQ?PS!mu6dTz$R,SxYp/Z#q-yYAx^V /e}JE{v#<ڜM 4ALnA!ig,o[>7V pjsETP >vǼ~/̭M, nI3vYꝂ!1a4t>iܦ#DS1?G3H$6ZFqs ޒKv,Y˄K6 !,.% F)z9N\B RlfQֽ8)*aR7Lt)^fmͮ+( x_ s\\Ru/-|`Cn {땭0DZNi6j'D c:!zV8&TN`GZjG{9]Y~wjʽ n235( QHF\KZ͂\JSWzNKKT:> Hº KeG}NlRoqd <ᨼ$QD8q&Q(,۹$%9uu!vs*?O^e <$p 8%gF.3rw?].^|i\OMlnQI}.mT ôJE@h레J=5|Wz/ F@J0;oN?ǣ?v=ZԫWcA])`?7gJ./\0]8iu'\Mb?*++2X>]> ޶.BR5_A>׍߃d8 Rxtu{Яxtx]=yFyck߁ #r7;̔;G*nX%"oN9lH&ayg$pl1F.7[c-6VLj},.5x3$bwoIhC$fvu鎏tުJ) 5,_maY+ x׋̽q{u;-[5g5n⭍yz)C>ƽ%`/_̜|;?gVj-L |fi`\jXaV&ѶZN4Xu&=9U}6J-*T~)x%9:L 0*DR0ޔ1ZplRȗguGZش_b;J*H{ 8h(%}E|)d7\H1(':C;7CHx'_N絒!4[<#4Fn\ h8vkBܒfЃQHD ]L*b7%KR*(*_1sd3M[{KJX`] BӅ?%MM[}°E|7q(-R;D([9>i1A>Ve,_/1׋Uu>0aWNɦVuxc]J gU0u"]qXc*菢ͣ,_P}w'+prTrabUF6d5|S n=1PČsAsR|%cV_+F{nF uU%JdZ6D5@ZA'UF ڨϛƤ{\\0&)*1?Mvr7Ql mNT<(W,BSB>ڤPZC>.)%)%#Fԝ:{Iɣ"sJqωWˮl}*v Ͽ J_y2slkVyyן( E㪏KgM#D4*/Gn4G!DhrKy5YmN5ywCJ9HqPce|iD!?MN!nS`\RĒSC>Q`"z%(qzq@i;qH ͡n@-)>F3PsoQfnJS¨FNsQҞzvx혟csPB' č0Ut;IN\p}X+v#4.G~6TXji h@i& 8~J @;'L<Y@=ٵ`) kv{5jK-d[_=BCV08fȁ~Bz4pΙrLw\KOtG_7DΉt`ZHBW}ťэ3 ]WsUXtS{9za\HZ$>` X~7 ZP89?/g5MS(1EɱLJ\$&(sjRt;(GP*3^ոbbYB: r,k5Oِm>%1殲UCїF6zMc!5%WR8Ԑ{ :\'#KǶl=wWc*ܖxuC 7/ D|MU--n0 .;Z@\GGDIo :E*ȨW]F>B8~O!rW|FpTQN6EU.YܫڄUSm7i:G]J 6)i:k\uRrUS/8rC]մX4jFl`[2І2n6sG7S۹X l=1U1_<O_~iL<ދVf v1^&ewΒ3܂ bfMU&onl~9C ١ͣ@W\ }0oL+1P7a<#r1X̎βƚJt έI hN'\6'-Llr72hICʯ{T{}lߡ-끁'gHҏITq"I͚Bu'PJڡ yf>nkה}6С,A顺cHӐcȸ^|Nj6cR)Hf[HiNC1StCʓ6ag`ugy(d 3d̳ + ?Cؑ8Ir;%LlnW ZФ 0Z)8PZ/ oᕣIc.w޿5GJqX [|$qH#acL}!l @bV6= 7(ǔJ.%,-r0 >I42(/+lƂ(m@z*Vy׫otG D1{zچqI)8g+ f v@+8jLQ&DGN^|2Vbk:M)@HJ7[b\Tw.CMB aGhv3a}FW `L^=~Op }==CS'Q^[7o7yп?ͫ5 t(h< lCRI}oՇ.= FU$gL5d6Y)fL)ǚGYpx=cT=~~|C~'gS͌x+fevtR((%\ph+HnVL XCisq$bpI>\b$̄"RpX=~a90(I*ٌ#R`)~1vX ZsNs8ĸ-|`γ8 V3o7ڣÆ9X.*  oCZG`꛸fr/TQeÀ1UsJF}D4 `ꀲK?e Ꝡ7_(1RZtMvI7&َ{*4)!b'BQ#C>}#wBvJRRĢJ qd-1)1J &ص!3t*.ߍ-apq|4HW.c[獪 34FD;d>م57L[T1`g{78xs%8^G58!dqgG_kazܟK/MQrDF8|:ߜQ]O笮[ BpT)V~Ųb#/ ؃hyLwV̺bO4 )c >t??z@q?vcIgHs0_C3$Eg&]V}ك-A=Lr&eI啋1ƕ=aFEt(k%=DN_P1|mHq[7+{X.2 # #jjĸY^ "c agǖܵ`Tk1w~O%,StGd=XG}wlZ饈gQn8,[H\;>lc\y} &g0d%>l+ook/7Cěf368ZyʛV}'+U|ׁ"}b!,n<j!-hI 1?ZIr=ӈ1fS۰W3"؉}VnrYfz75z3{agfn&E܇pp6˭s2@ rjEp%?8ieT_͘yƘ"` $W[9Y9®DWp_͑I%jLũw"+gxA h;}'hw^d`fqF)_$-R}wG*;-~d!oݐggPp&-N5pM丝1xq1d_)}.ˏ7|}w"LqR"^8 :p"ɹt<~oy\ V]TI=LHΤF+]~ UEN73iWU-M nUvҶ~?_7.k"Edz]XLΛ->S_n}CQ?Zo_>V៏w5k3 u0d $2GsjqD;.񫇞{D-.*Q{>kd$ԣ#jh"&.aASÁp;x(sˈ@pw'ɹt,m[.~t-#{I^yS/ո},XĐzK`P}Ea=]F:i% Wfc e@a/PIZ^J}J(ܲd\:B9~jaL^R,PaW!出흠7I?#r<ɱ1'fi{}6EzNwx_|F*a`uqg_x5 Q5V ] `И Rfjc3 5Ъ %"?۫-#x lBRH{3$c׃q\F9^_|+IDX%Ih;U1~8%LS.)~"y=qSF ( u,Q7[ ~Ff(N['mǷc9^nX4[Uf cCbqBJ.h(I5斴gC1fZqGc[yX@EX4;}>g߭>5cC^Ci6Gde*Q%?t1s+&hVH+D:)Ժ&#DhJhCЖFz4}d)4\bjM%hd>Wf5g.Z[0DwO+{1SBT3@\Ԫ*T[<&]m=RD:0/ cյlo*cwg7,K.6o:@a nL@n9ZQF@F2jǶJ.!5!K΍Spr%.!2)xƢh۩MǫGt]cM sa:4rp';iگ,p5rG_R@mP65(5rCx;]1i9 i^higM蒚sW7\; j`$}Ru)Ibriɑ9;%q8ƒ~MSY]ʹGy|(ò w>$j:n6).3 %5&gh͠%3)?Ƀ{8x9~9bÜd^y;\fi)lc6 (ZUXc&ͱ^ܸET<Ѫ]&+1-&_PSkP^k X+] ɉobent-(4|n]mw:0T`4HaxD1D d4S,`o`U]d$S;|КYq4ɡ0^+c6GwL2l?~f8 +p5u s6DϛN@ZNC) L#rw+ɳ!^9&2(fۜV\ٶI6#q)gKU=1v^>{1J>͠oLq*(( m{3jPۆ 4x5FA[({ N< pֱߧd =G{~C./ն~E4_.&({9HR`pּ̚s?̸v2Aڸ}ʙT3>b}exziA3- uy(ڟ;bcMУ&;Vxwpݗ;InW& g64 ~kvz+o39 D aOG>T?=ٓn͠P][s㶒+*=ظupV*'Dle6 ~Ś}rv4q? | +.Z(=Q' cƚh-)8Qg IɨN672Nat/kh_l\|(Rm![ %YYh>}xVszI5(8IBNY`k9b޳ cܣQ^fFHx֯W N^(uD$,8~LFM+6&TlX ZegeB͓%d2]!옡޾l ZBǨvjܭGEdºa0[I~ rBX_oFld+w5\}k0BKY!PZ6>م e̎[[ўA+4()6ViC/j!$688)Z(BzʀcIaCP6ѧ?~˻a*[^#]X1)c#Vg}tVȰ۞Я'2iy+I=}y\\6C/O(vA'<hIX B9(0\ ,-8 $LOI^VBUY+(vhEKw/v::)n#E̓[:"q V8Í{Ѹߔ@fisfT lvzxOcvR(蟽v͓-"=cͧJ@6l.lcC} X6fe~Y<}BˊDuO}nQr˅_/ wt/U^Ht,OfZe9#2#$YZ`-r[65HeaWXx4B Cm7d"ܒ?IkRh@&CUû_lx'P8izUO6k0p(;y\sJ-/Aq[B}1זIP.k6Zz6 R5~,{>G4éLפ=F'c_9$נCHWYr?G)]''J6ڰ2AaOvJO]_ơ_hwdބ* `:$XfeAhڇ|89ІAHJ ~؀{h?aC=tN/.r"K4g/JəaTO#&4T3yynOZ2gp BRVsA5ALTIL-⒋\fdIy}g%ոxz2M€awcvЪsO1ٞm喟'"[:sIh{Kॳގi伻3~|iPR Kis &xNvS;[Cl"yl~=޿p T;55E }EX\L^;+F6Py'+P}*ռիr"3#T6TC(2î\3~!cg{4ȥ\ƹI:abZ7DH{P"^j5쾹?qg_IN&wv6\=~k]lG:q}Fa(1%%j J3I$F1#&Y7ia*n3VvӮ] "cuYzj[D u۪Teku--bR33\8?pyWu^ؼ|sVņ1RĨå렆V j8 XP=jx%wN7(6Z<(иzkA Y1$Flaʢ=wϽR˔넱½AE; Ǩ0: _-:$+ݚ*$V೚*P (:Pf:Q^[%E: _s'&*C:a֜_O0cj킡1GPZ'Et5DW3jxJV׿,7۵-pwfU)ٵ9 >j7UX)GjN.Y=:?.2"CO O"e9B*M5`,3QO1ࢃcPvZ<2 KQW"d4=!؞/ ŅǤ"K ݀$!TgJ)X/i2h>6}\:\*jFjZϫ'V;n\=<+G~KC;kςN9dNL+2^&s6'DÙFk* ³eAS*0eVnX` 0MsY5F\.S.ZrlM;A{".%GllQyCs 1h_[ʆ#qn[o*!&T}f"LƦb*jMO!]ñ2$O: )u4`:$-y]%. `34(h\J9(U)'U  V"ư(57uEۛ}ʏVNT.biW 3{qOzaESh 72֔;x`4Bih>gCk]Kᜍ*|Tc}.4Ρi \]~֛;p}FP1!K&3pڴiBqH,&3 It,MDba3 ,rHW>յH;Tv=A0A ɝV]\T*:EtqE;H#`}CP 94l.a{q -Vr}L;*(Lي4IFq9ևm%@T~NIַܝd2? _~||-GpeCp>-G7teb|EMod\څæ #fea*I-޸Rcikp PQ' k*EڣNR^h!T~i-s cILKfNTQO!޴F+@'т@Uy e|B|,5߉JأF Zf٪*B?Bߙ5lz5kYc6? ]Yql=$J!![B{CHT<t#kD=$! -thZk ۿR-?GmMx+?>0h@/xڣ4b68IQZCj"|4/竩blƖ-QseJsD}F4"+/j4U(HuC toGI4JY( {<y5$^R$D)+͘EFc.d~{'*^TXW*U)λ)wCE(Fy(J09TЊQP3[(FyސD!Ӻiu1WQHCi^0.)k *ZIp7YDx:":Yeya PC%2>[Z'#qoJkLrGNŕ>G6d]sEMNTarA,PLH/*'MއFKhƪk$ 7(ꗮth.GEcCC]0PcR4$X &|hld}VB4DUBg/'ŗy#<648IװA :Gaޏ)_dH{XtދbLjqC8Qho@n-2\v Ƥ؂ݸYvx7><|;lQ@L;$_donhN?-gH_aWx1N&oD^C\8 t|MGS_%_U/.O}'<4Ⱥ+I ǎDחæCY!MQ: Hn[N >܆s=Nh]d紞X&nRźmK2?YcǏg{2ZW7>}.4[m0RW9;iGXt?Uc{b=g<,e|!V[^(W88/4hI-aNr6y⧡Q]!D?~iϩ5aNa,ʌoˋ:'.'.L' M+`if 0ƽIZZ hVi&<{h|3C$˯n_ڝ{gKV_J,"B );] +*zJӊNՍ7RVɘwzN ےS.Uhh?]IqÚ;aH~،'KtǗSHsTz@e\0AxN*I~ha?#uEoЛldw 87ѩ=RF?W;K<"`m@:s07s9\TTl+ʂ+T%C Ռ@D}|--p"]̊{ D,3Х /}xhHQ{zPNx:29bFY+w, RD;vQ:ch#6߀YFڥ{K=/SFSea[=|( b%IN3 d59\s u ü3gLv.6`z"- zW4!8Hˁ妎q8bC ^hѠ6^jd'5(I|c]7w[s#9W/ e5?>.sChJ9Ӌ%r?Qʭ-uhN5)RjpaR'|NWL(_/S;#X=KN2T2G◑Ap ׂ/{y@-/ͽ\Sc%Nq8xVH}I垲YX7$z4B2H Lr3'9c9m^Zy'kFЩ-Skws=WԬ&ake%,J UPkI#}Ax}}nE@(/m=bgGP:ӈE}caY'Wj쫷B$Rmr@Ȇ\\[ε=a(c;rCUk5ó폿?([i\ Tuqy4r+^y*NCuFjZ @w ~}jdOޠkX ႐QcQF zh,z2RĽE^("ѬDDjM/ͫ>H'?p/e15s^+bAebA$Bk 窫F`he*N<vDHoFը$hPBǗk^V$P-]]_W3KR#-Jқ~]ZSss1DC"F "ZQv# dAbyH-KG9FEIf%hKA`BO8%-CTH ظ1B'|5sbԚ-ĈreXHa%cvUM%{.TsC}9@rB2r(iɓoUʕU\j)Ւ8J\iԤ._>bJՐr3c/F:\W,aK*,Z"y/.#>cLRMLRP:䭳 r u1TpءdP]'L%M)FVZVdjjeϮ聸Rk-Gf:E2M' ͮ S4מ+ mT<PhHIfWtӆrQ+¨\j1IxyLoڌ*I$aOKIqZ+gl/Ϧ.d8;vQ/ e{0j7 >D,ne8ǧd>[:/y<{C-CQĨ_j>T,8':oSY!߇Εmލ;dP |'T9{՘d> ZU}7f \dмKмֆ D|M[FM7mJБfxɀ=D0Z#ڭ}=w?wmM9I;E2?rldJWb ~FN`?զFс#d $dI%F2\Kn0֢Ȃ> iz&юNrxs7~އڈj4^ueA_"-Hk8J5/Ih(Iou1jM6bQpA6jCFŹ$;hy*FTH V@KJMch9EсZ4j̢χ Q|#!:Fp C\ЂV֨M₌ p=N(a6EYBDz{ 1?QYo M&r 6 nWMkMbD\x` m>AUs^P'BpR RNh2mʗ17YePc2л\Op"x)`(xPe0shI0h Z5[K" &CF3ݳ̊֎~6mugG=ma^~1H:a--ip+)C1(EkbZ^=Q/G_rtlf7w*}^fbK{K߾[{NWNg/IW݅:辏Ʒ*BQNSo0u~g8%ݟ'V9Mpan),;CZTF UM9;6LcI %AἘI<4 G7ߓ Mv7+ZuЏϚl:vWǁ_.˽X-D7_&(CwVyog+&O]ٛ&ɰȪɗjkU7Lvn=hJ^)=~(ʻ6ee / -aw=&Wϟb&9)S sW{5 %77zxe)?&)66<[Y|OqtFrfvM\{h̛)Mֽ]i÷Yk042OkrFI>-F&me*%Rb!ik!L{f$(k֠DOkDMi]>ʧ% TΧmB;J)yHI 564QטC 6!VÉc78!Uew_ٻ6$XFC,l'@r1l'=ݖ6#8"U%gHZ"b)"9Uuu=}Qox1C>&:hMf܏ zC(˨M&N?Ae2$8wE'/p8t:Wv:m#! ׺Fwpl<Ż*GjZBƑP{xt:{8SJG<249yycG#$'vL_Ea-#ݿR0%c)Z"Ƶ,dOt%G-yb0 GJBx>ݍFWg"3޽ 'ߕ=o>-(?&crr?_o>㱉eqhbB/چe_uxȘY(r) 8V_YKV]pSylRZ8䅳hOitKvuZӬLnNdG݅dOEa'Yv ʨng4'܀p˽mK6p-&Wϩ|wsUiʅ7y.QX%?A_fnUC෺Wn:LcU4iԹ09$%t(\yCZ'\Bp{ y_>/JYl2 ;;oWNn2',qQd,lxȘ(miߏoVF/2X?\v _R+ v Pw}ȧr^=[3~*=,V´G*<:<3]KY G_t#YL񛔂]|K7ЙfLnq&,Um6+QI[wwSfv+K}!gBDk l uLRea 4h)8Tl2 sQ 5FN@ġa zqa~;+[>YG#Ar[[ ǦeOj[S !`#Bt$H;"I!kB =SMqq< KLKu0Q{V*3L! SGdY#68[z]QфL VShn_KV5#֮Qj}]a=jF08TM2{;FśxݛM*1Q#4[[Cmwח?m9MkrrNk 'm"aɁWI~DP~椷m#И uruE+!iu[PF!Ѻz}qii&kHpjkrB'NxTDDs!MNHى4: N)DM&q}(x ϧ%‘čܸ,jR:o34h>ZdM"fueol5]?j㲌hp+Yм %6F߿\o+e)ա>e#t6E7,/W77dr_}5^3ᠼōx7f oͯU tg?{ō+~F^u2ook؟M? U2*qEmR+kƯC)OP;uAcj@1 s֔{vƎjR5 &(9bυaFkA2X3TP}r9Lles>CS).a>FZ^yC&0&0I +3%O!r wAԒszow+W`єJHnV K]*Yr*Yrի|~ͯ0R)UH </vsN f ݰV_5“PL'F}fJ fY*;Y+:TuGŗ(-+ޖ򥺺g2O%Wd:R)(+ސ7gW;Rf\ >p _DDhVa.}BП?:xRqۇ~3gx]2; wS8 pelS.B)A{v9/ÂIy6S=\'PTGqjg`te{7ԏoojPfXs,9&KNX *@ [[w0rCR8 Kcr˿8u H@>hՋԽ8f ^>.78Sg1 @Z@3σI h]1%⊩Blo?s0X~{on6zӻ|k12rT#M\쿑;%)QiB!ɺ{F2.^w->)Mlf!.v&!CiW`# Bcn8Jkڼ غHU 88kut͏L :(";d浓ԇAIS;+Z?s]q s1 ^32Q׷ 0q/ֲ\ +;4f9NT ,(`GCX{F-X9 Zpdή6N^\`tN\@Л1tY'ZDYS)yLN"ka/ 68lus@\.U]j?k ;y@ jwu6w->6cb^oU0֎[%;gnM(/2)@ANgKAQVZ]n_B[D- l䵊VSiiSfx( ;|ܙonra s h `ehaBkFv.aCHrAs炃aWp TWew)Nd6Tf~E3D`k]Lo(9;0 v2T]r#Gle +df tBO3r5W [B5ǀ% q:.ZLӂe܈0'-x4gͽB|Q !rB# ? 0L;jpT V(Vb y[|h1Ъy dY&F{^<(WLW(Wm,&5n~4RDYJe^7&h~A݈n.W+z >׊+ ڲ=9ŒŦ\W,JݨV=.JӍTnԈp1tҍN#DQ:⮞lA Wyoq3o]7fNVjvHb&u$n$cQvcD܎9*wQc%)F&Q'b)yyɳ}rcNɌ$ct޾ZoƵlމ3 f:y]f^Y+Qݡ{Y(ZOկ=gyoքfZNvv;On׷kqRw Zky+V<ʝM'&^f!kZ{ؽ ePhw@@@ͪUZgZ6_aCn6%~V[uoݺ;RxJ((Y}g>8$e!>=====יfƍwm(WB;t2g:hcBՄj=b 8p@:ċEլexD桍Ov8Qhjc:v9ƒ05z;=,S"JU z|SfW3x!/PUtUf;S(*ͭeXkbwصq8 њ BawulyEA@CA>dq&z(M)UlwкwpO|u:=vAz3 C = WA_)鄖n ɮBvcYJzP܍!hX>E7dWg a1P%=`jN Jf"s3 k5@qe) ݍ|lohWZn,ΐ >R5]1dW@Zj$ɆZ$W= z$+X˼N~WSt{2=xs=s5 .z>9-r)/ut\5ŸLqW"=\̻HDȡy5= LtQL ߠfĴۍ0]1dd LC 2AЍ!#CvwЊ0Cd9\"ƻ˵2l]%PPAw.q2+>"@C3( ",7o/>^([ܦ8ŗn_0 [|-cқV+o.t>RIӥyXfUC hI9ci堛4壝ev,Њ/y!;dx2beSɮBv[ȊzpeSdgx" QHЎLvuƐ8du2.QKvru68:ixgaaR-}D]HBv~dWCV*E)s&E5+.OWۂD yA& KAO# zZ[}s*!mD*r5c@ .镑0w镵xWp%cޅdWg;(W"5eGaCard˕j/%QRN9ozEe5Y# )B!AHBVc 1V'p(lz@Pq=%} D∷ 0[ɐ"K3]/ˎ YATƻ!%@TN ]f!ѥ(-쎃>1{W&%qP㠈\ mTJbw,Ke.VʌL88 _a0B_ b~.d2}tbU }!n`yS@LSnfX~hkC,+7oO>!scNp}xiOb'2M& )&^+K eVNabDV7D)mzY_nd-_@"$MlJ/6"`ו""v HO5Eʽ`)~Ճe}J)) P*Y6od)&{ijA(FM- A ,y^IP SLyjFQ\nBT/Pk˵uKRL+1o^qqp>Au&U, 4 LƢYILݾVftIUYgZqmק} WtԸu*- } | 1Rڨ<V'pݸwZpY̑}ؿIhqOEڣm<,ZTͧ}gS\YV 7QF^СC/ FA?Dϗ1J#H%bp3aX5ƙz$_-o»߻+u"{~hdݧXwq6EVd_B B1ݢOnan]),ݴ۟Z/nv grOUO a-D:V6߽w77)8RiQ5%`hq fQ!޵b(Jc&k;q`]ad[FjAz֖Z>vLVC!VNsrq%kq{‘^Y?_Ǜ<ݰD"Qƃ ԩ>MLTQRvԆܼVLJ҆fnP1n.G e+<]WM!{?̙blyǵ~|bVSE.~~^i9 /xG}(ׇ*ygJX0NkNBDLeEǽy~$qp!c.}Hjm.Q+Y9{UU,Z-p);*%J0,@L>F!<">cA#eMBFP0sYw!b/=ח7o/vl}Fii܁%_.~~v@,Ys4`m:C6:hCzzK@;'UQˉ/Ve`I2w5a🺬r/J^^a aET-{׳%MfI'\ӋѝMW5\&;}E4D&,V >Qx_z7oE1<J_˟sWCY);Ֆhp6Iovz6s|w;j?QnWؽgx8XIпM.Yߢ;V]67չ^\z;+7(6En0YQFPsm'8{5>dt{I~>Mڳfa2?BOrt0UMA+Տ:R⯳'T|9}P6T<ʗ$quC#uTmssџFQ k PEA@ P-锅G^al=_X°Zsh$KEً0r3?wtuvˮX}wSZSQe|q$ y}eIddC+$]ZUޙݭN'=>) ՒeX] (8Sr0$[+gU#A;Ry;%*FVWK|`^ <4y͆:oX<{s߮.Պ1u}g?3W8AhYK5y"UHF͠z'2]IjF'J;Y~Ow!8d7NO\'H|5ҹ>ߢWkgVWؔˢXB'NMZ5yxmϙ|,L%krpMD [2ǖ/<]/@"Vjm]q`_*waC+ŗ'DdAW>#]9 ". ;.候ac؞cabHg?1 = ɦgbiw|8y63|x1 -e>d2X2T\YX:gi8IY4@GDd܈j*O&gXOv  (,D!y`Ǟ9(m#{7"hs J@7[<½xi'm;]kB\lΪ!PJTh LVYq`0C.@ 7p}*F^FykMG \՗d8e#S5^۔oY)%mܕ؋(p0qs*@< LF!w֞>8$mSt#X!/:J[ Nw!sdG'#33GQ"j1JA`ܹPQ;%^7 vW#);ef] Oin'il P\u{%wmXPd}mv ARc-}_Â|k[D4).wH5"x'p \'W$h+,Du9ud%)N"e*TPC΍`U/ %3w}qƞ䧱K"$B!e8$hDOrY4J ^E[#?o f35=6\2@/p-xNZ`݌ټEu$+.-on2U^o :+DNkQt}s@m,\*zOXso0,[ţ8 !5{9uV"zo[-["m3.6{"Qlv>J>k@oGaNׄ8'ȆD6V8JǶDz]x@wcJ5櫙]$i`vu0~N=~ݐ% )HWJ 5{YkTsގ5,Zi}*}+_cכ*o@;Q=۷^O;e/àxt{nۑ~U<  W$_drSLB&)uWaew*+?v߮R%+IuZ/Sپ_7IZEf#w8\c6ލr3}hu9ϲϒ^P9? ϱy>t0r )ϝGLG'fE&aT=8J0ت<%} Fʮv䓢v//֗_/|ɿKzanq]H,("i|_$FB9,/iKʳ%M6\F=BN"rcfbVB:3܎o<²IjLB30  .R G"/ b ͉=+*哘<$9Q1W9rag xDl̵zFp>\aE[(p%4 3ǐB i։άl4Gc1DJ} kW$=]IT+:WGgL}e^i?*r"c(}.iKzD*]n%ݺKb3>CeTphO}rH#}9 "aKa)m.H7k¯5Q`(2>ҧKkWTU YET ~\6iEv^y_ٝ΋vk=s_r tyEic#q$Jb3p$+MΑܘkMyJvRd" c";/"z(΋ ȯ[\HB?{W#aͳH`l":X` T;YE5CDagWESjI!AuHh#̅\Ȩ\d)9+1FXg$5"|E^EENݝlԿ=B/RI,2S<'>/GRc$\d)<+_2F^6y#jQium_/8(K7o. "0 j؈Q#JCv&6f 'ٯȸ,.1?9 Z*hC9ȭˑ%Db#&Mьr /Ͱ^)B ϼTd\0K s!' V#cL*U<-h07VS?E0ῨErw_݆\~x3|i w^yJ#( kJq[/j۪Ft_P݀uhw ^:1hѠ`~LmB'<} 9ItP $4._1#F3 ` 9r!TZ`'3NLD9ByN;\>\ySTiM}J#?Ϋ4n߭wIhp"On*>Y2h0v_~t;HZtf.BgЙޙzm3ɸYT[)LNNBs251hַ}Q+ַ}VN1OvTZ3vRTAw 3z7Fđ1͛,cX{{a0]x J<x1ΐoHK&6c4OKㇱ9qiW1/12WIGX\cotrZ0 Gn0hwdr0JD`,f:h \3;kDb:b,F -ZbX> lʈ-zhs>EL6&CZh#̵)c 7ʢѢ9og0 b͋,4XrhtX8SXSn@Ŷs+Qxw(oI"ݦv:A0w0BN2oȔc詠(A(Hj68*FM.]ui4)Ґ1Z"Ld>@9)RX=aE&mPHS95JX`m;&OZa,N]BuX<0- z nL،(3˝'Zg1dY2U9 4y=:hV*\ȤÑ' R?SXB(.LPD"T8A9.}`5gйs`l\`ˆ ָdYy[4 ZQI^p-sH=bZ%C׍Ĝٽ=C ^.8 G   pyqM5in?|M`PwFfKvV qi/Y0lӏr&ivg9j6v'D'qOAf(&~LG{F=/ZO*8.xI 0Iհ6J30P]6J#3}`v`WZPFsy`i!rQɣM63u7Д3ݓ/}2gB5WNJ4ۆN*c~ۼݮ w_5:S*ݿ.r/ao^/GT++d@7hpw/a.A&KKoyZ2Lg4וܣ5-$ԿsM1jO/s;&cna $˿pFZktAg,.}%= n)b*ˉpXlGċ->ApNESGycY¸LC%.=&4#JϞ 2dQ!~&t r9~z̈́hgr&[ܘ>(e 4ZI?[(swmB16&aD hKG #Mƒ P@4=Zk3!/Klw<!mbXv9U8_#T-G<ʀМ FwHK)Kg`Re"ʌ(iq9eH֯ L&՞cɇc2&s#ql\f0R=jzSSHؚT t o֧ 4)> C-j ZQ-g{j  HJ eF: Oj F2U?F\u8.c[z)B+&yk]:?Da$QÝ]ȑ0 ͳ8Ҳ.Vj)g݉X5e+Һ4 J {~eS> f(i90 ݰ <߷IYL1o?.蛏ޢG1}{s*PDV? hPۛ/LS<ըE~^bC֗SyϓV5fQه_{qz1`r|5F+ڬkQ^=TyE3U+:kH퇛dzB v"z;o+ƬX6G6p(fWT?C_ ﲫ{َ[s'­r~VU>>> #Yy4$?fM86? :ۉ.3!^,HwInB_4j!l*%s܊ǖdpZݣn$wO!D[(7G5{6[{Cy^iyV?ލ+Ƈ(FJ KH< LngPϿ)e) 9 wiPo:y(J.ƉbALK!)w k,(>x,dJuIcVF$58zA+&%d?ĴLs t4#FN{ub櫷ߝ:1Z=İ5bu7M,$B[rN'F6׻0ػ,(Yw*5Ci~vB_Ǵa%yǰ_#yoyi8۟-&R;`RW7W`]s Tp;m=;xbMI fkNO-N*'ᘎϖ>(g-<͗P]Ajgf5Zs3wAy\9 M0֪k+oQ2UȽLqwq_UeUvWU!FEmdRj0C)rDL\֮X>V&^;Kr G.?|yT9kEYn})dOX8(՝!~k@ JSvQ|V7|, >&C>vOhMJoq\24oxk]H5c!c5cZ5M%)"cF Z1Ɉ<(\w"J&(!yr6Ieyɍ' S/qQ^fR'U/Mߧeyh@* +\2TaKc(Iļԓ~8i$MW(mϖW@<Xص{LF=KJJf*kSFpۦr'cP5vƼv';rg֐!%EOVrۇ?CsЗ tKky:O1Ş!ԥЧ 9F`Gr,"%N"k<%Bp 8pMS,.O[ByFM[GۻAki-ׄv͐oQߺ|2 #lR4d:%! ICHև: )5yzN ;)[n83Mb̾!љ)({(i> ]+8;iLBٹ焍 n_{oF7@w3K.HpAP sg#yuid҃Ir> @0|%\HuMc{ޱ01Ϩmfs!GQ-. Ńy`\cy`j/bZ޺1py;wELt ͏fMㆷ|ZUƛhמ~0+󽮣]uWMb@<{i?d/,/, Ɉ:B{=]bw Y\Kf:1HGY>{3>yA|& thIt&UR@jwDH>d3{97г$ @hzt+)~\$/Bb"$/Bb?NƷ G2EPHk9T{2( :WRk^>N Ѩjޕ<9''o&Ժb|:H$m_tVD6`v_ f;٥mZ`֚Zr2C'T?XZMOш2z!XKcPZ9RN_QBSO8FsAuecXYB+eCCV=Fq<{F%itX쑌$'wlyXx5k -@sGje&Aل* ]V;jjdRq̎R \ *IN kBkȜ4x97>De<35BY-VX}+!G'Q~m3ɸTfcSJ3iiBs\5@:*GԵ N-v>̶FKi،2Ŏ/6ו8*WE@QAt pRRFTPYDD( -T٫q>3jaac+adw=2??ݎ7Qqe` 漺Ye UInH.^5JS(/>kARwL|\syB{ b3 r|kƙ3FjBC{ǔNS\i.ts2p)X`򭕧 ~$[mi??K_ԄD̾<멸 5z!eY\)\c1򖣭Aa+m$GE6RyiTzznI:~SLejՍ*+d`BXA0dsQ^JiGh+ɎZO/5@hR{*GaݹĦvcFis]Y s˺xiά 翿[W$<}#~i7?Fhg_$XA_ |nߋtIBhQNSNg]lL~~ww|q%с۽۫895ݟ~ڨB%9Ui|ԩVߏ|I LJXظ坸Ni6*)_z̎6 .0S@ᒞC0LMD}Sduc<ݻxG ZIcxvgbӋ3X.VMć 5~T$=zx]f7%FՀ (C!eϏ7sOs,whvnMB|- },{3k 3gAx_W 1Uó '}ϛ)m""”}AM JAT(YNi}1r|@9X>B=/W6%`dEٸa5rwbOohk  "))ZQE=tg \mQsCШ(-W +z%epi棣ܕ)i9{bp7t&hBk[jP8ŋX:Z( d%ZEP6 E_!=T`MC0iAkT:u2sr@@;`qRueikPېlBֈo/]snh0:V/{*'/&z4M>睼|,fۀ߾\-^B.U~UUitףڳ&.F >Kšnٯ.ǻCÙDͽ"Vw=A0.2Pnf/dT_^\:'}E'{CAvu)4P0PZ \ DW*jcCE|N{j ZrXSGA9oNaA:r4etX8ז!ԑa%`pﳲd2H9aڰСe뼎c#Z(Fː˺fk}3RulnޖSRîd Nשz L z6N5EMMv6N 0nA%l3ϼN$YHtRM̆:[bk5c2+Xu]AND,7[nX7"ߤ_*F];xv#iMU[KCNG/5aۿToBKy|xn77+b-,-TYui+]tHӢ=.ksfqbE-\Ŕ$E2ڋLXIVs?'![\1z.Oq)_ϕntMKMlCM% ?هh1C%;* VE@.1bJ u4q' Z肙H yY50ǂa@hw `r5 x'yf'oX s@_Ȅt6rӋ#A)]>JS?Q2^ ;8Q 5"V;؎|ϺBl05rjE'60œVK4s%c墌6:(FyVX{:Pˏ HB1R'g>gϐl F޵OjA+-cnʴXԟ?_>hO!ʎW}m'LW##_nW_S_ jea4NѺ)v4(9fLƕZLv*:>d -}:uၩR)jifFћ-9g+d82[&*nܹaڹa} R1 \ㄘePt ,. t$k[1lہZm;:3jBs& Il_޾=)FYXFxƚة755-ևh(z?A6ڽt' o󵇷 27Ag-ZӲ-@jZyNؽk-87SA v6o? O`jld5X#@~[_kZ0_ێ[2T3$'rg*ZO}Bq萀Q7ok"a]Z :7v[nm 'Uuj]Rk !(-ZGC9B\޻!|B{|ETCjz0V9״%.=g Gyveob9KKc%sQcyXvۿY%;8c)"9ِ,nCR5-9-~suKϚR S<5pSqu(1i2s^^`4 Q)l{?o' \^pPE.<1Cΰ[60 Fd!š斅σ w0~~bѶ8olѶ8]mZ'_nU-١U=i>I@i/wa'\.3_.6fp83::ǞI}OnndP7}\ u<ةФ4 D|t-ݽ;?oCnW[$G'Ai5l%FcNQV;ߩyDyD%kBBkNV::`+ ťL.;Q>ĕMUlDO}g4N(Ur$lO#ze:XqV0!{l3l*eQ8ǹ:pfeL-耫4hN"9R3ZvLJ63/=.y7m6V_ iJ.fTh\:9s;0VCG)PReJW*# 1J4#ya 0Q6R- 5,ٌJ\P%1(:F1!J)- FKҘ~5AT)Rjsk!߈7AJ?gHVD3nu݄CLw`NRЩr^%(՜$Lg*=GP-O0 >֢_4WM*I8v4~2qv[a u3=L//.(0d 7g7_~_]R6Rj~CUd&3è#S2 [v5ʇ?󵦸Ggjyye^}ŖXU?M&Ĺ\ [g!Lѩ/ 3ȔT*~Mv}ezwsL\aʰ *pV3Qtܺyou0S3.(lb3LMl05֛ {@{ ks1W*j>3KNsCx'RFQB06ْvVv`?t,5Ô|0k|M]::R̓L$&R2YHb/DKV2Xlh;P+h;vgdv{=ϜR8 '('BR4k@)4[XxArZ 1PrbmYs+c$^xA1\8gkc rr,iiѬ a<> S;y8jptL0l/y|_oU|ӧ?.?[Ps)QΛI>M;NA.@?~]@q0iH!%yaƁK2uof;3;2[UCArāi#FL"yppB~TLcګJHch%-m.nH:j9)Fg-|qz1QZJ$gfO(2 9[dbFBGK8Cb#Fd*k3DT2 n*8Pȗ(ӈ5 Hom )%)9Ԃ!$V4 RsA7j)r0!,9:pTK ]n*@$!CSո<]N]Y* #Іj4B{=-Zl E9eJ5FP&5jCk5(FC ~Ꙍ3){h 8SADc5LgPF,~4Lz4Z)6uw(ggI~|=ʬގBG ;EE̍CO2nd*9.WϜǕf<<"?3w)|X`6x$?5[IE5uc? lqȈd;K"H4hΰ:Sʇ-+{wWO-73\jKԶ]!wۂ{GrJ{%ܪ+m/?WOp`)T{(yr>;U4]|E|:=#?_vl\Us`ʟ6;VXF7"cf[l6.0LsXӇ_્ S|1h.|/Ә{z7 ?[WX]aFѰ>T"j `^KO >NFF_bXR=׷*2T|:Gx޴}f_]]b׷cw~uN`j<^ LS3zkc`^+;)ϬUt+A@`uJF&47CZXU Qc KcH꼬r I?)u# vWԒ d_S @D/Ioê%[,4zrw݅J,+8I_=ن` G m(6Cv$!Gj6ԸoaQ}Ͽ*gI5tbŽd#[!ōL\n{Z$pC&%w4qeSDΌQ\=K OVrHD ZAչ4"&9X)  2$tn`DtX9 BLiNFa;Ǿ8l/4~ UZ ܉yjǑꘛ2A9bǤ=PO|s#WrMX*)ȋ§:xq| BuuVOQe&>2+t &NET6`h+ Zxi)kw+aTmkfB jg 9/jfwKzSmx]U~Gb5=w5(UpfNx&@:NF`yjDޓꃂ0l4߷:wptRpM9lԜ&M%D' 2*p$$!U> Tt;Ι0;.FniP*12"Jg$[bg0DmZN/785Br)ye:KCnHS酱Qc>wZ*ĉOk(dLπ^|)RD;YFYB K!Jkx$N@#ג04I1Tol@co4s"uyj޺dᤩz><m0>cJĻ˞x{r Aw( 䇇i\?d<8e>}Y!9yRLy8n(/1GQ}Ft8Vt|!.K\!sS}$kg)'n]1:]ĺ.2-ںufuCC~*I>y׺9 :M &GnkА߹vҩ҉h#ֹq=ՊikA  W(x`IRnu:"#RGbTk%dK*)6b V Z#D`&HZiNKDS%v9YMو rgxsQ/y}m/t>X|7o]/3;M#0|ޑ瘯}8Ũ*%lGjoH[#b%mzG'pLi:$xhͅWBo>lOGH1z(Bc'wsVXmJBjIL#QA3d]TtA[z)ha0' nɡR: G:mQ6^Fb p.*nԛ^ל] "yރ"22Q{e`q7Q$5F!h{E_^ _W0Ҹ}fZ$b$&B0G/d{`\ggR0D@yL+P݉Q <jg "`P: v!F*#HG{:Fl7o|E)f%B?[P{ˈdɆtCj#G<:wO^? ^G̹b10+ODXLbsu +Wi8+ԧ^7E$pUӜ wzI/':IɑYIFT"s׏(>WIs,[e'ɽKkn*% \ 4Z<<]Mޅ< M>z&j*Pl~r}(b~x\CxaGY#<;IjvG}V!BI;&OTBP|QG_듅Ht,.B*OB)>'nJ7tɃHMNXQ.M,91vH3'?eХ˼Cf=3jn.aE}"%|.7^Ӌd ŽSqx7Žpϼ ǀdiy.amcQdR8X|fZwƒ5bކUAu,{pV8&oũVHy{hcmpLeUzlʹ?8}~]isGr+ñ;Ⱥ&,2)i^2t EYĂr%3ǐ@cp4)Z =<뮪g-.+͢1\-'sV$XLNP$rvJ }}<ޚdr"仪\~duC{n<$IO ,I*5c>`,1~(pެ|u?[ZC%Iw3H'CVbbC<Ċ~97B0t>ܬ94}x076DKx( Mنys9`'6~l}N+Y&tcgpjuKwN PyYyc?-s+/eG;zcVf/=paCH_-M߇s /Eڴ~ͫ<]øFk&mjfkwvde|T;x_ze>==Ds w>osqώgʛ3|d}3 m7%;gߜ̳C/|Ф?|yu| {W|;{KѳWf_r?/ $pحp}ߗ/q~لl TCQSb:0ˀKdocy}A%[/s}QgZ~cj~:sG\j:Y[txn>E m`S>gʓ~?a,r>x}bpv͢R;C#(Oݿmefx Q닋M<7y5~%8`Y𗓆"n=*߾?g-_ .bmyɶ|#nEx"guUgkeu*{d1Xw=[ܶwE^W?γ2@&@cN`-XtFirF8eI^"r02P[F³#'Ws̩ 6z5@Y`f^U6yr΄4iqQTx"F($F5KIDf(`R][RF+i3fm Q]yW YR N ě*' h"IZ(`R"v6v.(12B+B$ʠQ9X Ț옰=>;@?:FE{3@ʚ8C!xUEE Q;u`foyFgPzl ">e*ܜt< ;VZk# tLX@&áZB4#S({6ĕ={p!a]]jfpzZ(RTcE 56S5\SA冪G9pèl{_v;_ S ۍ"jw*pWOՄRA(8ʠF4P iJabUTZDdl:"ymBGpW'rѡ e5\n?`qn}aR 1#`zvGT ⮀ BUL1V{D<`!*yI su\"?-jT G* IэT 8 P+3!qqaQQ,i=ʠ@@W@%.2w6X)%oV'WPRX_sm>;/" J/X r xiWEclfJa3P@ rL`Hr&(!2Z#s8MDžCFHw*2*]N BRF8*Hu 9!!(f 指L_ .L[H):Q @5cHh@ ,l`;xԄ0. 4`~=mq+IFMk c* l,X3F*]"&$2WL:[AlţV0u‚$I)D= '*)& rj%%/ N; >.I$n{Rqu)"de]s+ 9hBe,{ $7>mŤ TJ~3G4n4n(W=WW>mmX[6<͵=hzoκٛ:7vgY#<]SeMJ$}a܃b/ n;d)›;۱Ջ'8/^ &vϠ./:uhbZEY&鋯~?h 1,~e?Zn\L8BhkJ՛VLl%%=HTwuM0yz LkK?lj` Vmidֽv> k:_cN69dTt5|o&I0͒|ҡɆ5#^lv)R@JX$ Tk @7~R/:y|ѷ$Z^wѽU9$;*_i: Zd+N̊OƂ#LîLEWM]00C٪7.ZapsXA1e#E#P %#v;/b˝y&@#k[N~e!hמ,'wx] ^<Y/zYtU:&|~q6p&&ߞ_8X7o';,W8ƕ"d܋  OGX>lpt?&:3MǛ6I]qޛgE+?n_8!Wg߹@NGIB ;tsㄇ_~5Ӏ.vv5lCKޯ<)SE|^$@{Χ'l\!r|,P_}o;%z;䋺@.xa!7+ɗt ybh D;Y5?iKǬx2n%v< ?#s!8vi=*Adu#j"[L{mB;EҶLt,G-Jh|R.{ˬْxueC^t~N5]H'9o2JNYݞPj;]Ah]ޢ9y'⇛!7G~"!p2:&KŃ{OԖ8= $q/ gkz jZqqQC5n+` VyMϷDzk)AVqV[)X{[}J&fBz~ >,PϦoޮ'zwowm(7a SWN 7GD>'uq?nFit}m SWNm-c8UMݼS@>'uqQgZcK[nCdg}rJ >覼n}}Omt;|rkK[nCdg}rJ;}sM3'z`س歭Bwѽh|B!sn=}פF6w;-Ĕ_ndgrSL(;g ,|_ś_٬yp=Շ0,C*B6s|x  A;~\-vbIo^Ab-?MMQvilp"x Vay|\ &2){R·zdj>u9dG=m%ى?jHB#PXX#Pyіik<_</-2\ί~Z6\8Yp@063wf׮ٹQޡ0tl{~ ŪdU2Ae^wr:g^K7QVvx8Oܸ^~%b8Sz9zjO#FO=)詧<xӛL )x@?EOnnFVS<[N?ݾS[CXqN@o}5֟{>Ay+|e52]M1gbڱ\//Aq-Ҳ΁dVQ1vA2e_KS%b'mTa:۠k5$fo $Ɖfcm 8FWLɪNsIi\FRS3Dz$eD(4JRȥbRcQfhw9 QpH\+Z]@J0j@{6ѠTIM ?8$(JTG$SQP\rPRTYTsm1EJzT#(RmM)6n \tN)JV՚D[C2DH%6 w}23@ j"爐EWE6" Yf# m(Q{le UHZ0DtY$X@4 A= E5Q(Q].@v%\AT"8Eu=$*2]*, s9+f &mWCJv`bɤoTiDc @_P@%P Rc I@\IAѠp`42Z QTmMB 5U[t as9s9kQr1 bʠm*!\\}g+FPj/@ѳ ?`+S骐B ~p,Q lP|N4C 1K(S3L!t;# 7[Yڕ`t(&B)+Y]so6.at:D))H ]1oJN e([Vʘ,8AĸP0)jR\vbt-HCiY"P& =dV\?r'8*;T9$lT#4 [I1Rڂ9`pncQQrHc5hBA]FH[MƀCG&TUCе l+ P$ؗQ gIvs1.%3ƙF56lu6#% K* ȴ ydN c5q@:5m̠5&#F_j`@_ yj4aM6$HqNpg% j&\}YRl  G_T"K(NMHVC@Z"@fӦirqɦԦʥ86"iB7/\dH4oVmbm^F6^Gנз&ʼnԘ21< #e;vL|ľ^,ra{C-wk2E HPCv%@ .ua,AisU5HY;R@ۮDYD`k.֜=CL'mvPq DJ tSh[VMgQ{CE8];oكJ6h/њf-[0rJ`cbb6-/j~l1(7k2 0N@ԍ*)ip t*!m(7f ;D/0?m#ɿ҇݋H:_ՖM]$!r  1_!HYzD"@`(9eY t{ 8,:`X/Ҙ$ 0Hрډ 1[? qfN V>$lJp H<1~aYˍOux98ђ)}.8۝ZT`t \`C`:-A'Xrs`}dZ;Kca9HO/() z%3Ή8VcJZ1 b& B[d$ec(hR)`cZOl(m eBoڠ؄:jJT J-@ZZz`hQ`[)('aImx XZ'46溺圓QVnL-uf6R4n/q{3}?o/nMxťg6J9x-ߵBَYx)߷(CQY{[T8".r*ŕsKͽk^!͊g7opnfzwh˝5 (}zb!;g}Qӏ/eY=4_~]xpŏwKOȺqrz)nFo_53oۻ\>O[]{ r׽|lw=,/_&/M.wynIWI:"2SU4ୡ rPT04܀&QrC?O Sp9),$m6{ n>ݱ7X;B>ݱJ,*y1> /WS V╙ܘ_&g7ja-$o6fP3NZnV8Ml7n cVSXaBvv-q|֎``y=>#,aCs3e nex|t`v}׊qu4FL?3] :IZ¾rorڡq 6E 1E ȟ Ͱ%CYEEf gOȬ[ډS@3tWj]yL+)u.lٸ&H2J1X.~=Da*ǔSj%A 1Y[MyTߍ/Ds4;Tݘ_N?l\iviD>d8Ip"Dr"JbtzD8hѨ$O6S$pB9eiI9$HdEO⯛|Q0p֠dI+UHM KӆS&ipɬmw䎵Sc!RS3Z%)ľ~rG~0e B'SGR2N0sI3mDbR,#Ye3ч݄YĪ:$3XBc6N4HT\t _ST +Z?_ WF@ۥ:[I3+*:i'(Vۄs*T_'ƒmLH~6A8,mI2X"t_[6|ǣk e(rTNθwuujD\ւ^P..ڣƧK}O?vG|ju^;ڧfDHRXFGNP[ї2hљ6Q3†]T20;y.B:Ąj]o1Кj5 Q4kx1)UÊql>zi4ٶlڣ@ 3N<.ݲGWGO^ o2T5 R;j;k,h#Lkp{י Vc.z6Ty^;*ml[bgq $ f0VUifN}Þ{xcђq-NDc?z{iYvN; %i)1~d.|9P_@+y,]0`GZ0Na.L%ldw^eW6,HkIkgo Wy&`$+2~Ғ4 a5Em XE X B[,0xXXޑ 3~ccX:"k _UE d8Q"4gn>!?ꠥ(>#!$_]i:k^]aQᕈNtft(db3n!,rh1[`0nf.~{: ED;Bۗv5P4T㊹W:v ??"E WGj/8rHU.S:z_]2XAQu*Ֆ0tu;4\@%1 f }V֖Jq`F w&G*-.`"۾+;diԀW!qdtY,,:4W2-J|)gNvj T ՞crsAT{̇~=aaî))4X#ʂ3Oķ>}e.o@Ǿ~R!9']Y.rk uOv;#oj BS0jz@AS"W]i mjcW0|7/&x.'. N|!лo@CK h=O F@Oi*W"(XLZYe¢ȧ}mYcm2 ,\SiK<|ұW'I״t+M*`c{I<)MVA$Z!0ֵɚ;d8ő;>'/w7C;S -}vu7\2i8˧="6ʧOT/$:XW;yNK3F)NWWS0$ qM hF#о?m$+LDv߇w d< MG,)"k1&),jJy6cQfUuuuuwկ֨UpP.:İ-Ka1 Cj?QHuD+>_5ѢH#"j]/NʹU)?2>RW߽ HyocdxKHm@6lqa${|RC:E뱗(B?|[|ƣU'E &ծbt7`vJ.K7:3JbVc4!c"̧P8)sk#ƣ=h{uՄ?t{5wW^2s<|7^Qj6sft nAh~%7=ٻӃhs DRsId 'ό *&qD)oylvͳ KFَ;6ϪVdimSL"zaQ~s>is"| n()3j}Ҙ KMcq26 YavnG>=WnN56=K|ns0JE]m$auڅdde-H8㬘QPɅʵXa,KL'" ;h6kvvfe4~[ӫ!P-N"TQBJ{EEC *EUhAd&,FRUG+0pe8HPΩ]K-'JjZ˗/O\TVf5ކ3 M5H\S3l.R凿[|`ϲ]bHva9Q$,!}kJ՜z qWhtKw%ŗo ^(u ;MOubD&p|`hjms $,~%n'[V`g-r/{wSz8;=U34nS3iCa*D Ӂ<В0N"BHbs\T+z#j=jH`N㻌Tk<^BKUI^2ÉAAxf:=wufNB/O|{78yIBoC $hb0v];[%= it(h cfBKăԗHZ+6X~3Qd癥][_]_dXY˫:y L獲yG3>W>_ُʩ`t0 *1μ^-&Yz޵IGY?=];ϧF Bai7 ĸ4*yWqE 罜AGWwJu/|Xjqkd歯~|s 0}[YE[{k?}yobzr0-yetVJYͨNLZa1{}Rދ/~̗sayoM 6:o6kr4^dO@߲m_; ̬Cc$x,r4®o#R~y`+&p”> (Nm_6韵iD@r|hgLwC'u^94ha24vEfЋ҃kA3tkӯB `c(M,>=kvki-~e WhO a+V8"K"$?-O B?-O ?v! Mw ciAz2<|SVP=O;iZiz%Y[p#-+|*}% |bFܬ-7l B>2ӟCق#yp8 z>{O'Vv/hK)~}c]`FV?.YbL,a_$7&H@EAPB]똥m?V1PG 謕C-+2;BKydj]hE>-*]^WǶłsWRdFó:!Tg–P r{ՄWW\­ݗH5@ a\hߍ p8 `$sq#\! -c(p(%lkx"̢#ʺ9q65HR9F^j .iǠW>f!5Uv8|5ʁ 62XǛm'k `==V)C^ b"{@Ge|دisk2{ok-g:G p[S8, ^eG3j΂Ƙ4%`׿^b4?-_d6 z9owV-E<0z76bX/#D&XD#, S (}SȏU5xAaZ5(G~B4A omN YAN(/,nh7"TZ!+B;<āXaD*p F 64AhA 9j#*9.YZ@T8DWpxS+y9;6`I0NW~tNཽ6\ L2{j;a3S"U]Boƾsة;`1BǴ}l5k ;3ko>RrX\5$ ABsw'Iƾj$~Q2Q%R*DЎTV1?bҎ鎷VeU`HGr@VCinPVҴ=轿cj}k5&nVbaqRYU_`k ( ffVwkb2( 0TA^w)A|C,OL\&-J%NB9ՑV'S>-4u! CI {xX"FHDȉ>^N"KtS*A (VD (`q 1&rnl̈ awwqvFkIi_}vQץj7jaտA:0ɞ\EwŌ.EW?q$-8b3#6խWR5T5Gx5Տ]=9}Tg(cY痉Ueb޺QLƞ.(e,+ֱ\m(pi䡼U1195Uv*+՚YaAΜ>N;5ļG-r-o'TI[wBG(f @Nz05C~wv*=DBw?v)pw^Uvڹ`Uj|-I o/ƕ[yGq3oZg!d`t'izd冼y~ &Bb}8_S+Bsr(]y[EyZVO:$R@p% n+̿ۈx""}dO@ MUƜPx@?R3aMXLP2ڊ6 o=.U8:(VKkckFEd1խb$^$Y•k e)8?u Q#%4LE )?aH0:a`jb| %;lHK@صL}cp5F:hYBh^b r CT9^ɵY #וڕݿK)(Tsj2PӇ$"+ox?ZwL$e w=NΒ #9? <7,˛2靭F{ܸ/vi~qRn2K@%L*."b4LRL@}a6QzbI><:bZN{0b5w51^f.&2rsL/rG)^g_C9/ Crzq^U7oWLΨ|eP:mngt9@j&SԋSTv~9.-ZΨ|eP:mng\[LK)wCk7VnM0+8My0kj`+t;B!+db!\d!20@qq:^*\ױJ5P!(I.qRSQTsA!kla=uE5$WUoufô m*ڿ&SԋSz-Z]Q' PǷ:ށŔִ@j&SԋSBtnH+*mv;* ݼZU5[Xmk@Ġᬮ.oٵdw5x.PӃvcWy) 7es4ֈ$y MqkJ9I$#!6o@ۧ*w]^oXI&O$.!pvf+Vj/Γx=q?5++橻s4G36 B9 23K6}z𗫧,upZ͓Qi=TALOfll+nwe vuICr-x!pxo$u&j)=w$5hsJ[@\smgD(NZ*P)ONno^tG3em\"Jɑ]'f)TJED0R8 *dTҔI%TڕYƩ"ц$&K+6Az3԰5a8e,&Ôo=5?qijn{>m׺vۤt8<,.@ZٵO6gI)V>0v@0yh 7^޼ު&5y`晆R i$Q$RsA:(݌A.,҄򢐜&-%jF!/Hjy{ )7IqIP *|ܾU3V2rN!aiҿ'-ҤŚ5`meۈdm'e40%{XC曗y3Gs;Y8–j O3Nznd^vvv6N swlʼnV܇wzmJ,G<3$F9.X¬y|SEG\ʭRjoʭ@m27'j˒Ǫu7$k Vdt>+yhFšS(,it(oO\HLEdN eW/z4gQMvX 5S?v4΢Nc@q7 P<'[RS@&.?n^]tl@H[+C"sz2J?FEth8l\4i 6ڣve{4 Σ't#Qn ӉuL*TH)wZEI>:Ф8Xo-!EZITGO|&9R!؄ !3 *-,u0ACX@u.Au,i+V0Ң O%0 *Aq,_aiELM*C4& Vb(I$tQJ))LR4O0?޻7|zMJPIHxkV*j=\#WFzEt6y,*~˪n^:ݫ T5m杙\r1^}NQ3-嬰S/{ዀ~_U${ *)hEUSdNa ])^͎o?~CGcۗP=GYৣP"Zt Ú Cўc֖G@IM#P{ Ej:X;j=JZF2j <p{EM`e>zX^+](A"[R(xbP^ȘC~Ma?[^?NN04Ƀ̠~ iCoA64Ϭqf~q(-F{čF-ox)ZtP#ZOosciZcXPJ àk%8p*8?D9 ûoGv:^#_[=gұ?z4" [dvi[?6u q\GĬ~◙+K>8>8>8>xA2N"ɘH SBB@ R%H$Zƾ7Mdh`܂?no;r1Kp&߆n8CV#-,1TC;`E^kX^o3Wj{Gxf<ĉF Y72h0Jj&V)1;u f{J oj W]aPvif> ܸL=N>?d3P%!vOV8xz_ g_w=EAz-z{|<\yjEG [J\YMh-w/wW0pWHͳ-qTd!_cN0Ȑ@|_ &eʫ0PK9AF ׯ' <_AU ap|`N匪)z8zw棚݌͢TJlsh{kT"bJ{mDDUH42Sa /<>lefMS{ (Q4NNTV=-+\(ź!dyJmBȮ$:B$2:Qs%Z0w?gv^dqTv7v06!Ϋ$M~5'//f}a֗unӕm\(zmζ3/_ Om\"zYQfoڂaa(דMCB .78XPqj¹h}oH@/+1 B%~ڃiP~0ak!`[?[NM8Q8A$X^(rz.^ï\r\lj%u%kkkkʬ-@IH XQHLL#CN1a.jT zXoC^a9Lk밗z_ԤfEV%WXTp?< aa\Rb SfI"HiX cz CE9 0hLN] <~j@N0t6 'W0;X1'\pk,.NyHa @7+k9s8_d8 EY%Ov*,C2vymckSel[O޶ЪBY;Wefl]}g2>e:R<5ʞ\?=TALO^ek%ֿ֟<-![R;^fa9PV<%MEg4QVT\HNfl:r^BdMoyvw;yQ v]]^laж}B5x l3=nA1CNHIpDae1|@+둫JyeSKj}E9v7Ibchcj!똠&jBJc*I*.Ud2fK"'|h"P4985M4akc"$VhBa,ZJRBS0bF'K?*h pJ-10DR&1zфs$dre,+Vڇ13٢b4yJ%̲",Oʮ&02+$R $)*,5%VI s 861 ьQɤ1HRh͌~RE٧)͇5&c!-0 &r[VPvu~5XbMR)Y b X4c) T4F9) AAY.@ B X\D$&إfs˝Lj*ԌrqĘcSR[SvfZ(b( 6 iA0H4_Ckk9Fԇ+zSxN'[DcgDZЏO±an sHZ]9,8fс6cU$Z:!gAdE7 w:+xAhM6&U/zFrKh)LW EB Y9r8z"6w9ͧJbܽt?,_{rjnޛM0;辌:J;!oQpHRB:ռ?/IHrbFdwFGCC'&hĭHEPUtEDwE`ĀfQ[*2}$RF/l**;|y:(n9}$8}8FgE΀]Q$5ɨj%[~j,YH¨k}?0⫿7|7?ы}-h5Auw5DWRiQ!s?{OƑ_!% (Y-yx.DֺBQ>$exX=#ɁYp>HvɊo5W)W7пQ$_TsSoornJ>We7F%sS-k C!{kD4D4+y_ŏɉ%6C\9̗j/|^O'>fgGiQXc v{܏m"WӹTϊ]e Kߨ!]kwc\v-إB& 88E*!la +J‹@Qn{b:9A5!RAN H 0󺂣|zQxlN{>+A!-)܄ڒUP[x3l>׈K8xf {C,5\"޲Ơܘ"do5D@zi^o歫cSR:y[bvwzb\]/QJF@-_K9r;X-k Z4 cigi1h.hjGj|CLn8cxwnF J1o)ԵU^v=/.VR ,0/Uȋes"C S:72Kv=!4lVaD*U{:C$KoEn%SMRV4K 5!Fo= J`dmuy&`f$ZZ@ju-8DW;* *2(4 2$R'6sB:8[(aU!%ˌe:9j#DJ;逃\4D`%uŰ'M< v-ȁ^ǨJ@ٴ!oQћYg0-A 7U$"PЬwNx4V7tnؿ:i'Nji#h2 *0!&E-HfE*DD"PM 8741֦h ׯ$,w~B,YAр[} 74dDd(<'h uK;fP"CkX$=Q#ԭ3S!; 2pe%ŐPc݄/8\?Zɵ1t4* | Ds5h9˹c.dт ScVO7kٵ ѕFf(`[#h"AĘ'*b(5 3 UKD/BJQ l2:Mxϯ|Nkb(\/#Oqys:) p2~{{汃s WP'hF WK(Lfo+!G?ϜګٯnץX9>7ioF*}󟅂C)%bNqz0#҆wt=UYÄMH {jÿ[jpI1h REV:?X0(F v# ⤦ؖޙC?_k5Cݸ\^ǁo(`T.B u K F%)L'!S`< Fƀ*iӻ0AtEj Y7%E43R3EGJh -$HL*pE6bR4ٵֺSݱ6@2PIڣIhGzk`Jmj;ٯQfGkLI2!x1ZB$Xh 6;6*qMUViOJ y1}8M$ PQ $GO[)o:o]+"|x6q,3[ 6:Ea#ЋܻX{<}7-wuN(FkIk,pTyT&RxI062!rh_^lVW~^w6jm_^O=^l%vfOfh4?'75t-㗜(}l|Y靚|,ݞԈawoldN{aI$I: I%ٍ$=$۸&mFDtQvbYLd9t4\ͦ_Q*vT^N$J T,dDiWR ̷DWCYD  g6%UهJժM)XR؈.YcUXъ"l*hEUҁVTa%ViSJ^yRҎ٧)Ic6@Еظ*UrK9C3@ڻ6rj S.r94vǗw4 z}o?`i:(FlA{"_(]]^D-ltEm<;شp o}:"Eڲc `rQ\`Sjh hCB*TGT/CliB -܀Fhta/p/4eK ibC+pm8H904Uo2tٰیghB<#{}úpd*uJ)4^ڱցJieu mT̽ڨ}@bgn?lprWE>`I$ioH5k\xh5+a`Q01ӳ&CrOtRCJ=]*pfzAET CoD}׌/ל=gt?V+ xQ ߣ_=3LeX4x[%wu<"(>ˋf NOp7ߎɝƭ7INNrbwjbw?=FqkJ b8% "R2_y;hj:EaH'NYBGp$@?mv*|5f5GهIh,E t̓rs /N\_. e(1YG{ig(k0%%q"hf JIJB͔bw [-E7Kٶ^(=Kx[=jwo}o6-a^ ?`XCP$HLhRm26EnS@uaROa`^\4AR ve8A 57A!p}ZC>iEKIVdj/1eQip2ĺ`~U^ymA! SiGJ+,B㕁i%OcmͶuil&QQN뚳G c&5!踢mP>1F`A+* jvE;TsQ 9:.Q `B,ʐTL’ [͖1L)mE4ԣ ƨ Pe  K rҕǿ2"g>Xi{ 1FPDL%ZC HsYȇR%|LAa |'E0HCQ⪌ \:!0nsA$pId^%w43K ; *a ;ʳTD(g^%T\ bRwq4W02/0yI Lr^`R TcHQ2@Dy',"Q"UL"<׀Lgx`p JcȦ 5!.|~d/'ٚ?We68Lc:$wv;~@t]x3='[ۏy=Г-:"%ڑ(e0NO):$"x(hGf%QPب16DoAp1nѠc̄G\8ǝ ip::fO WH / 6g"ъKC]"y*'ckX1FQL [1<3A gLfrukBPIv)v :cԠ'8:cBR -vƐ( Ԕ^3NҬbq$9hᕑ}gʚ8_AecvP݇"di6&3VUI@n @T K١BwUYUyTbK2+(z, Քبi#ҌB3?m왪@ ʗ`G8vTB14]ꀭN tGhjb-P"P)>0 ᎈh0(DpkQ{DŽfGq&4S5!n3M5Hd,Sq!,sZSUϑcmPmG(*x5`U)7me$ךrJjD GLu2(LBmm<d`fH4F̓6*'*v[mKLAuTT^@?y+64)O2Nc*Ip:J᧚aEyI)E nʋ>bLZRuDF ՁILAT A %Jpn# " +roP29HaSt?TKNYR zKWyݝ!O{*`nD@UXT.J`Z#+i:&ȘN>f7crͤz?F.QkQdCg=r[wVE]H@ꗲZ\9D7">E㕺tT3f7K|tVo.SJ#o<u!]xy,E&b8h>FI=X %}(0>pQ>9Oڧř{oL`HlKAIK\i= CKw3P)gD0_ݯ.a{jurLߦ7}bgf>YI{Dkp2m._ϗ}@^ PW5]w&l\^ز sy6VFx] 7`U2QFI~%s1hzBt_=F΅uw`[?*)OhUim<_,fW_Gf/!6&$E5j/IƼ{x8i yE-&iڈq)q:ԭ#(.Eʶ`.`62;-Sb~f7UI3n2YӪC1'uQj=v ތfŠYY ΡϊԼٕN+ը+Q^fumAr IT-fߎXBBi\H| FwpϥQ<{sׇ9(`q~4p9*/ӓP)X#=Z)]LS)eJI hmX5a(x/ՐѷL(vT?gRGA_ha *W B_o<{}ƙs[JB-$=go/\l`ՑB^b"G۫HFq2p NuUC5,% ;S)>;IBr55FJnE;:%-`l=p#ᡘsMy4w ()SB11'xl5͖Jq1޸SO[tMuE=z+1(~ˊf%S'߭?}䷋P~:R Ѣ ٟӷ8[,W>ܗ./S\Oz[,WÚss[# On X&H)F'חM5MQ \ɋNgaLtڦVǙ EK$6{cbJ-S)dorrKNp`s3x]z{kT;gܑ; ^?Hd,5L~'rN Y@]%*Fk6c{k N2}=32󮺤iuD7*!1tLp><H bwuV6x_d9fgICH̬Ή*AXD#%tAG9#-+P2ˋ4xqSqy̒fvr{3#R^R?*XOF+iV؆3QIVof\-@N5-h+gޓbC@W8i* ?Q+/z X렩rW@}eDDJ=^ DH=%ZjpsSLx$< ǀ=A,^",2Dht b(ш< (6Vƀ冮ˍMfZ'n4( #11:P! b$TsèVTʐtczUX9ki,XHӘ6"3DDՙa7N/Q?]DfnMmzҐٻ*uryOσV?5;1O[uRz0IE׻۬]~VfM_)+jJ( X%J+Jq0Al.b6FpCD]xJ5؃R麷 +x'0y=rɉ_bve5Su^ˊA/;IT{W6J?N _ȇE],s7A TIǝi;,_vْ)Jr[̤)dʇӊ:,M"D$DW՟يӺ D4 JMɃ[:ˮRc>֢0fϖSKV9-3%rgJii)%Siao vQ:B"ÛF zW\pf:9b!+Ei0 #4bfAE.9L+kxτIQ:ϸSh]rg0ۍٷĖJvXsd9kJ~qof؆s: o1!o&9 ;i>ʄ-k(F[ؖ%uMLbFm]m[Owu7D#shAk.F7B )b R0BK<$E#(9]/;.T.L~(E!өsxS8uOPE6SQD++s8NPJ(r%%SJ?9a(?ffTk!ȊrRKmPU9d/˘:}q]|`+,9[7vJ0[I1gs3EQ+YE{٪vV35z[ZIQ HW,]^7n更yĉ*+P 2˸,pIYE{ZqY] `RdWF,hlܣ0=Dpq4 A38䗚`E!LtW k}~߯5rZGM,վBF(V!Rt1u - 21ŀ/0xk\+j6^h0,nN: $"^ c(jIl 1jPh lX("Ruti"URW>^3 U>BQ"O`pF>S4 ~CGOMdu}mr$XPwȱ0}b).ڇ)Ev@#›h%O[|AB*Q60KvOT?0״WW%vq Gf%BV)K3Oy ((axo"gԘRH)A h߬/+RJu@q7ϱQeƂٲ†o(s6o(7ygENH`{2-KE--9H츃 )rpm8ِS<9/a/S4l?ByR$(UqBEr HsPDq^)rs#~ﳛM̬ ~|5G;e@o`0հEy^:K3KC&BSjN$gwil4BҕT`E1"sp,aSb4 q\;e6N{tKenn@4ĀRdS;n +邍0`{1͍% ^H[e[|# SE6`˱-H.;caP`D|܇Ӣ5ܩ,^  k0&2[p^L^Xd))sDRV#0GgK~7X^$in?,W~u} K$`{ 64/QH m%0Ǎ`I,<3IcA&CgRo|+XS1I % 0Ci`kz_/=a] d7q4뀇Zn>,֦8c>dDm&!;˧W:@r̨4ֶi*/p2-|1 }Y= 0lp/Cbun| t:}s!mÆ .O`hڧf l=߂-PX]Wf>,ʳEZa_[ۃ̾y۾VCt1,O|72.v[mb;3s72V""``SY07 rPAC.X5Mpny D.'o+ Ի_UoqUC%[փ>V)a Üꇶu0[]S&~x{ڂ[Ѷ{y=CAUr:ָA=Ltadz`u[w-@?ѡD[ĕm9_PJ6ꜳ{aN `ꓺ?sѷ̜"JJ|9{dHM(˿vg8C[dp߷ӸOBڽ#J O:g.8EӆJtNֺv2}ġn}w{8D沍@J:98Dvpq) R o,=~c=, K\Q03޼I͜r~j|Q "hwoܻy~w'6AhWp?ˤ>?ͯlaMʽ`3^TS__X`pDEV(>郏GAH6b!$(=%e;:a^`^T%1zgUD tm&b(ݭ:}{atpsh;~qCt཭.21tTe}i*Zl"٭'@ ~yv"tHj!l>)QEb鍓.,:BØ=(>4ς*ޮq<?s3 13-Eͳz2z3\R 't@21 ʨ/\O{!(WpXvi׿s&׀+l hqK; pK޴ lKػւmA kK`F@Mi@/)%U` fK0q)2, BϵbU^r^>~)Ń _\P/~o&>BM>юQ_9]j+ͼ ^y͙{~U~N±.'++TuE1g &utaP̳},+kN$ꭰK`-3;>"guKc]<|8w^B06{?:qbq%EFhdN@ٱIP`xY Ȥ`h vU @#-86 ɑQu ʞȨxXNZkᱵ'U= auS':F)PR`kr4Ajd5b+ZPl^ѦAjԩnHLtAF*7oofwƽzF~9YV9(e_J(F4m~8~Ԝ&o1-? DTǠ*cti(,G(<9ckPX" Z0$0FtWIj3o7&E'Ǡ(c`9)}~p06eAc|}Oi`#T1r1ԫ&}JRnǘZ$gtct"ڔJ8wڃKʋCE|/O,b_s~?xn)mCHPAEð tf? =G8TĎ7gm+?WݺmoO&w!c?yf?<m$G-E۰Ib}V$)6_p-Jxxt? clӥ Kو)jp|dx`z$.|Ԕr$O|g]tĢkeOC4 $,5tB5K.5%[#1芡Vf~-5r+|yTk<ϖUt}O|YTEA+wwq{kZwn ]Y73{s{cs_2T"+xf},%7nol#m26pD苂o ܲ{vIdc=67-կ p)NX7=ZX NlM[pShbj=кٻ涍lWX|*=dlMjMիŘ";~O f!eq|gtELIv6uAѩ{]OEc֭6U[EL1~ 7¥=hbPEtcvi)2n jݚ'.5dJ3Naݴ*79fޭ&5Sh{ϻ%rޭ y"ZKJzjrݺIn511ּNΉ쀾[]B+n HZ2%2lZ7!O"2Aѩ{]SYڄVnMHZ2֘^̟MYH֘s $;/뫆n'6c.̖οoEn:0w>܁9y翗Gg馠O|4t.fF; [W 9Ƥ(i륓QW d,*C`!8(ڀԎ5X%"cKd1s@_:I Nv'uo$3.Wa]kȴTmBvR{8fmBׁOm;DV:mBہШ&L=V؁(mBہФ&PDu (n;&5n}~ԕ$ v <$gwY$E5H|ɴlb7N; ywKI0!<[%_^M&3dKmv#=tؗ*=W ~S3'f 4Eb?`a.F@`#{G$d e xiI<_--oI\. :ApmT +םн~y7Vő8%ZI:M2MW `Vf|,yarx0a kNCsSP+bP!jBZnݿjFX'Q}|;WRPjBY e&kB두QD/ ` B(8xhT!R_Rm%-Cɔ B2Ts0J4'7,]<|B>UffjM*)E/ed-[)㷡uw\pa2Z\yCQ0ZL;;Y}ㇳηE2䄥Bg)y# !,LlׇҾWBuDŽ _~Gy4 ;5"J-(i*m`,"ԣ؅!#? RBȜ/< s<*<:n,$v&㰁vE*jAEYVEW*o]9,kH7NMA~/fKx&8|j1F" >98?^$ɶ¤j"-҅nb/ =XӼyĒ~gyo1(jfeMOo ~x$p0l~Y_0XB)] ,b]4XdaoX)a^rAABad3*?[F{5CǹjPQRlVfM4RG҃fl8.ʬrqɼbi!F0"PP@!B kx)⥳! 6CS0}=ZaipK=Mjxmu^x՗NJW|XfNHY&NJǂgn"kϯ_7aa=kq\vQt;nTRy8}o;j>.MS\-EBV_\z. bK{e/,*.XG;!3#H̘Ci ~HATr?a\ vD<,+y\TQ/l^I$,uBDQS\FCjLf={뗊s+raK`l?EpX}J YOk,A3WT^2ol;{ o>5.{ r遥g^aHb=7b:E>]&'^Y!?SRU0*=fݴsET9/Jb:gىtQ{wb낼哇m.7l>C0dc͝zy 22k#l?a\Ied($+c/FIL RxUy;::0byw_OGIWa#5^Rʰ^`#)Q}Lo-!ƍ>>ַϑtN&%gMRʕat&7G,r/y';f>qu%8w*|w~wk]5-9S{A_yL *|&=t>DgU-Ap+`U/PCI؟d{D D"C B")+ 2b-Ыk{ÝXO:I4AUK`,El0Zu -s3#9PRkg([S|@.J'.vfZ [=< l{Rk?s4R$9:WuJZ-$78YA7lTѓJ;PtRu vSĝ-7 X[G<|G5ngc7O l"*Y FFm$AEJǑ+oJ<7)8M\0(IN$of_%4ՕAmmP`Ӛ+"{!$Qe<?JawY)TpB BoSwnmN5#p9]0V?Ճ偫 [2%,a, Yoq+9U.E'iIp,[I0o-9C`xUWGVbΖ{7#5Ȧ-WsZBE6(w@[TK5dAAAAY"2K^!*!'*RQbŃz#FTy\FQe)$͌ybkr{Mu!**Dq$,iT:[: 2(@\Nqh4Ex+om( ZXCQ5z\1%i2IDd"1D2 !@ ө|]dE% >qq4 6tSPA%E$8:奧.Rps] Д<ay aH=vU( :ؤ#tUm۽J 򃴖"nSThy`íVo&~UsV*}xį}?7/څIK7ж‡ж8JL%SśWq+U7?Snn3vف+L&HWX+ŸLxGnH3tD{X'.,Veٞ Ͼ$pŠUL=AJ'F +. d^ʬ \/ĖCt0ptˍO0I Ad71tk ( xMV+X87flql_1< U f]ޞ]UW7Ub*3eB d0`D0 hYqgV8ǣ-E4ATd ?;?ۆ߬fxrt:{BYr]h&HGpgIvzU\Gq' msј-ucvRX7-Jem:v(kiXarg꾳m*W8)W N> "I)ByGuLd|*0rխtC*6|Y$>>KqIz VIruP4yfC0!$F[p`ER`N;@J|u$܀\Y(\)Ec s>2I(YNar.0pe2 -w ^\g2<۲1i\ZٖkA+`,.f%ƅM /g#1њ5nW1 5bZSSMGBUU0OE*5zxe'|!C8_' 3o :9gG~д%UFP)g;2z஧r[H^xA>=;]/t( 8o~7Oe_*cڀf`n\PkYy<E(^^# ʍTH9)9ڲǢo':}%m#=BRagK_On74WW Hh(}J~4l{Kʚ3\1M 6=lz%&Z%B@&k*IJ;:%M]7Ԥu9|[%ګ調#".| ߞ537eioO&IiJU# 2MbM`-n4A>ɡE֙Rsԫi3pceqTjy9VVqiHIRbFt2фOݴ&hR3U=H5P&M'Wad G׷ :wd1?OCܐ^?Ҽ UU0Ty4_>'ywu"^5i1s . 7CrgSt| ^Yp}u[R(;fR]1y @pL|R7@oO>v| ~`qcIn /u9c'@C!S=2ׇLJC~Cwѽ^I  ?έ9܈פR-zq+ݷ39~40}"=RB-{fv8MF$ W]yșhFE}GRc%md"/*N?r1bA>TnʷJnqLu1J-N[n%ѭ00nG6nGAjl~ź?~}-{| > 5~R pQpQC3sl(2zmn[#h](޸3"R`G`9p\?^p|3fL8?*=!Q 6zJٛ::Gj73ޏ6*x5* 8Tr@AegR~k>{J1mE3ޮ;4M.xsi::K#&tY6= #o='giQR1U2 w&cj<{92ds=J\MͯO* - վiۢ ?L=Q<0Ad[wk7m$+ `1ܬd;Z&]S℣7Dϯ AMSe{Z@-U 2~t1|P^RFg'oV=63ie|41|ŤHe@@517 mXlr6y8ʄ:d8l="*mՓD3yVQlzzg4~$L^Zj%0ߙFkNau1_:wʦ& zuq.3<Fie>MU~ d9W^`ً Vw'z`8V ;^ƴJio^񚥄8@J56EOMG} >6Qp}>Tn!8 "1nA)=RC!Zj{,3c+wo&c߃#$DQE-w3̈*h+4wyJՉU5FMifhQZ )..#(!,T}雋Nx;s'y9}+-(`s:z+*Ru2iLk8ԇepTyDjbͮuu2_PrF Y^ܿ~ϛM$\$륌DdTGMF[gFaESъ2/bQ' U݋yZ\)82!- <;ف<yӁlJA&.QF'dw,ŪECR@RN3֒dmvݨv|ouN \KnJJ!d'-9}VPhZ+NZRA|p`PqqmWK>S U;SX4sɒJ0;,Dar[ȒdJ5nHd=[mn83d l<=Q>xZx4OȮ抹Fʿ ǻ޲( Ɇ[aJ༻jz7Qlwp++ &H.''TqAVɺZ[1TdyT̘nDLdo`m3du؀DeL~)"]偡2*grbҒiPW$GL=g܆m!vAfz \n X0nl1"J;Mp/4GZ@$Ķ˙-X( \zű%䘑J-n7 pRO? fY/"/QSC 1dluHZѤG6s2?e$f''0LTmćQcVWdmg%5.]3XΖ92wxxgy\~ūYId1Z[}QWآp ֊мZo}}_noWnn=>Pg3go9Nf+*]ovV~UM>VmoQ+RTJo7l5XL^ո>$ЦRn}গvD9~ )/LCk[}{U),fuf⃻mwBƌP&Qǣ\aMW5T0}PUU8ZvْrBrSG j)Ekko4("(U#.Fi棊W# Vy.P/1$!_ۜ`;"ϐEm5zJ)!Xn=K@QB)F<|8viw~yb\m,(z-tY 㠁rM_'0$<1BߥIQo_] * e˾GQaMt@Ҋ?_%PK`QHA҇T\zF"1d LI!$ _syD(paK5~~>E!bk1d!`RDx$U$q疓ș!e(Jqš7dڷvs崢fEsVAHiPazCTpmea pYI|Me\HS3)yڠ֖ӤN'nZS)Og>^]{>9m\ga:Dlv7yoœĶ ߞ}r+6,klen(-h9*CnP=i 70襘er˯F#Eo <Jՠ>s=֡BZو@Џ~śUuqi(RTh]vXENxk")hν!hCQr>dBJ:MGPp0\oE X˷~H獶h10 ?:oޜL'K=O"KyjN8I73'1ĻL3?~4=qW73a$$}ʾI>I?ȗCw'CF47ڤT2OrT7BTVIVl{u<*u3orp@&w7hNNc([P]>| `c!8f<ɐ@SI^fg'P_[4hF?u( 1u>}")UHJ+$Q_&D"3t6>I|E R^9 gŎwҿ2eP`0IU3y K}QajY!6 9,WB7cR䬧>btD9sX;j`o mlOC :41ףǿ9%RbH#$"c,I=S>(Fӌ~u&LS$ 4[U2472d * (?槥U>(R[`Zl3C, }٘@a@_1қ%T,hq SJc^J*5 X 4eTZY%ޡF?[7pD O][O{7)'O_]b?VlLICgۗ!thIठÂPH )DÍ;1F\1TJ `!^am( AQp+*4Of8H'6H Gtzxg\dm{\mzxgݖ_*-«5: mx;d"$(RgΞ=һj>8u{: m.;#1)a  K.]hTaD9]BH;qV*^vI!Y)d AEς Ll8z t ZqI@z^akﳧNq!s(DekyN.8^c"ߓ㐁NPS\1% GƆ: JaY.e+R A#HHCp <8fN5c;fK4DhݼƄ` :"N*Lvw~p'y4_2V c(+oJ)z 51U"12K.,v Q+\ Ke,a~IZ5XM^IP*|}G(fT\3..R (Go{ k2뽨|RIigNh]KxG5yҨz(~nu]J@mH3#YD=($"#Rw|>c^<ݻsD }^r_Dٿ?8Pz]ÿiFoݲ/[iؑ˜zdqC$Rf6m_OCh6 vOab{: _93[ dsǜ87@ӬކL&1V*U$eu95L_~:|u1܉F]9`P sXIH"DAV,/AOoAm|YPyYS|)H=:LkWmz?nX;p4Y\m n-l1JA[ \pC,ֱ@ܙ/9UJ{Qˇ=?ic=R! yb1"0@1wCOTR/GW}Up4{rg#PiVoy](zc_~ g[n05|'AW:X`[~2\%/@;KE:(5IʒVZSjYewD7S 9ǹW~[01g1=sd!Dw>ф5U5W]ai8Va ^!Òr5)ΑU=%y.Z0@mG@֯bT|wnt{nfEW$~/5x8,zf!sD9 b gckW($ @,QI%diP:\t@A^Ud,$.>E_WCc$3iOr$g?zgzG}xwbE=/?^x?G*I۾N >{tRu?I91~b9c\Pmy p:um7ޯU1WnR pmd ƽ6l* KJ7UQVx"Q0q" !EdC 6Q0`1I%FbҠ_+h )0G"p b**Wύ!s:֛;7-(M }2$)e 3RՒ8 s }N1kG^N&")/M%@ܤ< G qdsl׈9s?X~$mu{5+# nubt2&-c-N`0DZѮR)%QE~s! ' W|P$(lPcBNo(4*9a@F3E~ΡaPI"  мԖ1Kћ$ yRi6z#Ac,⤥`U*aGJh/L Gp$⼃tgp%'t^pk/^֊3ZQ_Fm ))Q2s[/dI(!}e=u0184ɡ/ I6,L)hUHYhb'6ӓr ㅲZ\HKc!К=M"F3"]nJee1=PH%tZ|buu%ʮ6(`O/J,p U>B^`"GfS3&-Nµ]Қ((H$S\+{Xq 8Xd)XwUSV(\bq+X)XSzsW;)շLr#=ijXtCR=6-^ s#(#H> [ 93]QَaJ;}eYN{ `דGc'(XXJliJ9㑌ѽT;IHOH|tOamsH9[H @,)*KEf;V8Z ǼtEڮ&/Aq W&TQ(v ZBp*%;î˩͡#/̝B@͛g&jv\!`ϳ`']BCDxl=60\LorY=# |{4D KfW6szJS;/6?{&S{^ C=|.>ѥnQsnvŕ"Zs-E4'#DFS?| }T~)yH[- HhFk:%ϱ qYPLAхTT iIKl 8ਃ1n_<6I_*׾PAҤUv $Jbov N+Gڝ)Xw{PImv5xC tFɥiZ x4)49s^eô?z.[Sy H;1J;q&Imrך>1Cm7]A+n<8KhE-Nܞ`-&=« `ib{Q~qbB'vl+/ޫy۵?<تaœTq5~J.Z(D%tT5L7z/F*2Ul;7d,j`z rUmS!ȯ@?k1>^4mΙE * q5JfU%ґ^UEݝmX'@'8],ypXGؿv&odG4A5A!W^kʇGSYE_e$쿺CZ[:c%ƽ%ugK(OșyWRa:ю-56ZyE DADSj^6pjc2Gj."*4)Z`ZKaBc>`+ y՟UXd[A$bYo vco2lc-s :O1x 9 dch-v(!Nb@<3pjT ď `KNkjckjn$#VMb [K9p*Mp5% LHn͆K1y 1/aa%Xa1UP)@#XuL3XK)1 ;*BDpQL6bRxoRncOK!u~Q\Q536lrӯzeHM jxaxmr=˶aYrˮuR l_0#3^E`cP _.ChCj}XF/} =NC[![ͻ~ ?K?nLWd!):F4:e$: S$`1grJ-;/-k#13!]*k(@qh9ԠtV -b{WTII+ = `YJ0h7چ&P:$Ac뼣ϥwRfyb[JLre}DʉQ`aJRkDTYWu_;KĆQS:QRSR.Z(`xmg.o\c;|Z0-{UsshI8/T2Qޡ@TPiEY5ď~~F\?GC.+mkGEEy-9GG [xȯ(T7󷫓e>oxwp4p9tzy3g7xT&jn{S`Vl<8E]<Kcѡ,9 o(Ҩ`QaܣK%8lbA#fLL)HfUբfc@[5%~IFBȈ9fP 5s%c F+Č=^Lg*[6D2 7Bb%L߂A#&TqzɓJL*J =*c\iYxQ; ~О豊Pekb~zp\5!e,#1cnNnp7uw7MG~78덿<2MsC4d}|NyG\6.PLѰ"w0{q+ǜw.zXMdѥ#dZ$oٟrCZuNFҞK z9.*:0OoӌѠVN, YJP|kr"6^l 'm@Cy`* fܕ jΒ,KW2e&$զ!f\\)QFh5vi8&bzYXnTjOI5 ]5h3I%%Ly&V-D@ۦ=g+&->Z'.UAXFe'RO^D ⎀qkuc^gJ b*C([6[_SJTvUlŬ9&Ox~m<3<;'"Vׄrf1R6EDV-|)E#lDG0bVU>Z[P%;ERaEArlBƗ "TYSPr ̽*.#STɤKb1EhI,L,z f a1Z2lCQۚ#;6ˆP1# 1s>{ J! sea,N$s&-yA@BTy f XC8'6|PgECIGqsU@bR]-kf0 YV1SkoޫV;AX N+gē;vZ[QN2ri5WV1TF6"X/Тg{tWkeҞEnx~K\p2;}{CkMD[McZ6] ,9o8|2\Vi2/!_JP|] x0d\T)jpCSnoBicؼ(4m}uL*UOU#LD(qC}-:VaaROE=l' 3N9Dki.<| 1`b& R4 HY1ggq $ׄ[vA<yb֫$[.R4_?m ;Sc}CS.#ձ-5_Vo#c)s .xxC'u͏'aQԚ +Fg09"kPu9; @eUjzBV2=Xav:4ͯ0R6e.ar6kGH(\.H%LHy%(Ol`[tg-sr/FC,+IXKsV fr!v9 dI/:Qu$03\Bm ~l)QB<]$ξ|:s`ͳZ2HL3f]@K;ZאX b E'%;J$*:\.O+}M/:ԉ8O,K"{@<- (~^N{al}W=i#4w-'K@v(R/xD\2wj5*|9Aj.5.j&`5E.(1R-֘UYWj\̚B#+'-+mDV+lS `@)W`} 5˄uPm&]"7Z>AAmm+__S/z1q1gDZ;}˅o\B-(j+jDIu.o0z\OO^d־_?ûrP'W曘"v/+AYsr=[n;`gvNf":s"3PY﷒b=i/*YG,.!$ PjN&2OzAHux'쾙!"Wr9\tIYT F#S1LE܉v.: )TEM yQqx7w[NY͗@^zN>_tLg3*,RyttM2]CWzqM%c~A\{}='kp&`Z¨OnWq v\7(oN}}wk6RG7 ow_8u0u:;ejx+|v.NLl%cE?ʘ !&O:N+)yŃi6Hα--Ӯ^1 u;zXKN}+a2̯C $73/ j%19G:8b.ɖTEm-cUe{=W]S+ +W']1r \cn9,/PY76b~CʩtƜ-dV-B.(sW6DDX'Чω(&PNC:XSz7Ra OZ$\H| g Ղ {׋u BG+ Z!2Vؘ*g {2od(F7=ՉC1YQ4I2ao}'"iVv].?ޅ^hݎh.qIG﷝eznr{ws!Xiw+癨Kݎo6=2z?Dj/&M+/wyJ9m \ _Q>|N.O(JBVO_mmm|l6 ]LIGbbum'hc Dk5ETVÜLGh狋y^"ۼ3Ct4aY)ev%")i!m6hv\2ih`D~97LrHe@ AidW۪C(l je}P~͍XV5'BI Fo75lRJVS|qC! Kio4F~ ʻ@MFq$b)do-r@ )P"0$]T51"CBﮛ.k.-`W^P0_Wqˢ /SFL^C`&tKYA,L؋Am[^OMh-`D|MT7<ؘ }TKvjw>%;' E|z^h0m떦Pqݚ'UǏTomNv5prWs3JL<=/,uNVJqrG `zcI-8Dtv'Vz |Cu쾌[x&(؉Vܦ |iCC=mzXh `6(61RC3k&"h<;BHVV1w=# 1_P~.9':1j "*m}kSY5\?߉׍qeBVmdKvLOgBF _B˩uiCrv:e]S1/ϛ[yK_tKDs؉fl觃PiQGL]FyNFt>z[mq<˿b-+tJȞ"ͽ힉kkn9/CJ%9S\\$Ee;8 זwKeS f\f2 YY(uw} OhaycX U`*``餒9U_0 $; *)/X+8-Q'tk#HFݚZYоkP[Fsg#mݓ(#\U/5$2TK{7uEYos.P+@4zoϳȿi TyoL=ĬdRZ,|m|l(VL]qwxHj*%,N[  jО6,Y8PgO s_' &L .7t||)"<&boUn0ʧ EQEb~3 Aq8l ]mk^h`uzOr9]nv{_Yb"ViWo>=T Lu󦕄K=߲_96Cmamɋǯ 08^ .2w!+NdiϝYF,t_}-=UnUrHBsM)N7n q -1:FvxGT[2ڭ E4EOn4[*bD'u&^`j8-Xօ|"^ҒXQв6#!'0|Q;X1!l|JSjY15`3[~5<5FN1cQ9{|稺u$[D16LQ%-'YĀu3m/: +tӃ5A<eD^M]7b~Y=l(HZfs_Iq1Dj, jJ f<Z*-3d:a֚0Un0ЯXȞN/U` &=GT1֥XZ]S[Q49ڂf)u;2c0d%dd#YS3ѱg<@ļ΁1PoF=/ĦNhH`K X:cZ7꛱WHC+4! v>m.h_[P<9QumA.$-(ۂ"N{ND1E"-( '+fkp"il[a(hu}5 &mAbfラ 0?J)b (=Y'ί)ӕ wdrˊj;+S' Qbx"f'LRVwl#Cf1߮={ӈ`ұY$kO/#QlY.ÅjIeF F=<^٥ʀV,γMYdVxL>5/̖]߬7gg9&r?BI/]>]]3Nhe棇O _&0˅7`eBv~yd^iҳeﺩے;˾m^1`^=vY) rg6g\I{w;}f?t6(!IBZ{4VlMŗNJ ?]+=1"e[^>* rqj}KrJu9B q1TH5H":{捧E8'XSt!rjk4} t_6\߂HLD:] kY  5Zy,u` Y:Mj!zҁ8ynWJDêQxRzc礠gR<ǖ-5XمQLXj$j7vܭwX-E^7dx,U 55s-ZsfsdN-rQH"19{/,55=p9&R^䁧ׯ|MlYX^fc)2<&Cd7Uat/ir/`O]U pmu+4ݕXzO`bvU/=TmBw (ۏG oamD쁥S}W``&20 { a>$]V,< $]+VLz l*[PJOJdjnp!5GR4;Z)Z_hMI=urQ-2l2˨P^L%w~ s>TCOx]$lAggQ$P|L_-:2thl^H=M;Y%j޷ص!.6ɨW" U|R*{%ܥWO?V\\ "nRa椵LQEXve)ehfI5~q~vYѺE,l5s1rrB f"Z-TSV ʚZ5`y;76`iqJg4S + ^q1nneW|^T+P( w$YͬhYws+GAdIYr+812SdjP4|., ] Ͳ1@nTPwfG}S<ދŠLj&̓+e.^v[ <Ԋ#7h!F(ꃌaayz1z}`_ڊ2'eD`1\,u퉣k޳A[KR҄>ΐ=S &&Q1x(iJ:ކ;Xo:ׂzm1vM^x0iL9d f 1S8D4K `69h B9rk0a @TS41!bLj~1}@H|ǰN/'OG'31{Ƙo?_& 4m )R,NTയ~/~|}j}=#\TwxB?zR99#HV^wu̸͕EBVa\ ƻ42Y\&i`Z?@I&tZמ6@y*ݾ)pY:ǰڨL/2V)v",p1 Fs߰-b;-M5ɐ٥l`(GY3 /=b2||Qhf-2[|Ux!s\ՌqosE^7fv'`zWOA[\ 1a axABs\XO˙b^J+ĬP$ .*v䵧 N7{aVp;Bl:?^J %yp!$}\%)`6/qK=!|l1 /K6TrspțhKuVS gX}׿Eߞ3w_9v LKl隹%NfrY6owx@▦kbQ{+lYvr1 ּ[\MVaф(ADd k[՘LiF~׫?aɘeoݏ0W?Go](wmMʊ=[`y&Bޭ( .;Kɿ&b$Lt2;q Č?{F {%*únv&JV#TOp0l (rx4pV$U:γCˀkpV[ @@@iK!YԐZFquBc2=܇ ,6[ Z`3UV= ($5zl>WIEYR1u\2DRkL _9q<9 iKq+- 8NXNw00G`10М9bR"&1gKf0j` @V1%SF)j1 XU/՘Bwf5Ӈ5L]V:Gd9ƒM>_3%JV/՘w|"CcUBsMQ @D)l5UZndAtA:EͅoMIcdXDӱ@#&xvgd"'!Hs߆$Ӵ#K[\._Q%xlls hBn] DYGHHy2Qw\ ;@M»V{TèRZ1H^NώّKUΕb"J_}ל8P':1T}l#\CW%qJk&P”-?х?!!p ^õblϧ<n28p`^j#5& MŴ,:jhxz ^J&UPLSKkIa漏;\akBO1&jc&")֤nA{Y;3p s M $vc\ځ˄&kLaRH[Me& i4uWI{Xo{7t1% U5īH5#ZQ}̣VDl,$RD,XR1vqÔ{u_[&qdteAi4L∰p"4 E$ ȽwEX:X[Y)0֩iNd}2ck׉ C<|cJF$+I8qiYj}i7",]UpM&!S$G5(pP(F0 8Y&>9F%WXPc66jIW7DwYW֦Yߣ4坝=bQF?j2t ;q'ezS)t^!؀()ZQƲ̈́y\wӌu8SrtWyk'!W)Ǧ3f*&W}dfyJQ&~5]}t6 67=|]p{6L8,*W깶*\UzV:_XX~Գ&DJƘYC+vY깭8\gu״ *9p"wؕ=eh~ 6, MÈzW4[4KBG.5:w\[Ш$px*}&zz".4+1NbjL W= ֆڢ&'mMb=K$"L^r;T Owķ<YhzZSj?[>Ż $])ɞpVā čvF"#$cb.q/Ƹ$2®kLa} ;[wH fẃ܇5hHק`˦ߵŠ>mC4`s`d?S#:[<-23C>MXk&i4aSl-vO?$%[clg/K1ViĥVHH < d_VwIܻ Ru]ʉߦ51h{pjֽ+/ijpWKaYOW3U?nU^:7!?:Ϊޘ Gtt6ӭ0jؖ&ޏAU鏟v ^(KSy IޕݣfW#oAyg,!!(p.C(z[nH3%jX@ y).Vwu؟9U_[[w(*}jA?fnk6JT}ՙtR*RV@hj7RB5YY}`kjV\̠n2rbȟ`Nr`hc݈'q<8|xc9t:x'rX.7O mϙ6mC޶ރV.OH;tUR.bX62a郅gd Ƈ<|3__m# ;t#SqvƻO7r?* 1b؈ܘʠ0|i4KEx91yt:;ط a BChzoQٱ }{\eFmpprÝ= E d)qd5z`U(֯=AeH>d刓hQn׉,e.;2qdIrWyig!P&%ؒ*k"i6JdHZ"Jnu1'%+&|iF"hI-DK!h_;+UߔV/??M0kUgF\ ~ .+tëxMMl,X)d,J%:i-3D)W3ɫ!KtQ{E`C08//%wb-Q+$ L#&%,", NJ:1%;#6S'x`x }$҅hBkX%D"`"<)T!9.kap9iqWRP 9fb[J 6q:iHQF{ R9$R<U,8,|Q(u9i҆W(˺h>/Z +8 PR 8ü H/.!T㬐;vA4FV;u@!JL%B c!M'#1&bt " s%jκIGu0wjV˄n.e%K#mc|^g%w7W%f 1%!hb1*XjB)8A EJQky6L`)Ӓ16 Dx8 z&Nu)ry"0`Ѿdn>?N;95BHI)R6VJrG}hp.%~ɍIOm^8&ZfⰅppܯ_~GS?i hj6jG98_]ň cf0Tžw1x&Jƒ.{(˘F-Cq]nnw;Hc\0ity֣Pwٱ!YPǒ[@=6_oCd' RlҀ+TOM@|B1el۵vm]۹mvmW( # $Nd4cKaL2q*=PA9 Y(KہZei;ܡ9+_7ẉVn'6Onk6ceCpJU#z@Q lI` L cuh67J "tX0]瓝?A vv  7N`Clz߷.yo޾~[dGs^xb6~Uoop27l篇{ٹ?<l xľl|߾kl~oG^}܏O!t_|N>OY{1g rsr0Tja^p?0[_ݢy =yUP Gqܜ3y@*?͟ݙ;vs]-xz$^|If&oO?˒ɠy_l/}+[lxc?lv0 6߁I+omW:]]ioȲ+?.0[ &l &AkEҕ$IYڼDH6ꮣýWܴCۜ'M̷O+vφ+ۇkL)xn={=%}0:$Ͼ$|r[oUl7J Sj/ܹk|W~p|t8-kwߗ]Y}q|_wɇAHZw>,F0g܅=36oCy=>ioCG]/m2}>aEť7n-~ig`JWnw? h iVzۡߡk~L Tu?྘$M_}<&;A]z?FG;^Q͠i]R#"nó^i.7Lu/N'=9h{wEܻkH?/oooo6 ~U@q{l6zǍ#c-|6d%[*&cPwƀ菢4Q׃c*fXoѽ7[#iy}gkfJ]l}gma^ۋg[UEH|K;#f/K y/8%c&6៵Zp "Zp>8@HGhuو4QcYMjh\䡦q5i\Mj>̌TJ✻r Rq/IK8dPpSXDa+bı.(ULskN Ƥk-U k8`gk8Xp냃]!nyZn`aYǂ 93>%##UԱPT>"ԥ]'"@yV)ŘޏQ#p$(G *!8c0YA(R53L 6ĥ&._iM\jRe}e K@1G#A&HLIwC\(eSh ^2c+cʱ$ {Y4˶a-6lZl%.'5`&Z\oZ k(jMU\vh"וvv( 6e6 @Recit$\HHiV) M$j-BQYEֳK/䮴$h*Q1 trc€?ꉱԩ R B5)r7r{=XXCV5rA89L!Al RYb*2a  gRL8Æ?Mgߙ?L#piÂgr Y É[VU MV~D5^9rYSԗ)>/ uVds1c؋Vyb^4U<^EA$x4u/ZY\WÖKZJ*:XҬinsY^evN3FRxJQ('F%ld S_̓w\{WὋ(#DdBjw[a=M4BZhB|2ÅXPuL @ E\LJKa[W/?r5ѭc0ʌ5jS M|B_|KJӳzR$_iՋGbgQةb cLҀ P`5ʓ@p"c"qa`!BsćghpgЮ*\^Q[܃<Ւ ,`+ْ{6ؾgK[ d3ؗKht1+$5WtՎB+M_\tid7DB_N)yiQPכL\v(toJ}ۘA;':'  C,ܔb=RS9R* mE3a}ޙsT۞ZH*9iN@#޺{E&דF -ST2"mN8pajlOI^:w`2w:G]v+TN{= S*B=̏ƾzZuSh,2M`7vxZV?X NޕN Dɫ쵃$kt8э 'a`gL1i]F#r>JIdp78g[Evvn~dʞk^pPiN ;!_DM C#]s*Z=RZQUwX5mٔB ZsfW1}TM;۫gkRs\@CR1z^L,&юRTHAݯĘT0,p1XzFḘ9*̄k_U\ŰCw(bdjyW)-/V7baN<~V{.$Xƞ鸐;O͡.1_.e mvn?2 tmkke,FQDa-.4\:l7+g̋$d吏Cx7RWLKze$ž'1苢P&oޕp w7|:ۻYC%a#NiO^Jټ3}? Ϛ ) w\v!3?xv.vnoh/y}+ў.\|M41^v[;=yqzPˌ^~uk 陓?^쏞}zw~3e= zwnut@yyt{/sK^5QF6._pXfC^I!IMnowL<ݙr!B1,fmm\S\H?LפN/ǰvzr@?ϛY.nOzb L| wDW'lnϼ/Bc_[{Џ0.R/"a2⿋/;7O.6*y6v]*Fi_ j^Fa_cw -+KaIB'{]7 |~t7|~h 8p|e0<;>+pr pۜ>ISHy߸0˭WU(5Ncp'^ӆ2|t8-=oyE|݅&!Bj|O.\왞yŮ>XOW> z? y|iKyAť7n-~ig90ҕOZgڥtŻgvhF0wں_*Ӄb>n_7}ȻF''w(~AӺF_2RG^Q"WK)A S]IO+*2r1PߥI{ab>#rOmQ;v 836,:PBW3ql6SG>ǭ%wufV4oրΏ23WQ*F>b8C`MwU !O+tD180 h7%]:hJinJ85 c"͆ nMpk[ܚஏ.Z[%5E#\LpK: tNH!ߺrUE%v|̯U "$*# !11ӀEQB ,"*6JkĹ+IP v/Sܚ>&e&5 nMpk>;P]3 [gyp n^'2BXJ{UYBQ96!%B.I ׁC Ҧ>+p nvrwE+‰-CmTH5jaaGZG#Rq%j-ݐljz[Gq&<<;9i7lyvgyvgy񤡫{Vt/{+ҋ.uZ]69EA%zaLM,PRF NӟBpy3؅98uOX/@ j[07>`#+"~ 70umQ5 Qp[7E *JS &dCQ5ls1ל١K:lrF1.Wnb7T9 UwCPnb7TqC7J EV8١(DoOEv>YV9 dEE."GN8\1?rk AЬm} "oL)p?@x{>Ns?= 1^-W :If=ĐM>'/uA&LQ#UTAr՛fԻwjG^ ;'s}B}&~t.V*',B~>ѳPW__8' _]yb>lO~ޜ0 #k&We1߄*^\/]K:w'jI]Nu7ջ>HJ%x֏{~?}3PK\ 33H c^[^+{v$([2B6+Z<5mDI H-Pw !26(>OC@N$",6ND_Qdh&Oo\C䇧_>dAqUzTO8ƉVc_^^2%}{j?.Agɻ0K/'l.09hMO9q˗"2xXV!FJ9ꞷ^iɿ:C 2_H'"JY9%NƩ")YvL̵P\Rf & WYtBMn'܅goO'L{[fgy[r>}G_| 'IY8S#?@|r&ק[enIV|\竱2Uq;`Z~˽=I¸\*$W?ZO_փgg;_wb)r$j2jbJdVG,VXJc{Ө5'7`B?5}=;{Q^I \X ; 11;;^-)h8萄FhR5^Aʹ(6*X kv\5JTQQGE֫Mf Q_o$5+b6QpJf I/=5P1{E'@nO[$QQd eF,d4K[t(-gMo*urrI9[q[ƴ  NU.Q`KMҠRJ&+DJh9geR%&lfDUc`%FFiQ{*Kb|5d0Ӄc"E  9G>IT{]#S\0IJc? &VYǐ= 0 /o0=="ihYC^˅ MES EhFQxA߀w_4;hx -W~jQPhۼt!,b2%6_^ULA %je= flU8d!/Ɇ긧q6lGvyJsCN./uձ8(8R$Xюrs$9^'NhXNߊ;#"4oڊjd(y7oz[(k{]^nFmdS7Qś!焎)TDJZG+kŽJeLHN)dnYن&IR4(_0ci K$YQ"RJ0-%aN1bF!# d!w$G$Yf ky{U֬j$jN]K3cnVQ2fTf>0 d=D&xVPF |5#"%j"{.qME&vh? 3,b5|=&n#ya[[p8eL2^yt .J2Zad LAW Ĭĕd)6np $CҷO+Du3q荌~\N9%EQ "Y($~V(. 0#ΐ VINJiWW2,wүg'L?mbE[矞'ee=ʭKtV@ OoLE\]W J3L5UWzм^F{;͘ ස̵*XJ"<K4m" uAKp`)^ Kz`i ]Y`IpX @Kn-%uӺK4i0W"f:RΤ5VUʍu LCJ9͠s)ϵ9IނSȼtoWZsdG S;[R> vR!+W-3l Ie0YS+4ޕ15rn1/(Ar?r.tb 2qP6Mup=:9;-vqN$$KnL.k2kAyFqJ0њ먻qGFn&1 @|!A#E sH5ml5YY)G5po70-00i}`^Ҷ/.F1暶*n 4gdIR=?f,۪DvHdL%l|toW,[DENo)YߺHa%#XjocG?~#F1{_;sv^.È} e(G,^ Yzrڎ-y 89̫.vvp8?8PONQǿ\i-LbotmYn8Z(LʻQ̳5u K^Ъ)i%7Dk)ˋ~-B1cmۢ**֠$vXB:qhPEtOD*9'G%!CZVST 8@mG$E!++a=5mo~i;úX" f}YR&dw|Cn0 A!d85ئ :‘ˍҔӎ2İ'!Ȟ5.u>OꑗL[MԺ)T%BN1B)\!ۮơR s 7:&k^&;Δ]dNpL t$B ."DM IW)9HBJI%qjdes*5OE16*/Sb(x3 VYV>L)X,2).KzBA1u?P*E1os[B9~ʡȠM--jUUTSˌٻ7r$W @d ؇A, ,֘m$3ӥ3Aե2L!}3_P(dL'Q8ƨxaOv|b1Kd_WBGJS&H5q`%E<|!gsy" "ױ'Ҵ"8op؍TEA9I44;qîÔ`>Ja_ j(ʠbM-G_2  /(46 r#Fq >´d"eH=^n[ YCXlʁٟ{7Ekve.&s Jո е좑CРGYpbQhuR[V~2Wa$d꣩ȴ) p?PZ 7Hvƥm*֖w Z#J{ ge,[SPA% s,rS!p5Y)tZ{h;gh:؆*`rtg\fDe{\hmwLE+S8ZK򒒣h!=^)n`"& 0Pl)ɯ(%@B)"R6&*JLhA'uzv$X6 Hn`rUP I[xc2$@IX=$Ymu,`:@nu bXȤGE{7ުEN-:%oߪ\If-vHvUةVDǦuw5FͧJ;I]z͉3Tʖ+O\ꌱi.R4X$Y f5=q\Ԥ:Xg>SکC2 NzPYJ( z)d# m*!ihM9lp>gJ/ISUJi뤗n{@jr&%NX.?֑2)]KEOZy纕I؄얪}:G֪ (]KHJWZL#r#ZM*J&/ؙJvQ70ȥ80fK~꾭Jox(tbKѨgԺ\ H_2P)b0 6ϩ-4:: %2r@AZl` HcUvdyu\]([㤀A ʤQUA Bl`Eq Bd*܄rMBMzcT(v^[EN'Lu/oB؜5Ȏ΃[l8c^T;Jv}7ž887־zx!hha:ū T):kǜVș PFyY9NDh+V>TX#垊(W8bׁAxcX Zp]1܅Ցp!]TGr/ lª`բrSe[|Fu.ʧyEB(Ssd'/ӄ&h?͂G`i~sG2gz!\90VwZ 6R Kh/ ZneoYq+;@;zԣ9 [3bkH{i]WRѨ{ReY-7[?f6#:y=Wi3Vht.hw߿ .K^1?y4V}uBvvfߙt h\X^qJ2'/{&{.hY(E4t%/Y>+6Ӻ@ȭtکRGe#%w*[(Ll!L~N`ׇɧٌu䯺l&]?ay5A@ِb5>]/9_K1_Ny87höi6X+ca )%HH E ke9oÒikL;Y-P ýU;';*\C " d"yfOg;߂Lަ^9%ip]u'Q1VIߝr4 4폿2< od:deDJ+Ky[͹ "&ko{SŨI aĩ`jqbrt䲩p,_8j|>Me"PmO$ù:RbO3G֟UAn֟j#Y`z)(lȶJ-u-5!:hԤMn?.4l 6xQS?Tf#Kfw(aLTxe}ϸx7\H6cgp9T_CsY~x&$p($ XuP7\'Hv@o(Ҍ2*YDqHH'^k:,o6@1|N/a)< 0|;[=r_>o:V&G`xlf43pPo6&1Ck A{uV Κݾ&<bWe`eO4䨸ݥޞjh nZbBo~z6>N~y|S!w>d|GW<2 gpUA _vKnI/lz1Byֹ0tHRKZy|F B!vyO+O9ɧ7_' 4s?~?A {|xreIU/tsm9nveݧbQh=&=%K{A< qY_8*CG--G7K(<uŅ~sH>K}Bpa}Tr ;XTDG% g͞cXTnِƜׂ6kI s}rIM;,8  -8tđ`'vڧucQ',Q瘱n{`F=Wޛi&~a?z 0 ωiׅBw x~jMPz.Tq '5bjmNpV пXӹ#Ѥs!>Za!TTd,ꈅYD,\" 'HYj2uژZ- mm#4xh\{wDZ!^_ ܛޡކdVu;J/}o;Q,ߧV#Q7&w%mC[r$mF(2-5R ٔEʔE g3n(+A{`(kOw9wswn_tL؂30/^d-٠zJ|p擄mGPz7 Fz. Aauk ,y`$*ArƏm`K5{`^,2(0f[:e \<^R3$%d h9tv4oJlxa!3beЈcs oA0qa%GgbS"!ɔ#<i0]J3 (f1Ijs$: cri}΃}m"4PÁOv+-ħnBOYKdKԭ;Up82Ɏm?Z l"ݰWxgKFT)M {{Z o;}84TKoea%zVn6X^;a."-u\*k(6Κ2& d+gf(b%ɷShW3Z+-MakgLb_͔ƊPfJgOױ(OZgWn%ZWxFQ3%٬.?:[O.,u/JX`E,:'UP ^9:9iqŋ`8bhǖ`& 7)L t cݸCgx{ȹicgE7Ia@xzF@ԅP3g4{rgd:;bq o:I?(tyߔw:hinOl9Cր@2c:"$`@Κ2(fJa4ɢ(yvE+Ĝ.S=^w6nV{@նWMs[YQS*(݂B-jT|ulkܢfcI =Pr/hrEĵyher $Wu!qV&|5$٬^f"U+B!sU& `>p  KqAUW:*n^FVH/"-Ɏ /KH!j[ą(ҶWcvKSg@6}iN"U^ѸՈ{Cz{WnB@,j0DAQ}+)ZQ{H b^e: ۴("3z Ŵ*|yEy/.Cd)md$Ȣn\GɊtn×H~n#=esV9/0'6VI4w[ 5/ΉrQM\jiMM!f f*ĎoGN"v8JRYȗ7]'foM՚2sn=ۣo@0Oc"F#=7*_v=2?vDC0ցٔ0*_ ~&WtCtIn6~~#IkHvlf>/c?tBNoWJ+B,y(dY"w/XCsDxYf>wHJ]Ldg/h6Ċ'{$%u!)N> D.S1bm $"\0Z.B &si(fh]p*B ҍAW\k,7vmW9Ѯ$a*Phe''޾h-˒߃mvVMJ݃贌gKh٣2)nROڗLj؈S# fR½}P4iWw GO4yZ '\ek()"\;nq-lyV?>+VW6Ɨ>s sRgV-B2)/0*K\saw0%"k0}#j'R.J|G|X+,*>d_a7limp]xo7GzF( ='15M(4d.-ͯ bVgSL0ޱ4^iN$~\K|Ķjڣy1bCno#4_ |Q2jusqP0t}8}I/f7ƩܱGwM"e2sTˡmݪ50%4H MUq`M_53t ^A~6ϡ@Toݷ7 pwX~sm ly%lqZ\5pnù{Kkqe Z`A$UX6Y+-Nxy- Y˰esI?kڝX橘 c S)_$L|qGMwr3m޹;j5C3I`43+tvr-O]<5'Gt[Nhv*ϝ}-_q]-'qc&HޥcYs q|?F`[HR@BԾwƿo̴%y ;' &C)Z$mmۍ;lOLBoǧp㞉$K~'9`v/.鿐 lw"|`n=kZn>8ӷͷo}Dƾ ׫lXpo`/gw'XOǯ᭛~w9 ޜ{_V(yI;#?([5u2^ n$'v&Cv"3XAPcg2K҅c B.=V W['%< Np2>6-UNtiq{f5;[8׃tLrH5+8]"c8v/&Oڱh?q'v_..׮΅v\-C]Ws"_1'f-G>p^|5!SO2@ 4ŎEvcB-#`LV8Hs\ڃp~<54UǣS}68<.spW ܣ8o_<ʢEteFteKBR d-u?o 6*پ2 %RcpȀuS!UjQw0ÎۻyWmGzr;:טŋꂁ=iΠs}E s=b|^k$]ҝkok$j^U Y۪i,2Lаzw wn5܍O? DX:@NkM#ߣ?n9[}Ak'yq׀+sk R,q? omud'RrNw)̘@ⅥڜϿ$д/іjW!b#X36181Ha]wm`hlؒś;+~hyE׳t:x\7J*9+'" !rMuO.þSF3ER7y0@C1EE&ǏK[ȮԤ3Iב(=!]wvoczRGs\d?H6q

wzr*a'UT6B.Q-W;˺Q,_)Wd` C/N *v 0hqh+ld峖T--[Bg~+ZRP\nClHʼn}@&MAn$ ĩiBH*tJaE*6BM) I9aC츒R꺾G 66s ]KeQeT v6t BK*=o= ;RX5z. $ Nk~vtb*Bت# l;,Tu4!ĝ%SK!d\vlU$\otM s~yyߓ]olɺlA|(x/G[3zR,bJO:Zn)޲ $-{K-iAZخ l;MUg~|;WyOSmDCpgd"FXcU:&~|^6m;2)<װ\nE4C7^k`'JZ;ޕ5q\뿂Cn9/KURT<8M_˦{/$LImhϵ<[}YM}BU$آsNDgŇIYv@Vݰ7L&9rKzBlN4.\#1!-8ojnRpl\:&l4Jb ^H4rGChjy!Ec7!n9(=v+֚ %gQZv𦭺:G`P\IܯS 4 2ƓGoDx W\&*yov6 dBJJ\ӻ?BTJ„mzLm-m4`UgFww+I晵ۻ2zi^`g~-,A/NS,TW qŌAWV z@PH9gϮFw>ZN$Q8^bK9N) Ŝrti1;X.eoy iK/^,#3+wizRob =&i^(3)ٛ B妥p|G0dWLs"ъ+_Yu}(J)vhC= s`)Ʊ@$^r$eMŲG5ZMohӈI.eD O+o7͎5Q([ӛ/tB%Ap5ik`|?߽}[SE݆ځjW)Qm M vLNBd+j6 P o?:bnc _S+l,- ςO HN{0uf(@qӕ!:mSLbm N"皀'O8+8TXm ?F7aV\gKdɰ+ ?,i,Jbb:HYHnX c<CYYsgܣfyRiw}j^9ڧt[K&V]5݄aǠB E *)PT 9AænRаb"; > BDZY hLKLBxN<{rCNc^  h aaw$mpDj&?ܯ.<ញ~PN-~LMQuOJ0פe& Нm'/mWnhJ6Sj?2A;hR\B gm,ĭhA#k46Ҿ;T9jex"5p>b)a3MasGZ-ng-+FWE3I_.+gޝά3 $U3,o{̐^"x[ŏ5)!ګP k,:-k] $,>#ɼ %8I\}xBiܣlQL[^*'m⒡?9nLn3"5Z{Bu`S)^NfJv">Z%?[^»H4o!m@$g-%g{I]PASPޘk2+4WL Vsbe PLLLrl[fŏ5 q~,l6&Ï~OhrH~U| Ckq܌b $vt<.CзAL5io@^u!vRյ@H' CӯKQ"yy&ކ@R0_:o<@TIfrC`uI )Y/ . " -- "gwl0)*'!a("ꠅ$!d(hg", )eRfTpإMjlCTlbe?m/~=8+ðT6>< U*aKqJ)O?xZ[x3m+M4E]r, PQH7TOSR3RXGdSmK&),jTB{ơ: `N4/ד7w>\FhFK=Qid6o bepCdM:xݔ5_7S]ksE%Ԉ{QYo= %9Y/ TJyIqkj =0Pk|B;0#$_3L#:2Ϟ2:=z+m5mMQ϶ ?et3 TM}Mxk>>||vxdRh(גAsRŠ;|٠juY.k/>Ay=4CBY3mLY3y5up53%-kG&ۧ;IL_giu9;u^UoༀoTӆ)Bj9'S0E*57•@(Akc@5ްH#af3 "&X3b+r(t=vtwwnニ'.Hٝjaw6k>=6=`1IX! a#p{!xd<>'f99 '79gVʘr6XX5 jx$N{Xb1 D8OGR c+c4bI@HT:gH#Ln$R(}aZd$Fc1A9yC 3x˨!JUDv3cxΩ`(n[,/+m)GA#&034 vEH"(3ia30ՇVT~5Ky=HɩZ.Ʋ΅:աYJt9K2N4'Q&>zMI+13Ml$)#g rt\ ܙ؉2c!ef TF$l d[7|BsVVo9'?1s_ng~K0W+8(勢Gd.qo޻it\.^77 a2ʟڙӕC~/R x?jF7|vf{7ykiW͵sXٲ]0=,ya@!)f e 51pOG, &ʘ+^ ar< JDS…\܄Y͂kkYS΄ ARvjtړYLKGqQ)yʣ)JƔʀhXQ=C'Gk>)L)MJO0a2_@b$th̗y>`VoVJl|S ]8!92wE:$d ަPP|Sb'%Po$A I9v+򦟥'j.ORCUX#Egc))ԹyöB{cEV< %$1xMAr<nQ=P&R\d(TAL18=l#()uFPĀ>(̋VdIΟdbi#ͼ=KEGH)S-1@<&J^SYz!&U<!`|m4FaefhZ!/ LE) A2譙ZؔP$8n-T0JbC@@JN16n#J ƇjvAKf;f [Q G 76Vkx6%u愊USL6o%RA8W9WI!L+eR0Q*`1!ir?o7T H#GY̌T^xRΓ*9CD TPMr Rf, >6[UFn-rRz Rt9jk PO3IĒ!8O53yMbLQZa(9Prechɭ1JbNcY?,۪)0k#, C""drZuRU P | Q I.gJi!pYmiOi|;0BqVYV+2 1䆹!-@j¤A؃0;RiIl)8#gJ_a {»F#Mҽ~w~pE7<epJ8y=scbENZ)SMY\@n3KiFFRЕ 4HZ  ,^n#X}r˸PH1|T Ք|<@,Gl%,.C9{#eaXfTbzznzԆYf3h)%IB0-'xʁL-ShuKUgNG sQanW"Hx )_zy zLR+p!CXp!Bs)A.TUd1wt JEASO5$9%:"_ F7$4@0 ,. ^.+JbH !7RY;muWÖYЪ2 GThi;r[Ƈ(~lir8uҎăl{Iu}j~#]'8NK\l?}Y^>h fWմ߮'nSk:7:2)d w}-oˎEE#2z矿wՆD%RʂWtR Cr_}7e/xe}|Ww煫=4hO 8f W VI5: ٚz%Wn~ɕmib DCEc8óTÍD{Ey`UP^Gzék`z.zR uz~8- EIB:`y۫ W'Xt#8|2(a-%FgFOGp U[^1?YoJD߻k~}'8΁ OnOch*n)Gj]ce=l y-o GsSf u2}6ƾ'19g'/bjnĮ dO?=#4/n]/?($fÁ5.~ؔVp|'_.oܭ8B=9EÏ%=y ک"(9%i7gق: _1zxt{Cro/j[~DJ/W~7"PE;A~:x @; kA16^s"Kސ,JH{]ZAh trXj{K$|^yx_jA/~fw72&W>M ia?F?kH?">}*~ ⾏Kjlvw7` mk- rw'ήϝiXrBBO:,kRHAi'˫sU<;ޱ 6~UvA[+\IX_16Gnrk:0}Y7CƮМfd%®u3=Aa E&vs*S RIRd3WNICRg$ vG" dC:SQՃũ_k& 8zEC*ZRJM%u4=܉XQVyG  b ]?#5N' WaHsF%#Ō/'Q9\0ބ>@7[ u9q,xJFT )\vlc91[Y@NRzØoV\u)#og\.tPݏё|?W?gٗPҵ?⫛]\_W_Vw?x!9jy#ψM4O嵮,Nf?dʳbq&ly_YlR ,e|Df8Ny:]te}:-@]\  Gw@@aE|O(Vi>{ ]Eֹcul (ehVa^s27 ٯBm`љsa-c^ #2D3 kO:nUm*(}Z%ՑU߱hYŁjFO)1a)iw*TuefZ 1u)H2kOR`v#xᡜIS7hȐO@k16]CR v+twk?2b[^+>VI&O|҅͘Z(;.CY)~ )wQKdv2+7)Wn s\/im˾p;+-tw%2Je_UJAm18ʛ_}JAc[ s]*x2Q٤y,ba/(y 6Oصy~j@lur_U*J&o뢠[2J9(ck] GR+1 + {9-:H؀C 1 !; ĔX(OT`ދB;0>Rjetw4ǂH5é3t}*CkgY}]J5;q/I+l]|:apE ; M ] 3Z3;Apq_҃cD7 Rt4x vPgA9zG8 o[jpdi)m+m0mYLGFl/^-Fx &EVIZ Ab`>"B6¾/F{B6÷j9rdd ݯ͍c7OoCZ\x}!b=c#1/T*y^b*3">1M/qqzxTWԊ-! 0\nKL=g[n#cE,P(K"bU.M-jk0WE`kuq3}: sOv#_ku*Q<RVxص`A^ ΰXt-R;\ !%kúͻ^ܘ__UbD2|j̮>ۻIxuz>ʫqvs5Ӵlx`t-n],8bLW+4ÿM?ٞ ֨ǹ1VLuE*in˧jm] ^ ȍdos];UDI♤1 +E{qs;kߍ~C~kޣja#(:rkkScK4q WVBwxJ2F)zBp.be{kd/q^YqJmz=|=` =/>la BJ'YR6,kkknFE=:qo`$sR[gJn^r8v[,ٓmPEI.c%SĢF~G*dZe|,B8ԟӁJEL5ì{}*(`x"dG$W#>q[J?]N)1; F|nhߏ({~KP/t0%_?7JTJgF20|F)K$ɪaCXbh@gҮO. Cc\ܖ2'G}d,@W ~9yRr,O'_Hܗ^1I#쵽Cv?m(PR W=<H3dE;b!s_ry}kdnc ?qgK֗0>v$W *etoF X1c _ˆ%VPMGI}4lVbrnr"D 9͜.8)j޿Ҋ?No‘ FӢ6I(+ C_M=|M4 $qelRΡoqT^zܒf OP' aRbwxsi{f/')v.I +6I6J*[ T`h |PGT!s RLIR6~]ؖ7b8g ɕ?5цwB?IrَLc[eA?{_E"# 3cn8ӚohCS0tW'%z蝫4~en fc~+.: ãYNcYVmZO?L߭J-O8!)xT&zX.7ҽ=ٚ)ctVXyOp~-&0Jpfb\&r2>>F$Ck[&ʔj]eb]B.;ůE"dl8N2oo|V6jqY?^[qh&h\~7wOǩo"ZKȢqDyH$;腽 &,j|j od'7vzbRNS{\Fb ,8Թq rW!F0PK:gVUMØ6|HOgJ6,T7G &>dbTmNf(da&ђ}?׉#{?އvgٴz6ƣޮPCxǫ l8E^G|J*侏ߖh#ǪhCOܐx͚^`M2QrL(bI s{FC6RDum*)SKjCRä́oO7bt<_K7֩r-&cLX*2˜#<&=u5f6sMusm{+T&>9y$ŀh$U H(wⅺ{2Ɍ"x,)XA{N,1dG)o|+[ne%ouL\{>hnGM ŋV?<[\w"RnD25чƑGLEX/L,B0AB'W%gͨT~7|]ʌ <$f2LbSxFΣY^]:0^2S1Xv1 ׼Ux<ғ3/>3zj:^fb=dcL)owlsY~CòJY}g9WYL}?b;2*wYbU^7sQ tQ."JEDi4=C.GmP-s6Y8mw&|~,ޑ{Uiv"]z7 "Z#*r4\ѿ1GPrt;]JGoA1 jj)-eݡJ٨,&]!'hXRT_}^[{N~g&4 4guT͇}*`|'yYN2HϛESժJ pU*/Hv2> [ʥ 5$ʫWѝdwU9Ợ4@TAq,"g"'*SB 9lnߗS{O֜1X I1Bls;j(oʍK?|w)?-GYٿQv/n"v/nۭ.h\gĵ'p40NF3]*,; 嘴^jg8E@它o >l:$  oQmTBث/5YOA)MiܗrRuAHKEFr$38rRPv#%_ #Ɉb heiT˧H;`C \ItޅEۡRoȍ!@WOr sI?ZI. jWhGw&/s-/q`Ae Ӊ6isZFxiޚ"f.bkjkjn)hFLk踝i\{+ B{a'7'=[ruNR,YhA#ne+yN^Fb@#Ɲ]+-+vax}<ʸo*|yPFhA׮Z~sJq@hAs'ril-h 2ǭtY-ې6ؼ&7m4_qf@<$ 1m5/@t9%ݯYa0] DIXv*E g;Hм РZhR51',ou*iLr|ak`D,?>UT+.bEEgx>s5}wZ*9l~cX0I &{ & Ln0!&6c3oLCS1\;ps 4JE<ZL!(2XjC 7Rp K4kmX'Hf#fWE~:PE> {*AJ%. (l i":`"z@QA9b)NuyٵJ/cZdo_fe i + <.^PQ9mʽdh@ Ǣ}JW,)RyoVvMKJ)^j5O{{iBQ,&)= dQ Gdv8ږV%H$L-"PBRޭ/T;S N%+9p;*!CJࡹ9[\l,jj)eMS\( ,uH@&X(7.ǑRL9Ŕ |Vo+("^ 0 =7/#S{0D+P>' 2>mGO@ !LS׶شR"]i>b}00;u3{NS*NVh.]0B!$]pW/w6\xٝ9 ڕ T Jc}Ke:9bL&Ę1W^S H"m*'y?GϕC7 "ga7 \:/'$=[#8FT4HMAM@'M]={ h"=jv G@L'8فwl x Us_I:ȑYCnSc̊A9B n5k1mDǸTzV[ޖF0otL/S5cW[5NqPmE: {g zVZ57bZ֮)iH U3{_hu- b{ kF=bg * M^oK5HO޲G!!黹 Ig]Q!O:ѣ\湡( 7 ۼ4kKM-o"B7'WK?=w{Ϋs ݊y}9t;7QǷw:g$-A̸VunwD{kP$ ؊n9Â=8?%: K*W=Aq'GTY3i7FULX^ 8#<$ġ׍C{> h D`/~knwraC^COhf5Oy`B۶NNfOf~zK{=hs6;yA 0dkڇ1yªs^ptU/8`X V1!A I}RDH(7Ptmon1BlВQ8U&D RLi9w{p`po=BS("yΧsq@]zÅ/%Qbn@JAyytNrrp>آm zf_7gFx/mw_nۻSf{DRP4L yVBŸ߽;pvw8݌Oy+noe>S&7E~PLq?&p( 7g7*L)/y¬)/neV&D췿Vz>wSD;N(rݯqG Qv..J gL#jtPd\ A[.xQx'Ete"hH#G 9 H1WW ?G;gAw{Wa ڡMy+Ŧ_E6 cr7f2)`쑰}qͻW#.] jcY`Yy- wƦ ?c#ٳJJb~{m*7\y1|s%Fn!d򟓿jhTSeUѿ_:5ú-3;0__Ǭ狺Q-ףB =/h0G JeR@7b m$!)MÛ@Uec)+Y7|l fwmjS;}6eut=t}?-wNj^->M*l9ƍj>mn)JA`EA(tnJK pS᪁㒠QeCH.?.lUd\E@~~mj ;bɚe6Y'58Ϫ)jI)HF)H5 3IK1扝>]즳-B+/di3QHh&ӄɎfPrh1&!&OY}~շ"ae(m'O@`4\ 3|9o~˝M+Yp#@!7/ >nQo:gˆYvM(1;-DKъ ZK"]{;lgx;{;p%DõշB1#L%p;qUV'm3,b!$}cqRZ_oğqhrL|,]V0xǨ!qs+:0!Ot$zvr~sVEu"_n?}t=)6= ﳀAJ7P$1T"MXj`}(l|~wPԁQ z,@a'.!k\x4~8BAv%lO1*6% ccmP<<6-ϯͫcivə!ܒB>U?h֙qoRۣbAFwF8M 6.&H(s )dA+!E+  W_LE.E Դ%c9PeU-80;a#98bJ2SA 2R"#Le k(.whhQ1EeSK T aKU(/q * 3ApN2d^W?VK=ӹ&Z&7~=P°Bj2#YQjP/ *"M *JX>)`MUID^O8I.*7STqT搫%40"d90dp:1!#-S=7Y&,$ș` GŁxe%63ʪ)7bIh7~Ũݫf[!pb?,$*IXr@RKZ_B[d 75nj/%8:RȔ5e˵fUw[~XrxnYTP/UkZOi.gjRo_~)nwgVjΔyc>էwuSܫ5h;a'~ c3fՀZ`߁pw|BF`@1lr]GT)& b*dB[AV=Kwly+pPIoVO{rW_pSH RGO)Ej gBZ) 6I. 3P@( +2液!X:+5]e/+ KLaմ w_YI{HxB A _K+orˠ2x+(7Bvvv\%\8^s[y~#@"5IzfubI*{VU(Yxl}:BKeDt\w_D|o7<rk]UV&jE #vb0'OCK>ˮ.q{,|ODk4ɁDY))$İVcx6o̷:Pg/&Q:Z,hٻheC$"CߏXXNaC80$䚊hXL]Չ gNꏈR222W%$0J`t.x~Ш{9#è-oc:Qt SDv#AV.ru/>s!i Yɥ绂 ld 1c+љ ߌKCJ hHdN '14x!98Sn";^e8WIN -#^#yX{9E1>n x XDy6?I%͏i?kHd,hlNC?e8)ЏkSL0 cHz#s@9:|╓UY+gt\?5K)CQX~PH)pC3/2@Nl ^ L[%nvЪ᤺rК+&_]_V~B_޽yˊ9-~pm_(1/pyVըVdH5 3DMs%V~Yg26O2;C0=ab%F8iie3 ":cY])]YoG+rΨ~N'2lAlJQ$CR^"ϩnjRT7t֩է$¸38<74yU gcgJV@ƅâ/Kw8u;O!*s~d2b"JjuxLora :7,ѝۭn; bV*ؐ bfF W޵x@7AvkhLAE"la\Iel{+|dM^LRpޛscVo~/{5hKLP-Ĉz sU^=eqz=sBl_֪mM\3y`PEV h8F%rVN*uTނ|{d5,k3VlNnY"p>S'#h&_ZBu~!!` G7Eߟ s(%[2ߒ܅^_mL~w#Mې[NfT n207in4!ZQ ׁUqH ‣B2jMl ".R"k%ZV2|DZT~w\ݑ0I .{ ٞr%FIMDK0kԽuT5>ϛ?'9\a4V`f˴6ƚH&vU:_-3!OǓ(ŗ5m%fägäh[Zp>a;wa[. y8uqp;^"i+)8~4~dEW)eOӞf~W,5] Gqo| \P[m>s&`z6rz D^fqE@YędXp*r>y[^rrFsuf{PbD]ixbljѝ5m#>.rT;oҨ+t=BA*? 5 %O۞`\k|O.)BS2*@˔(j%lWIV3` &y}ř5mEߪmMfds[DO0Fan henvrՀlʉjOS"Ƨ]pZmY{w< `Sq>{G['3,ɘ<2e l1[57+Z h96X IE3̔I;%YA)Ͳ5mΜV>JLA'^ެş[}.ܝ-l^/a4'p9 QѬ-Jv>fGShHwY CM.|+m\ЌnƞP֚/ WVVK{wŒ*I؀J1[°o@3iw3$ӓou5)u=Gт/fkڸnuڽUlQ98yn^8ȭ~70Vao@6Cσ.l\>}C ;^Ē΅&=ӦE.iQc>QB O0|I!m* Nn\z?Pfy3한;8eN퍆.Җ>tAwcn0:ܹ<|$\?|=TC@%y𙁇zOB29TI/) `70`wkoZN##VM Xxip$? : t!G}"aOԤu*A3j{_jڃ$BJ i6~O??fsx1geB?<S_hO|w*sUkF<:zokz;]~/ޱ؋W?2}ћ_/﮽wgކ/Q8Ɨ|ruu>N[='.S;*XrxY]f62?4ZbQLNaKN9Hwczh04'->8歔\6E_\?i7% HinKFb xzpxПS{yU'ތ,?[xrfqde'[̼ׯ؃k/_n,tNpyl_ qs5e tKxuz0:EKpru-x)Xx3kH :'Űp҉fhl'9XG=ڏ/_^~޼@XO/1x:_|9&ze.$gineyA`0` t$_ǓPWcWa4^;|WɁᾤd_t/C4DN졒χh MoG`20I ^LLeHRye8aU7T|\|, 84.#N/EO}XĎ A!;6c< )KiF!̳ȥ \K~B0a&mt8g3>ˇC8Lqx~+X3򪔌|[DN'K}h!Lw-Y&H2fbJe>łrv6ODiyh2Kq\g{?"po/Y{F2mWCxNHHp8! ʄ(rXJ@ QYXa?v*$8ѝfNp^Jnu)`^0`uC'dp<[8pr8Nk9WrA d{ȋx5X4BPiQE4^pRD;?i`'@LM4lP|\; _y%4T- d7^_>^-ɇ;g'QA b#X}>%R3HOj1";oCݹ4ӻInzG},{Q7kKdd*=y6S?eO>r4Bm;ل֟ialI>N$u:uf;`D\ 2ǐ$xQ Vg9[c#8UPD.&l D22'J*edu@5<@~D7h A w48)!F"IFN"x{rY?ZNibD1TT{ZMAqIR!Tj1Q#(-z"u;@ &X)4HP97YCb Q YV5s"q`[8AgZ+ ;IDuC.8$!q! H"dkX{dh Q F.J,j55p8yܚ~jӷ<Ԙ8^5h"]Kc:;Au4Lq\i}8ߝ>qȦҝQNa#TgbZA_W/g*7WHKTit>!UG޿؃ h4LlPLό'O3 D*ڀkrt?0@C{0Qq\RHϗ?ʚB#h2cGӀ i*o ^mCH˨H:#cHÜf3ᮩ&HJUoZJ呶HMRgi4#@&l20"=턆ܐoDϖR]p4LV[L 6r뱄{QTJN`n B`u"t)޵u#"e1;6 ,0I7_cɽ)>$p L˜|S%:H=GKMC pn5HԚj4J5+r+8ucCa`\)GR}k yiHo$q-[@ñm Q ~4r ٕWV&R#Y=/ r&]w8q6Z]:︻+_fj:eBN5hNDѾ22%7h1.~n?_ [+7(5.."dzON/6_?6i\߅>:@TB)C䱽dʨPaJ.Mbkˋryͭrh}}b$Rdkd Jh`-cykbN! ETQ `$ghǒ)LOfA1s;eQHߋ"4j{~},ޭ4o7b{nQ<گܻp6sKqf]4 \y`W(PfF= pnmP0\0bȰ2@1ojH->t)yTaY"FZW0dm 4Dc{i4aQudP)L9>*+4LC IJ#5RFC]=xOR':ćk$pf9%y@ImFd/D`L۴UqyC劕 ⊺.:V# .&MJi͑fA ki9OF $= 8**X3\PPj#HVL+,6K1|qlQԥ^R/|E:倒!(dC%SI 3yj7.1L|%1"J+dZh$h:{[j LBQdϲ_2+,WV ;N1&d;k8&kxAkYyM(Ɏwܗ xF!7Q334]Iv@%LX{X^u5 OjXBGa/]әº[[ /BlV-<\\.i=;b.fh M5l3fG;L:;{ RUhbQû8vr42lpwN0R1; ~*X!abv)vypAF'FڰQP! !m0s=.9<\,9Z}sEF$GY.DKY\>˴VdvC4*Oa ݠ![СdeeY8{K4,#R1Uא'R'ДV#%P748~q5rQȷ 4ͬ/bq[y4 +,w284MD_\_@s֞hyL ,!_xٜw[6 r&"^vѸ 1ϋ`EFK;Sr0.2Ib\ޓ7T5 h4b@س]u͇m0 :3@t]@B4 ~?.B{; 9{.n;$+Q;ޱ`m9 .Y=U{;- g ejtOVYrU1(B|2b:SXO)4磰yԙº [ /BǬi_47gbمQK&]},d{`TfU3i=mc"y z&d;{#TQwcz 0 iAdaf4&ܲabS O X4v}N_fSzUwy͗?{T8_Dng@Bb!3)$8lӔ@'{&DaA#)& FO{mvu>5/$oW\=|k{BX&g/7Ó3^.c]-sUAΘ&N[>[dB-ȗeIi狺uQui=q.Q7$z>Fo[W3i=lcZf>FTG[nᴋ1Q]ʶ 1S].H9AY'GQ6g JC&;8nRHpx=YM bj9#&?%>f%PZ\&r;3wC:'-(AxQƮVe-§gkzqZYIןIҍmiPzV~\in`ߊo?HLS#^_Nɓ/џѻ_^ۭ>Ň ?sqey9|Gg? v?OYL JyƮȸ8>ܢӧ zVX:[O+t݆?T?tn3AZQm]ېۿ@57w/$~MYzWYQ3]V@߼"dz.},(Y0_ Wޟ 2ꅾ#YJ痟гm )P}]u5X;>wקλUPÁ"tQed)|㻟i+-3&EbI"(㉳6H j빴;\ # H=]uxхv ]{Sgcձxw[O*}7d f>[㪹n5[DlOPu;jk#jjKju82z Q{J El]#PTUuйP*ó)=Ɇƹ$u&cRƾ`Fn\ޫey!ypfShb`mFdlag>AmGb ^8W٭4lD kHuw',n0yKG$6aG$w мERJC¾.BnHO4qs??wWGҗA'_>~URC $6 ++7k+Ԕ#g].> mãxypgwԿھU!˫w 0/%JfG>'͍ыER]Hs'(td|H{Ym++o/Ho佭y:߉xtVE͓_-dϚifE ̬:gbI!d|>z翞uHZEԎw|Mן!wzvɡ%rV>eUof*f]39\ǁdi7ce:e܉x<?͛*;-uzU{ovwpmx{mady1O0892<:5)G4)jz$HQ;T ^vnY?5cIS$xu٩EȢH1Fq8D|rBzDY6: KC !(òsGFk"`2qP̸~VfW pHMYfm!2)$ NrHJL L1XL%ONH9Q8 bQ`%? oIA 0<8@g4ܣS@hQiF]2yI lI'm@HFsP0~8m{LHbw}H_18,ßE~鹇 0O$R'm+^~8c9ERqr}dX@z3Y`sXV)+q둉%$/>YX̦sݲHZvY0lݕTظ"qKp5SXn(Q|Au]kUNXS4&7'I;USwj0Ra/dgvVMn1%exb,g9Ksno&ArLJ67VA,ט-9w/aq6S Cb8{Ǚ\+!;cBd w0нRye~0sJQ)AgLpB5XXjoxNKΏdJh挑gTY]M 3mܟ-`gkE5_"m|ܗQ ?aQJvp"(=`/ⶢ\{t7.zqF".E?ZH Lpȡ˞vwy4k@IM4)deTy<鮚U18kFY-,0&Q!aCf⵲C4$dQ S3rwƎϖ]*9?CA w+BbKzdD.5n1we4/pNo#@H8\HȞ/ܝm&,4!D[c@$V &SOF'mx[`I>޽^ѿЕFUa:3;R@dD`VrlhVWyU@M~NgF˅dcGgOVxgK-gnR,l*ߟVr~.Wr~bz0̉4 P*QX *3 %Xib2H KFPj Ot'JO>.wE]3i;I-򰜔 [<,JiWb5礸IMԴe ֝Tդ>,I@4MZI5SM3͊{PcKܨkʁ6u&Ǡ,A۷3vUǚD>N_\Oo>r7\^X%U+6D->9d8Za!'>g](ݚ'!IH#-PʃSt EUH~KC0Ow)Ƶj20L+R"(؁-,ZE\k* 3B$KFJcGiZ|9OFKtKAY)\PCr9%\ܧ!sA`VX8s%OHuh3d0G{$A>*{mWVb]eCRAs9QP)gnWzwW, 71oF@>OrYٰ}]o(\mYyPm]yA7ڽBl4b.Žw-ny6`=l)z@w%C_>[ ́GQ9]NA\1|>]Fw˒o` &V 眑q ҂%(%͓,ˏ>ٝʋ""X]CQO&`h58w^$%<3 aSq~4h=\D|h ;\$<_WO^1L4JT8Ofk1k@P7C!Mxpz1"F7JW s{A^Iy|y.Ábћo4@龱@_sQ y/ܗ!JE¹x/C*^g/xjNy/s]``NbϕkQijVkM#ճ d!el1NUa ,9_DA%D19e8tRm%3%j)s٬'͆}{睋{ECETȟ_t&2"7>1Ω00*pkIrDgHȌ"B˜\L0R+gWjn8kRR3vm_$Eىfjl;5\{UStZۏwm>FӒPǃ Տt4tI]MV~*[ZE\5pժ=QA*ٗ 8{=Y/bϜa2q2ݭ1ApߑtQqH*05R(Y JKŽy/d6FZAdH%"57j*>قꂫXZ@ X;Gh>/;~8]ᢘʸW8 1n9.ijvrOT 4u?8k7@t|{ߺ-[\n@NrsXe X'AC+ Ԑ$ܪt8o#M\w^AeNW$`9RI)IF2xb#ҢH7mŪ zЪNuŘ8jK>گhPeSvi@hTvK^<_^^&f!ZO[*z/> VW& m~KA=i҉.`VFˌ{?ybuҲSbOOكO4VJDn~W,WCszkݬ(0C_L&&t,6ʿ="?\oߙ%Ml߹|)!;@^'NT$Q%Vj-)_}}A*,dgϵzs%UNa=]keS8Bp*Ya( R,_M ikU1.͍To'y▹ݔ~>u͞X2;t8dDЈCngR$(Mu 脾'[v0+>\r=,=ʍͤ,3忶P H|YGPKvM+}јCn1yʕN~}ylkp<RܧM x#W >or%j\Zi5hQ4svзi+wO ̃iawAJs&wYX|&j8%f:RUrIY֡[1>k<PSu]C  :۞3l/y5ESi CU?/۟z{Q3@BU(1[RdL]/_,{5L}?$"w#瓸[W J`!tDkn$S# ) T\Ȱ@9qlN C8Sb<%ߜl*A7b,(ur*4}nYaA>oqeb)*4Mk)n ?8DGgSB0Y¾P0U`0N[]ͅ)Hx1R+7T=Ϙ(1QT*ܚMtgQTMƋ\9s7gwiZ7$?lv?7r+O[Ji㝱p+TйiԶ T'(%jdNU.[W bԧ r/XV NNm/j\[:ܳU%+ܶ("gIfoy.-FJS8;NR "9QYs!Cʑ2G88g(u>S;B:E_9w c8oL^ ;Jem]'+d:P"GF8.)&]ţ]y%:U )246SaءRAg 8hvH;[J6 tC$v7>g4?衾w}%F wvn^Էs Bsg-X9)͝U7#8_Y!))!I$ $9Q9՘Shne^~O{ݪϒί+N~H儻[[5n陉]͖%9'ōLmZL'oNFtsP,3of' )cћ W:?=ZM_1~ӿڋνFjV V R"8C x2dhi>RsPDfZnI)'`qFE/n|10v668_h~%LmVwK#q$Y*$#ӸYKz1`WV{\労i+*WW 5z|+""`JPCD^!S4*j0U 'yz0-=gS(,lƴ_“MW<\jƸjPSneVl9*$`ƧEZz#t0cuD'( ƭ:`!chB+% Q"3ΩF % h ay\vrW n쀉u#训S9SgdjΏnX }qkn$f_+zG]\uը,ף*ؙZ9x7 Ս*G6u:X}uW+݃Jɾ'\AS0㯣N-,bʇY,>]ۗ%x5~4Zc`f *W-lвBh d֦B }Nt[ׇE4_l7]1IeG3 gy D.Uy4:`M# _ָdZN5C&’rÈ7bzqeeB[xtY)ε0訁 "9cֶNҚE)&u/0kf\"rar]N]N]tL.kXS}Z2vjpi3qNF'*˓/DV6WcED: D,4 4UNcY ϐRo %6%VZ.J8p )B3`/mƊT@QZʺSUTuixU̕|?|?LL{5r7->xy0Ń)h޹i;+TsuRM6h43j-S5@!ZG&s)0`t5䜢[WjXc.*RIV_smYl -Z$7(9V5-B'HmUIDEs2b"<3)<`hа@uV4e- :0T@30(=\tX xŪAv"c;QBI2Ҫ4ubsy 4iX _@0)X8Ŋ rhJqU&r:ot*xry/`2Gr+zL Ͳ7Igo}_Tl%GUa  ȗ{!cвoƠUޡpt7UGfoB$V;29oDwT%;*vFwnynU]]^8H#8N ڜ{+x4^{~ ,Ogo{;b3x܊}=ø\V\QBvj:FV:IPKAݛacFO2Fat1f?$849~{m_py*!bM3!aQ[ϢtjB S6#(`;l)ЊUrfrKiD ,%F4)ݥDK(U ch8P13AN1|)OZ :UtV)D RsD9H% NtN5N-+abF4yNh ITɱB+A j)OEIپI[dgj@Hyh R{ 18;z)B(A0-zB|F+f6X7Q"oٺ[8zФ|R@~*ozAOO@eL% O? tx '&$׽`H_  jqCCq|Hhl0<یXfL 6  +\ELtI0ل")8BMm3}0\?U bjƙnTbbmj#ssI[*R vX L`  WS`㖇*`L5Qe&C׈,(B-0MץS1l3[x5ٮ8#+O6|s՛*9#6SyΏ:5@Ȭ)ݔkAU< JU1ޮ.J * ]u1'a$31-tXqCႳ:G+͚uBu g5^>{ , -+Ɯ4'K]%;(bE\dpV|nȧJ[ᦔ ]/P YxۛOV2@Uô}fvUM$][D:w?f=3nG.q_Ӄ_Zjq):RdTTK<ХDrp 1)Ǔ?Sζ M1 pu79"}:@+bk)m>CvMDJҦٞcqƷ&Gr'!oʼnyVoIa& N*D #CDtT([ftb}P"H4ulِESϔ#an4Lʓϖ LHFŚ9,l~>3t</DN9QžXm.\:񖰾 |x[bjQq_,З a~1G"҉7E87=/fc&Rhi3_Ŋ,p,Ǚb>Q\Ux;$SDGZ N!uG <*.0WQf sC#9r*@[FD1è΅ w^\"W/{4>y&l5X=۫H$-RIUnɯ|ʏ ]Cl.V# AH ?`'%Ջ]ol!P#J/>gq DZeM,FXJ5;m/9}KsݔKXf鐜 u3sb૜ dG83g>+.w;|ηD8J%LxD uvL]xe9k0݊8y-V7 /f:<ιNXSݱeO' =x`Cu.a+Nx*j\I)vD'®T>ipުir\U5KmbDaA`0"uG^78Jv䒙[SVI7],^Lre;H[d,俿NYELmN` Sn$Bɵf])o3 ۉq p($ |S)*sF@ˬ&c}&by$y-ajRWIBpA騣؃krFQ^-ծ1:i5-Rs {рKc8%y׹0?pv}$|??0Û}ǘy~uOU =L׀@>VZUa~Rj.(U 3WA2_y)ȃ$H!tRL3gMHRU i&O /f}tZ8ss?!ngÅZtZDXh9U0 V∙ku$1" 4IG,^K5dl4bo|pR'e4-( FG1}>Srfz4itW7ی2bG[]9(m{5W r[ 3YouSik9p Q\Will [RfHn<cDI{ITw[D'UEl;e径6ؼ{߃j`fK_xw>@Zla֚z. K--ĝ=~t;>:C4Fa]pvR .Kt>(QJ-ݚZ9[ m4=w2B*t_rx\ֹkt/ǣh:}2pᣙ}Y)$28#& P8*&X;o_%vT[}{Fkqv?CU$AD>/$m! /M8-&0xZjdI%':hYR|R?mPuq:p}Y^]U< 淋wW'F~wv0coFy}//.99w~xǞasrQ?+ s-LVյvN\k |{5m'SxpAb(;z$1a/oy;1zeǷpjޠߊ>q6vj?=l8;,4Jp<~v=< BcK^Mcm%%Ф I kw_?:&muf,ctySShbbU3eaa̶HZmzqa")#d:Ê@}ZwH"k5;_0+0gzv=%[YU 9oI3_Hx[J!\Njga Kg G30HxW]THA6C*PxU䡠$y-ń9P&[#AK4 ڢ.6PkFmyꊣ6R3ⶄ(PMh{'Ȩ#u:ޤyQX[uxAy~0@$DY"``Ĉ 1؎)?l9d鮳BKenKD6mќs+b3NP ͪWoԋqg4|oJz|{s-E^Z\Oٹ>2 }^k'"4z?X oD.&/e- A)'sh^`'ouu&݉ 2<j3ЎEBf* R8AvZsߣ,SM[I)׹;@f:ĺ]ƿ,=8sO=w hq22W&ɦ]Ť녻JAC1[03I0O(5=k0B,n\#n<{`B3խj(',aaF]EUV잽E-q  UHeoX.}}3%smGW=S'*=Jo_#G3<HXu#zյNt=&*w@ h ^4n_A~tbqÞTy c8a \ 5qT8 5pc@!p`N~ݞx(|80h(aHK@7hSa$}ɓe%y~!- h(śƔS]jr{PH?wH6uDE+,"96X6`12䖄1h\Y^lfY/_CYD̲RBv),7:i[Z &}$ 7kP4I[]sMEk3&ܬyV ) M+a"[%A0SS6 O(6(cp}Uu÷Ca#b3mu 7|u8F ?{{XIX TR4מƤ,xCU\c9~Sm} ihz njB}BSJMg{Yg-Czh殒Ucw46^[^ӣkt;d?T \ ~ܻƌ 1FpY_oǎzԸnҒ!6_()^ {(_=?F1mƕcpl~\K63+U<;_ )giU4<+OmgSRu6g_K__oOJ\48\ʵŭK_oX)*ys4yR irǵ<65 [y5_yQ1ܡ<4jl"2_yT(K7e0hܺp1[(Kd/΃rAVq1n/ !L 7`ayk,iKlp^sJ瞿5"di>[ }"L+%Ϙ8!6kXA ,5݆QF`j"҂o}~:t̯ DHX1p燬TBQo2 8dpB03p(Ҝ/B:ҺIk\:2841)P8XGQM5R ծ2# 찛 DW!q;`l?n=>2K1>2jlz}!>U N-*[9U é $/O2q~]FUIX@=|(QI,؞Ǩs |B|A3ʤN W]ޟR=FH顔T: K=΅﻾ X=„ P <%hs!0BՖ+}xv]O,%=< ? +mH v5TJ<'2|lOy IَUPIQH35 &gտjM8 ˙< % `R _ xE I#O1e K,gU۷vBjN8jO\fYZjM`EoʣTh\sIm08"b"A6Zp(Tg=BBK<5_Z;DnMQAxTĀ3Ìan`T8H P֢ЈRcD3/I mikjŝ54H4zlQ,xukyͥiFԕ0аҳi7bcӓJ̵SC"rV7m|9c9\6 ![Rζ6m&K~IԒ*v xG!HTX*LGJ܀Dx+&(BKMȔ ܨ魖*rųLԄL0F%4xq)aUJe}% }Ga@MOVF7OLIL]Ǵ.֘K'ZOe |0x޽ YL5X-qT}YYThƎǎ-wa Kv<NmqX4KŗE)gk𿄿#QL|aͦqOz\Ȇ/ݟltq|5Gs12ř2bx4 ]Rw0Gv} YA&rԄd8.9sظx)=rVfIFlL+VޢZ߽zvjȺH&xRMT74ptf;)\;n~wpn~nfM+Ƶvky|=tSF祼6|7uN -t^IEKqR)ۘhdhj8M~V<~IkŰ?y>rF[#Y=m2LS`JAM3d o$s)uY{Mj]n Rb{ Z45'qk1eާ /J'+YmJO#m#^Ŕ<9`DɗՓTD 7$&F 9NXk6mNșP;#!3Ȃh$eڄ7a uX+ cMOS'0!C-gm OI?M]!IZ˚_na?۶*k;.yk4q&BU~sVu^F2'-ƾX+OkvSln~z RDߚ"][- wvy ȒAI c02u޹')29aʖa92wDGk>aJ僆eonjt>(hpt?绕;`792v"O1{;ń ݔy}(v4eJ|3&F_ܨ|ϛgLp)% `@mP5.ˁմ tu3/?;xd{PKERgC3p t|;/0h{y(U?sjt ?UwXSnỹ d. &ñ>|addrg142bĠnV~NJt8fu?t_88LN*N;?pM4mX|ƿȋa1Zw<=mhpO;~?߾y2dwGa ycgǿO'o~}W<\PU.>N"/|re>N"a|ag_S;*X*7a3>`Y5/I1 eRwʽH1Ҝ=<a^wRf]Oò eBp]2^(|pa_n4̾H-)4nY?>X]v/&* !2 T$_~2}qP僲_͊x&ҵcUMNa?w_z?_@jԁDU>)w]z?~H/ɼoI'2GyGcP`T CW^~ x ʥS?=/Άoo&u/8T?}'0,vo`:n&UKW}a:Zd8뢆ayiKgcV,<2e+껓?"$KR-rĕo/yss)up}A0u|v1Vc7/ ܰPT<(J-6: D?|q+Y3[F'VzBx~J踎H 0ejbֻ]j%P@F$Da૧[( skb2Fo@5g1޹yNZ-RI NGrn}Md$ `L+|ҀD%Eڻek[h`&Z!#"gJ;Έ7$8̚k[TS 71!*T)kqE֠yD.%Y&Vۨ("}ZҌ1N-u?jJaFjz6f"R3[Qmx.)BHj<>ɟVI`./.̼`,LT_^H瓙q(IuJj18a5W"5M߽( ANWTx!}.<}.8sa#j]O/rGQ덑ۏ>lZ]L+K5m[ߑ1Rq袋;I^Tb\ h ~_bBQ2 zkrGw1m\`d!}D Y*#a.&839GI\:[V*ŨU[ ʸH"V0/9 1HN+[Fvw2rDA4KDA4[ Z[* 'k"'he{ȍƕjۜ?L^41pI@HdKJN(HpiIc㌏IBPZb 8:܍=KM`tReLb)tfQ =Suܹ@FrIb( P`VTs(VK zS|f`\;I;I:Ub1?RTJwTʕHyy|g?"G/.qR Z j廷Y:v6zGټwOýN̋dƄWձu]@JF;g]0J lxc>n<$@z>JtJD5FTp`u/1%[.R)GBrڝ\k+ `NŴKHCD43ʚ:aL j+] M`bIU#(B+SS]p>Dm#R6!6E`ǤVUAdCcD|4Oq\0sybP2yP`A b4TP2)AAh M01s!G.2KgFV"ZfpH37A2D75Txl6x-sl"gx,u0 bH+- u2.%(k`ؽ'xKbw0Lo%fkG!5 /лZ p).F"*1$K@6{9ZtOǯ30fڟdl#ʄI9niJ'!/X-8rWJM\7_qV[m堸X;,"3B,["scR[@(v61O>F*ig  ikM<:PS)(< ,81'?:[ Oso1I Wts#h<:~p(JL s&[89Lvă:]5Ga8& ~~9u)͸., t]2% ]P`"cN0G1Ț+ȨEg#~&!*p*+}+4_&{?5.gno NrP֠.g[n_5/cҊ9?Ey?BG r-g.b $& *=Pe-RK%(;TGy_Jh0vnZE쁪ay0c [-k7->Lbv~ @ )iqgo1QAnO&Y!?TѨNL0fM:6Ub:{qRtv$ڍ! >DCMp;9ŭpݮ_nHZn㿻 Ԯ=1o +y:noW'a} yVO$"ܘ(FC݈9kmkq`CS4hߛݠ7=foׂ&xv]i;?gtb۬g'2m-<ɖgd"SdI y&ߎR#ˎΖ!0l`FN~dDwPx~b3VtaSz08n5) ěѶ(QNzzqE鈥,'-x ޯ n暚 vVtJcnrt:TKUBoeP $2~X"S2 {7(׌*n]-@HY#%Ϋ,zXTrg6$˪݇t8oQ##JaT N  )*k+x!t(mŲы1 dc1T"󵹾Dz rITWHXc-GL Ȧ!8Q4t e9>TSrtぉH*$ӼIJKtDRM8͂3G26;1E/^ -Jfqɂ;UP>BF (2* )4^A\(J*4c\SMKIf)qEXd BIօA8F OsL13X U ^b2) ̌HF¸Im08pqLYE9P0g/wBWʀ%c|Fc>~z'kwoºx$oR}'2GjV&py% a#élSt"+U9:=":ӎN4#~ ϴNqlx)ep2y#ehtT4.s,,Z2(E[u0T+,q,#9MqA"gjBLI9D4RQYfcv.*cd'Zd0iIf:^C6 {ᥘHm0O0 |OE>> dxZ:Z+wxvR\~V,nͧoI:T|"lkzӑ5?Eq[3\`tf0 XY~\ڟ>|y,ň8)V OnU=,e0nc/aꧨ?c0\=E.smj[(|?р;H<Y?,~ M'`>#م!*Ј>$PN1[E]O(CR+BE<ʢBaK&37L AtJtOEՖxʍVjh@.M3]`@ltTLiLgBrVh'g9,j>D/*%Rs&xHSH +!)hvTPc֊"l:QiL-\ A S;Ω*p9]HVxϙNKk$Lh넖9nd$u;s:%P- $ha.K2`U,q՜J(;!GFI}w<} Q:3//0kc#Pec#ڦcSIͧ0Z#˩TNc~51Sb[¼ivKBigzH#411L[Sفo)i{\ mXƻaՈi+%!|^ipkN&Җ! .gWJ:$alr`mw9eUDa=RHZ?aDH۪0"$=av5 8jԈVhtXZ#gC#Ujb뉹e<5Nil$5Cl0'a֣I'Z3U-455֚ t:Z_tp-1xZ !R H-w |q2 QCu agJkg5(`Ɯqwi88a.*L3[8k!$PQZg5J(KbfC%<pg |j-%[y^4lG:ѵDZS:ێ{Ú+uN#:Ş`R23Zt~fsQxJɭ eJJ"BAQ^{@`ea(u4j\9[/*lne!/2$N~de=˅V^zSiXQ:d q T|F)4Rτv WJsJ\pѸޅx]Ay/Ek e0`z8CkCm)e &0CTRFB8=7ԣ!c.MU&b-op&(3h~ZSɥ>ԁhՄ_%DP`DW;!\%-'34L}ѕ\B"=x'j_p}oTt<Գl`lx'Tt4xFu5Lxo5Rc"XRq  ݪE( jJ,)t#FRS 3k!&SaدzS`42#]+; HMXfPV/5&xp\!n_qW1_k_Z<*;Zl znOkCr<r7otg^7ɾ8y֧U9z5~Eժ=} *J߸X[ڔbNyW5UKYzpͲ)%'z7LHjLgnN7xE-QݲplSj`K*fOߞ*Q^ JL/=-wvCj>#A_M{₱9vƐL ʮRg! @T_T֎tv.~;dŽjB+uh`$7 8pZ25k+yUɲ spѡX[!܄' -D H 40eI]OGɘVqd1g`ON"T{ Dv''5x_jRCT3jCG HyDG QcAXK}>anq$fs;8%8HYji$scbV6źfNو9afiϫe)+d4F5Og$|\ ůO.$oqf80[$^r6,8^'=a|]m=-mgQޣgݞ #PFgP,EOm;.Ser>-81q6Jy5ՉڊDJz>Wqjɩ.ádl de/"|u?b-]zZgbqp_n>}Igln}Tzxy"DP8G ة#*kJ_:S2XyJe,.rCv:vȦɑԮO2!} ̓Q&$&LE`q!F;☲KF$eSЁ;!eJPRF3fYHǍOhɩ)b0f?o_<Oݚ\JPUtd/N0mM@h_GRoϿ|H-wO"*t>\_F=Z~ !hO:)aL9&.GZv7(|YynOEzYhuFU n>8}FtցݣNη`@jgHN8Q'g,8U1NcګnhHt= O?os4ݜq~$K% OL~|N@ e,{}{mb5|7{#1l1&!}9b˭j~R#Qt@ _Gmf>;x.Lm&55fQX_97,%=H}fkvRK [HuC h%#ȏu Ԓ{Ƒl /Aޡ0`\,bop \ȔBRNE&)jgDcqS}NwuuuuÑFUU/ `:"G!ʰ% R9/^K< 2~VOO;P0r2UɜE[23'CWSϳ3ZvID+VΞ}r[$,i ijjgG#:\Q"e#9s!;)Gܳ"dK"sDj"R8#Vj5iaE6U_ʟoJD4CDS BJZLr48x3Dśu[WUֺ}UuF+5Ś6.nm|5R:ҳ&¶ֿ9AQlJ}V̸y4UBEXiIr(.=|f?o&tqL{yf_tejxꉁOr1 Qemۅ6VtR0*̑&I$^!"Gn3DŞs̍Dj\Qj#\׭2 \Q4 |(1mpl\]DqѸvq.nA3ǝ@ŴꂫNR'EK%[Nj :$Vx\*i/\/XWE<]NTC:t!p=V=/h?&wp`@ށ8|`1Jsyԗ]Єc-rb ,/pB Nemô/|>4"݃/A3C(OYVe'!{\<1bW-*u^ԝ5;ty_sZyF$̹Um.~*Ox4 Yԏ+(̂ cY,##r--dԜEQJ0Q%iwкQ{Mh͝oP4b 4)ļ2M'05a&ќ AD_+ ޡbh''K10ۅ*֒3͒K$^|$pz;raWXGavRc!S,uNXe `NHÕ4 sE3"$wþI:8)1~kn>:/z=H IRa6X9݄Kn#_Gc%,'"qxJS-ԭ[WQlCXPQVbZ lxz8}xRsOEE(,$`蔒yi)V G9Db -U5P}DYsr;5j'˥:U}T{]NqKHX[rIL<3Caţ)X>NOkq=Sz61)!N04Ob99 7TJawǏG9j:oI,Ιx (-A ? d:[2-,0۾#DrJ$O_||7L-z55?VEqxeIqGXq|5[bx m~mi/3r.F{|ȩԂK R>Gt5 !JaO2[[=+X0~$z9U#u&o˽GMN%M_j@! OL>jD@X7܂.ܫ.{u9v1pa4~zf *W=X>3J-3=DE|9W|XdưOgso/5ᔊ8u A v,PhY!Ei! Ze 3j`l i4VZ#Y$BH)~l[B*طFnfc׬,=jq yo($ȧDh,$XV"`K K0SQh氲JG|J"#BZ؋(9jVDQOYK8 . w 0aXpY@Ki)}Ha0/}5}Rue[w5:z{̈s`& O#9P>IH PPFԇ5Ɔ Gᦟ{zr&X>bEWA`?+ W[L%'>o p 'EX0d1|AXN ;.--E{MeZrI7o%@)y>vo7QYxG!W5ȫ[f5޲r&GgB;H [CZdQT"*[Et=?\zAKOcû懟F8B MRߢGw5y(U ?SHJ 佛Yƞr'8tfB/gQ .hͨ !m^l9`0m"XJPK .. z=)g-ɼ9`r_TÇ`ɲj7UX -jͰR,Pc[Ci8- 78%wW4f^ȣ`v,黻4Y\P٧[VKldbF%W6%.,SSXjwHO-Ta&4P=oȶ$tv~Q S -z3Xӫ*ڔБQueWAH 4!r S`Poegm]r޲5_R!pðs,RН%\wZLRvݝ%7`vaY^Y xY鎤#f9B9됗jYt 狰Lou/J3TROZe mE0G'/KE$'4b6 J^ٗF/уDɮ*LKm+d1 DF } `IAsZ|b\_lxs7} Q*+)t D1?" F /v oHpL eK~ ӇY3}Pwŷa,9}FC 1ǯc7{g&>%yhٛg\`y1 KqK'BqW5OPZ ƴTQk`eMR. &VƉ MA!3C(-+ ,Y&"A2j9UbL VkzbGf 3"pr\-d(+վiks:Vػt.P%,]< wxKod$R07X PA9G$J#wN %PAM@;`Oygdޗ78r+MXM]Mf;1#''@kݯQOE\=oY^_5?S LFd_z^;?E>'U&񲿾c5Nw9b\r3\avw? yKt${Bs|B ёMB%B~ ^^p_0eaVGLs۸!R(s9҆!d,N(BdR-lrnRJbCOr:hօn;%N7hXwm<p*:>G*>0^ W@_B!z&Pc[L!Vk3Noc,EЏ>2q=2i{|#q5%S=)x/1=zq\f]>k* 4McQML%CIy,q;fEgh%Uw1;SeӼCeӬs3/h(S|FFuQ);ʚ_!ʷ7{PɼRΙ*k"֚;s.@$ Sb<N*bBN-j :]V1!U]3֟.8QZE5SwG{h0{fWuٺDUҔ0ĎhE n<`M7wshY*Qǣ ,6sis{ ua*YxG!Hf̴HkewkCy ɋ_Z!Y:btPfRh:2 U;lTncTw,6 ;x{Q?'hu1@ѧC<rHxO>W(̚wrNybǧĵSO:n :a0ኔۏ9ZeomO?:XP<:Xs,~k2ٙo83Έ~g&1Tqa-di,OqtĎ1fd!\/Uͅ癤+}LᑵiT)3 5hS.ot BK0B˘U`BgeVWl}Q+aFO7sktkT0#VaÉ,CGak>PfXp 5§2Z>s#vtN y ؑ>̥\rN`NE :+|%Q ZhاGN35k:ȵ{WsɳRјP*#X~T|uZWB 4Lj]6}SVJcОhM@b\1D`!%dDH#83+aCH_je!qeЀ,P6 =[O#Y9F[ٿ~*gJOI(+H\$4ծTR)J>>|8$IB/ْ> Y6^y\SLR7,X b*2ʌH9o>N7?[ln+ѢHKﰅu̖r-s-[ !mGΓ'o $ g[xcƝ:]uN /ի*6)ඳ4iݫ fUdžBn1[zToLdǝ_qrw;@$ի'*r*3?4"[ 7ZiBX]EFE2~wsxF.VZ<;籡a<ե ` $dTΞ`ނT[\}m8.<,tHs%FT;?yBaR%819ܺQ^IP7DBhI#ˆ*`?Xs,} <7ey|I* q9?C n̓a >&5GX0[0=X*xPc*LhQ'+w- ]ג,p,f:]҂bj%Zzowa[UZuxB2ײ%7*Sj{-JBp%K+Lhpɼ=6=M[>f1.bBEkY! F'Y[Ns%I3ptws 3?勡pݑ"Qd?;rӤ*L1?JDvϞ'Y8Ʒ8bYfi|JdKRc!E«& ֬)~-܀/Q>tx2Q*rVTC گ`9RJ/=OBk :x4j>)%3/5igRjTmWHi5z|\ڝ~YSSҼ%WJ"s >|hzM=1WKZSӏf%3霐*/B~UROiNZh5˻ԗkaͩ$+/ܫ\‰JxTCͧ02&&dC o+3DHN30ʌ'>չ{4Jm%!*BnZȈfK [naE9$m ]w_N+=5ʖVUҚwg_yN#텅uM^1-OcvMK0;\-kmۨ7Xۯ"CI.7FfEwdO͙aU߁{~&76Hi+BBeă 87 $ nE8DZ+e(Y[M0uY`'X`21uX%SAYx k3"e<|'bǯB\sH5F7c_h)qgS]\ ԝJɱWv{b@PvqJ_%J˼F~=q5fFY1i վJDX%"WD"!!)=-!E}JrFU뒶7W`<:^}0}sU"s3N&&s^(#)'A4W:%D@vkQݨN.ݮN`d\UcOw=]Ll(Ȥi2QN{IHt^8ߔ)Qn@Se#Gxנnk!z&Ol2⡫+3܀0%TQF@)TkeV"J/,x+a9F<^L-A%YFPs,%b3TE")PT8"TCҔǣ!٪״NB$nP)-㴊B2O`xd'SJ<8sȰ td +\S5 Βf *2F@J*|PF؎3RDS A9ioޭ|yt.ZNsttgεb-&K.3n *T(%CAs56Na6o|5>߄Se x{mZp⒥o٧Apw>~$?-d6+?h*5uMfCBf`~:NG J'=j.}8{ :男aXA0O5D5P|;" sIwKVB?᪓kWH(NX9_=힂m7 8t4 U<0<*#Bq4&0 u"^ú<ŨU[q0VPƭF .=+"kw q!XDHA`*+@'*0OTx}N`NU<6dzzɅ܄ͭ1%1/;JjSwa2VR?M* ƛunx&c3DT+bGovZG7)!e d.LyJtŎMT;h@% ΢kό!NΠ5E"CgƐn-huZ'?~쀊k[?⬫ChSwFD($'m;i%x9ܝJr:Zu]#Zz_,cu u%eyQ 􏺼ByҞ 4ɳɧjFqÜPUW\I#fz_4 3% rcc*er  S}.pF _CQ]@bbז7Epqqb H:x"Î#KYFa{ݷ$Zy4l xM_Ȍ]|a4{kIH.z%[367 }Qs6G0?3Lƫy9^ $݁'J$k\Va,92_8QQc"ۓ!kvZ䣅+q` $I_Ƿey"Gw<up$Ww2\`}z=o']FZ(%r%ִzᕴ9x-%|{ {w$~pm=T3p]{܂9GK< 4..z=x7zW9X|3z++CaiziLmu޶ȰeR%i wֹ.>:W(hLDfx5W|>V]+g 3~xC˜-|=~:W>OZCb<ٲ|X t(w~O2_~{g[oeyur;Nwc.C.l0z$m$Ҝw!]Pɵ0s-870ֲ?炙,7$ w伭CA%vk-Si[')w֞/ !Pܑη<sPiխbԕF.|vOk;eb|d{#nJ1%C[lmu 20]bq3 سdqinm i~2O[#A핪I4?ј$N:jD "k3Ϣ~>U^?_~LU~8)ԁT}yˀ!mA? _fgfB󃊆Ï͔SldicJgq qnQ)#L\{f+ DA\~\$mΔK~jX9|[K8_OaVUU+Tς5#BYG|`B`#gZ8"XJ6`/n$|"觬 -iENU4$ gY`PhW讇3̄d:$|TȜQ:ˬ,D\ LPdYEK?%sfԐNӝ\j5vesv dj!ĩI> %Eǭ`$Ih*&Ӻ5u:"g=A-9juIk9quK(ši-YcՉW\1!4W`7LZ߬V"Me5UY kCnVS"RЭgIEzLJV!dd#xN2, z` } ZCvJ$bdW:[TR>rL!0_KwǛO꾻tӣ,oA)M\34Sd$ޢedI" }S$@Y'djdNC%TӭXN3Y!q YC[fp>\<F(ALsdFm#7 I=2P e7L{w mMkMPN.MI߾$bҕ9.v&e~LXޯR&yj"J58fiq&|Fm _u@XVSM}L*9i?S";\]^$wkXP90 JH76Hs@6$Ɋ ﲼo؍ ;ro ţz"o\ꢦ',('Yzз輾M,\ НUk5KJeX ՀCbbi~\T.X!_AN@S<Ri_K7w8 Ҁ/v S\qB1Dq3qچgܭ4{m;>/qh-ڈk'Q{kN#7U,ŒhԌS|X{ (7ȆgS\ ãѷmڡnZ岎Vy4k\Sqj 4;J'go526dVɷX%_sT6d}V 1>VPv/jkd|C}֞!2%X&~ !2+/Z^4%&蒅v\=0\qqYvX<4zxv7Zc25b1e;h{DMݲ࣍KA7By%-wF̼*BNr<bwl`s_䵨&E \}*OGeZ@ҨO0TRC42ޅuNFO-™&ϩ "`v,4q13wVo~) ^>gÝe^ܶ4b̆g8f MOؔ!mƪ7'Xhkڡ~\R˅}'~XUoߊ;=Zލy|#?udI}&6G=was;֊!VYaL~ۡϏu8CȠ8sZw&U2ƕз Gg/D3^p&&뺓r'T#5jfN@[DI Fɒxfg$co|7dje7}S!m@!v#̟H>Bq~F70#ߜ~|sO@[7beV=O[xM@!?XJNB8Xz#S͎ss=^*Q:ck&L7}@p)ɟ[>:lGve)BV;7Pkh[]?A,Br`Y sAISP0٬dt)8 R46i!(b+kERNGa^sg/mːIvYDZjؐ3S7n4 B*Mh8f#u;Tr8-jIe([J%,5Rh6x1+YVogHgYn<[wSFbԥo"z\"w~(?ew5Nq1<,7zEb6?qE>BҍɲxQȨqeFBxNMfAir?yU躿%ys?/v -)˷sVס͉Ι\i$66EhЌ/i2r*(HF;*ۉdR6jzkߦ)9+ R(r9WM>w/=˫NLT~gʍϦl& mq=| U5'QF @:Š/G?$7~9w:-\xG.(eN_䃙?s'cC!#/drZǔ:'F]EKJ%Y3^X(FVsa~iJoݖ_-V̓9q/gܥl7koX2u5_)\71W[V^I 3'H3j[8LTB^SʫUō.%Z~H0Z9Riidו =0wTf"/p"KT :@ B{43>"׎B8wEK(hrdGkVZa岽o,Qwg, L@Yb_u v8̈́|*%~ú@ +5[a9aN*~|]*RNС>]T.k-,EdR9D ֙[r-#Jd~;BW%l7 +wZrI3c59XIm+KFy>B$o$iIA,:A"B(De1EXs爑 +4&c%s7':cA_cyp8P±#p`C81Iqمơڽ[:"2mxi0;z&IqEN)d1Y@裸^ ἇ,:ց0j%C C@@-_Q޻-{ђ`QM| U0|>^ NP~j%}z|s9mK:嬘ycEY1dݛWw|+~07g}3ϲ_A|ps?dO=o̓ͷGJR싮ϘS^^uyIWeOwkP[,\g̜wG$lzOK%NJ:̥g4(kZa<' ELUZ5ݴƣn1h$:JiUٟv[2eI;j a`3'\ \CF @bCEҬ-}߮vwﻓi %rif֏/"vНZ՜,<g#Q>V.h^(>f)9+e(mc,}4kn* Sɚm0Lfx;ND%$%7wf,.X<ؐFٝ^Zy:M!RfJ2Lhr`W"B MhSFNZt!KNYmC9rVkZ ƃ0r\$g%2-d9GVA'3hɜUݼnvlVdż8~F?fHZ$Q|8+byYt_|@jm IX&߼ !NX3vƴ줜TY.{eb S1+ 9E$YxGsXڡEް>6oU|+ omngkn+pQlhуi;5 6zc>Gp^ -Pz~>>9kRSǩ7?҈JD‡v𡳯PAvQ+=P *yD ;u]=8BVst 6Cs$=pG투lSůtk2+JAqU ĞG"!>mz܎L{~FHWZ3|$GI!CQ"^"(e1]GprJ)Rmx-Fzu._SE R֋ORjg^:Kbv6P_A +jFi.nK: / /Xc+)7 '8 ˹\.3o4R_m_lP4N=!{X mP8 Basf4tQߎ):H3U@CN6srd`BnhF%ݻGuJ|.?|.ע]۬mNQC OS7.MS`f)qfOc oGe3/go'Zn2.VTʴ xP $ȃtJ/*Rw?nJ_1avCa`>4L&ds`1 Eu|t&[|ȷd<tlY"Udkß8f\x,ڟp.9jijmjs S)"hCY5U,əfEd}`kS8AM6 mqЉM=cuq[˗EKh|;}9r^w@/y]ПBP;}q^᭱ R ĄzMΜ70BPH|u3_ч\gAg)HDQ5;OJJќ/VॢUʚ|Mx/u 8۲ T!?kk1R1 .Ẁ>Psy,@5EHzᷤW4D/-,nlZR0N>,(t<gVߍ#܉I`Szc??>?Iu~Gn#o: =.J)blj5:A?>t'l=O%%/!lB%^$si^W | z~{wo>ŧ,/ނ 7|.<.~[!p~i;!tvL~mr˷틟 ~?/TA#8||ǣo$}ӎ߻L'k=dݐ2sftt;?ڣ^$7c+;2)I}xq3[2leqgz0ǃa`g/t,н7qpY$nvB`L ({RfTI$bzw5RzڻSo ? GW @ug}{L[+3W+'͝p:ƦgVë哬daU b`%qqi=vˣsJ/u)v#Gr3[PE@DĀoк{ӋwoN}ٍ38pOe~"i2n ^'l%zn⋹i4m&%bs{ڇa:6y5 s@hFCn;%mw xľ4ot]/`~xOXiиlN7}GMڕwL?zw^ tr AK&t^|hN:Uդ޸$: a0>\}ofu"o@ɛ|'@ 4TpnUnpȻ\H?'gDIO[ȨI.㞜ܗtdr9wsRwz M$JڇzHR[~5r6֕CQ s6UiK(ZjP; *"&yP585ri>?I O2+\K6qܤyMI̘oL%"TLqY!dc,ny?@LX^{S0 x:s%rS<g\V)9A\v.7>b e-4_o46c y62绽$bWA-tkP 8z9u6!8΅w5< 3όAt%4ֵ~CJYsO~*3LŃ9n2n{UjGƻW08܅pjr7a9+t^v!FR0C"G0a-9E:0 !bkT3Ĩ:dT[sf#N9UPc7`Y+B2ʄ,b.r¬ J08;$_]}6~VΛr~ O'XcAJE9Lo$iH4|,GZuTfE+\*O5f t6LI:( X1+WU'#A&JYM gԠRM(J,49!:Lo)YdAh)BJ5aRHr"TacsdGr\BFq4y5(d b 㦂 %U*4A +Zp鑩;|W'=.&叹jbEGY+׊,?(v*ZмqNkˬS3CkP@ 7jА[( "f4F`D&Ǒ6A?ՌhfQ]:1T3|y<*0e25[&89Jܛj3xO ,4n~8  רH:5q6),h&nF~+ b| K</ܠ/eNi~^bo?Q}tWY!"*ÏltV"3IŔvQNL|(J**SP)ƣR@r,g~Z8)?l s@0s%܇*J XFio]pEmޟhvvvX쌯 B*p:(5EMl{;ӞpepLy<GFA{N#619}eJk@p />őmMKճy,U!qЉM׹8DbO.;+ʎ_Dse9$Now~sqg;NA{mYirD!71+ e)D%䜩Md.h{ǃ)9&/@$^>`V>ޝ, vO ts3J_ gvlklbM֟ LuuZ1T_jjgHTи~yZ|9*x)X--pbly]d)TXqP \s2vՀVqrAlZ9j[5=U]D+DJ[EU0lxj$jÐv -9 X4>Zq&[Z礗_"EG՝~4ROH 6XK0yTDVQX@V [FRi}= B/Bŀz<¿2 /ذ-v~+S)wKCߴxiALb΅J RH Gw YqBSq ﬥRVǍJ6Ztֿ6bXl/>JlYZl欠 ZñyY\XpZgѼbzN: a! O L` SBN : \R&QP3n\ic/Qȏ:̩$.ɋ y(C^ۿ^h% h CP Tƪ,PP`ahWiԄaTW eO1ǐ(d>J*"&KP(d5WсT8Xm 0)*&L,UV7Zİ&gʨy$ @lg(?|$ )ֽ{0lGzFB~튷򌌭Z2jmA-4a6;:I1n/M2lگOU'&r1]M]q^q<-Ty=D$bcFo!qgsej`#~A$'UáwnFD\$c|9ܞ};h/|f un]ƀ[161`҂=2eR["Fp1G\Gڐ>$v`Kve_n$5>Z 1ӣM@%lzqX{^Ʃ2:9-V=RfƓB(;9\q4\MKl ׈ErsX2DI,VVFLS D- `qh4 Lٻ6r,2=;-]/8A0It~m,#K'X:\lP)qtG%{;d֑\Y)U 90?;rD ĩ.Lb 3I B`7)Z5x晓ā% XA{Dr[) SaqcZRS(2b$w$YS"H(%k@.bdJiPLVWtP)y @ /(]NCz:/ďu[H,4̈́䇟.!3!B"Wws &Ys>Ss{w%P o^vnO n·`γ dލhp;h lPe0РHwtN!tD. ?7KT]Ik!(RnRLR10iOCya d"N (3A9 d$`=%n߅ϝ4aX(蠃.`<+%})d6^t}1lz7E&T%ekDzѡ]ԞOU y6`kKK+JA~ďuIpaͮs c l3hݢeu{e0"o탳*o5IfΎCfh*fR yleGyPv-٨[>$EC)r1$6S Zr/#eHkIS[̛9{(X6Wb5E0 6(jw8lVD>m> u)s&i_eeY0ޡ/>d:O[2~;-;t9C5XPzFBX8zR3T0MJأ7KrL;U^ Zl'{q*Ш9v챁\{_q"O1bƋ`$U@o³žԕwY? R ,aRygw (r9?uR߰ kֶSo]p:ux6<7g, іmoK\m@- F7A~6|7r!q Xrag|gB$oOÐ* Wx|(hiIM7x|x~H(ANx Qh"sօR{ܦ$K8pG,g\.@ya@o^Fҽ 3~o{޹vYݢcfp;~(u/| c}2]6My/ λeQvpr>0|~ג+`1=e7E\bX:eO?n:z}3b\p7/ .6X޿|ח\~W}0ǘ$n^-'\h{%M;dSQTLgtqJfhډqݹtȊw 썰5ο _[J^];M~fκY)lrWT͓pYC.^u=f7QbajO A~O 1c~:{0G˅ޛ8s?{?N8M#&U|:A@2d]Fh0t~A5/][]FMIfj -O:ccY[=kgmճz\[ FjUw?ĸؗ-5)S A<g:V}DPkYOD&X&1TH\lukybT$yPK^Qe溋߇C%yo8>"2!% >JRk5bIX#&Q 3!9tXl ^9 Z,P%D2F6w1n|y 3/rgνacgwL>->s- ̵6"oLB: =h55GlujU}v,m= objEs;x @0vPyM9sj:>{pbqXwcbqXxaë;^8:[=-˭sD7Sdj/7Φ+ )! Rs!|s߆,i{ k: \ ڦy|[ _u[4^sZDBPꌡ$FI#E".N` FPS"ZOrR&F=7{z#sT9-BkG&Lz4\1#rDzTZ)YJ<ǥ̅:KsmRF%ܞ RR\Q2Vi1PO3oHN#\R8caL @011H '8( ͷ2^Z͌oҰL‚SXYaPxuyE,=dqQbYT ,Mj6pd ,x8B޹ ;xl`}%NGI=P}*˫ekiD2ãRX_ڑcv]]9Uh?H ֘` yy BʤRxxIROsU ~kLJBߙ~O1G|dyAUM<1bEO{{)m4)+'Y1ZA4QHiZSzЛXQ?SG=EX/7`?>_žwB(z1rH!^ąjPb'72ihrŎRK'Vj#p .d*õ0Q1H ^{( 1S"Y ߼.DujTH{?u(F&x5 QV"5}*K~d1te]s{ATm{p ,:f9?pP>৿,U(g$+FYmBt撢~|a}@挑A~"_dTJ76*%)Y0s,X9sR1:u2[isZNj(LE`i̫.X .0x0zqt >X3%}#b.& BI2L1m-68e$ZaEJ2r\$9ljexr Έ`PSLLA!҇{z*A pkIJqa<8BGd4vf ԿzG?,,'I{cVB>gL 6S" 5B)RC$)R)Z \E*nYJJKz9rSYPShs4%IunzPFqhbR3*1=]FWaJ}^9N𱋇0xJuf<1jRbIu`v0)GXk&1vQG ؓ-cР JhR}I ƴ'oӺl3m(`#' 3"I:`i&gZrns(ۓr r dς}Pc{m+׆<W`ٮM\jh.hINsF2(ܻOI0d!]0NG73st˛f3c}u}+k͕//d*&2qZs n L}V~?zh(Nfᨷ7 ~ gJ Ꟙj3$'?ʹ6BC=.3Ҙƾ 72/<% )33]_n_.ޕJEerX#èMCm$XD_,9r繖T)%qgIٷ]lj8+d+zp5qwU kusGAAâ:ͷ%DHcƪԭQ8?`Ä_eA5TԖ` QiiVsFL׵":WYB2R. J)rA(zjT1g9vYOF~<]?Kj7dF8^7\l~(4[i2&W?Ct0mpWI:}$~'҅⌇<͊Eڦq`9cjU~broouzUL^]o OA5i<)In9ٌ+@l)9ʒg¾1sjBI)JiRD(&L#1{^v>33z)2- !؂stb֐S–&֖4r݈f=7 I3}qgfiu/KߨͨǾD*N +Us|rfC3yL6- OO8™:3V!ΈeL‡[:p|,2b;^*q\SḾDH:ʗ 5 r PTe4GgFhB m woWh\E2i30.(#6=>'$/=i`Ee߆IC/]ՄvWp >q_qw;Sd/:}+v#u20ߟE՝Du!(DVcsXtXSt2Xg&Iv jdw{ق)1i+%J7 ܬ |G39d'9Z< @@Q9YSqt9ĺct0%'Uչ`JB@Uu3\0w!mT>nA}N,3seí/›)2Bwʭ3 ۋN"缼O<U3Aj5nгnڄPF0E9sķP@YZ;ʲ⋀P1z)rE*X9_Farq2D̼g=fE_4;Y~;`N +͋biMM RWe|zxh_,גJ) axMW`}{9')iYMٹW~)뇿ڜ|)$!x _ʈV8WS1ʙVHEA B"LqRFMОt4bbg)JK9tYqy)RJjx,H߸s2/ 홌t+3hC*&f')Neʌ*e)/`LnFWɵB̬㴉S,jk:!1{nHqV$+P㬔AmtVLF^ R'RF(G ׺:8pWTz^WQمͿRr-VD㯑X*ɪ pqR=Ņ [#_vXf Lj쓍g4ZGꖳ*T_LJ3$c2_La3'U E8Xբ6HRJRcu8)YΈĻ@+DdЮ{{%T*SvJa~5\׃H2[-Ŵc * hdi"_/x4$Iza_ٸԸ$YgC2uxIٰђK%Wd\dHVT}t5v!FpCqG# d N=q-d" ֚:B7f=E:K E& &] uE^QoPʣ>$!'Gw@|=c7A4x1XBѾ`⃇g٤t,` 1.8iLhLjh\lPqtIl  ɣ8mO$`D9Gs8ف/>W̴>Fûa1LսM77(h6$ra[PJzMz`#_Ǜ7hxp'4ޟU hfu 0Ḿsi2…4$t;6jcIJJap$}ﻟ?\t[^hZ;i-=47출ʼ ʼM5lsc"e8 ,xg&U]'R"*ܘރ\Ty>Gc 4kJ-ge^jֳbQ渫ώ؎=M:;&W{kڻ0p=wiYQO$;%oWW%Tӊڏ7v=N|wÖ(#T1qۋ̈́^w6 @VnA\ 00}-%]]k긷cסcTҎ d#+vpţ q\iҢU1]0@9?Ƿz"NخQg|{@bY9r7z<.7jf[ȲЖ/ Ʈ5u<+qu"U$Akf( s )_ZN{@k16Az}QR׈kE랲oz8GWdS㜀A\eS/K :!WKj [1¥G3aP`D aN'螊wk.zlZO[ޏE)uA)u!QX A͂B!9at3.!3M3J0foZ%T } ;H;VjgٜZ2KK1uIew>/E\_z^nO(5Bt7*R a#%.5u,yTf">P!ql6MPla.Sg\U`.V`6;]2kXUck$殌^Ì)Uvxtm:ev"Tޭ; fnW-],Ʊ9́Z"e|9vV _eS)"5or^jZsNn|A}47&O, MQOJO}h\b(͏njG,XR# MxEɚhmwS3`"BٓLTb;gtJ-Y%0\ 릪0'uo$oj: 紎$t=; MWMWX ^ CEQ/Rb lnJᙼD%VԱYl~#`7y/uQ yE,4s'(c{ذFuT3*/+<H CK QI =W/\ƪ&`c? ݇d~~]]D|G&緮.|wǓpke&y_3]d_]|wϘk^p ,~4Q2ug5B%ckFLhms0[r;. E@K8?`)X% Wn˾U}>77xZx}?t*\t)"+(|}zP\և} @w.Sa]3HM??O?Gc{R.p `M'ng9p+N:0|Dvj Ma'\;a%n,%~YL#3h;+hɦ9ӫ^޹;?Ӛ~@{e{M'ygqOƲz=]1fzk!y6wps{b; LBZ̃Q x^j߯_|7z7`|/p4z=12NGfGFمǣ7q1%(s'(.?Sk2:MZx:/tS7y8/8 <3M&Ϣ8_m:gwo ƈT#6ѠWpcq|_=7Sy39Tr|z:\efu+.ǂ9il?BT˩= ޗKD]cC\U m ;-3v05mt ũQHYXJhw#\dfb0?dIЃ'f `0/@rꅺҘq{G.\持9׋ |~] * j#-H/uC+û MDćjOc9=-7f2Uufn|^9~4"jJE~?߳?vVm  mO-ڎi>ZqsK6َi6ҋ.'{MG 9õA'5>tv+ܸ?k.t1 ?jk5۸n'ɝQ1h"P\Wp-^Z=>L%hBcwߎV/ViGYo'ffN3=C+ ,#W޶6c/苫{S['J6DRl~m(ҐsblS3A@4|:8\]7|7=A ?vP"Y88o}=C=7Hh˞F\C7/~I׹rH1qhrȬ5η54 = `##=8/ד;oQ锣h=m\LWCs'_@,;6FdLqK`_3Xi (לGat)khg T\vhԋ3+Y6?/ZLIexR0P>$ `S%l&i;$ K3[1XX4sVF]T%(6LJkW`KXkBU{MпjAfϝ^=wj9=FQSFP׵ Ech5XᗯC.Cd#/To%Ҫ̢1K^ NG8>vFksoF7or^նhk ŭ{6œ7+m1W<4mC&Ȅ}(a_Bu=(H]4a86uМC6O=icr>L:oAgiP)e0J(x$Q,ط@' v'A\=\OFWIpG~{P:Mz]$+uW8Wh۰Zo\#GetZ7_+wo0IVyn!zhDbh wk~}ӥ}QEԆ6bF2`yG-w MK2#IBg']n6Noʫ)R>spr~lU/]}~DF|>=:T*2IռԲ'D1ڱFvkR99=ijeAj@kխkSoj'Aebkސױ61Kv# +k{jU 0}:r 8悅Z#٣zY۲n]$3&Pg A- |{uh8)yuMم?.ݓ/C+J;$m( ͼVrh.P.urv?Uڠw,'d:=ߪX- sמЍZ ȱkhL2>g{i30_^tOOJ) npZ<M F |IYgK'3<q`;mcv9p_-+"}kbݬ{ږhrql@p1u~)٨;2\RP_=@`!cKw$XU'EAOD[9\|^]6P3܆LcqMz 9:Ҝ[⬧Sy_#R+`GSgw5AM_cMBK5'x_H5tkh3 R/7>T`x - @cWYzԱh=yܥR(%SҥzS:l}l2RtMyRXؖDVevs3yC8`R%DId}ު|Y1Gs:!d LsR" ٵaV1|%6j@ֽR"-uӺ֨F+%>.\YLs1N$oI,E5'{Hyer[I IyrZY2=LEq;>ZCE:IYU$ِ "fإ̸X,>s@K6hIU4>DM`ٔs+ c NJRgKˇo*ZP de {Lt$YI 6Mc͑ M4,##(r%B{0T ۋsqdGd9bеЇѥl#hQZ,2+GfՠiVճT 9p6ŬzL̉Q,2!1xK&``xT7{[!gw 7~{uCgdLu֫#R]}ZȢ5Um#>x{ R>3-}7 [PI3LD&RD)~sOgij_L8|"&O$,MŢܪͿr2$fg_Q7iH+dҝ7_MfaL8, %3ᰞpiնM~L_LiB8q8M +&ӓ/ 1? <ԝl!lXJ_JI| )՟e>?X_qvB**=b Wg_d ,@CfQP;@*ZlH\$6&\*,#[=BID"8ke0 %%UTRMe24X_[jM~c呐fj:7{ݡcðj̽M>jK`OOC+=68 ?m;҈ 1zFԛoxD3.GB"Es Vľj^yX5=l[} 5^-蟩G`?^UIWCU^j-Cv|b7I!D/ܰpc (t!V*[l-gl8~I?Wr?St&_'nr4#&隔Kw>iE)aQGiQ&g>poj6N߽} !!^0) 2euҋ5H\bh@Z3@z!k!'|Ԑ\G@N##% )ZP*HDwH ^Osm3 d/~n' uזͫIRKO_}L4^YT'RU,:B[pƹR5ޤe1J5Wɡ @0LGb,sżH.Im%CʕI$" {YXGH3|Q?xOmxSg'Vhm?M͸ mOV#:6QL՟G_fmD+!lߣӭ .b4v;u̿ '4:Ji64j(kWsKVp|K8J +#VNZ;4]7M^7kzy u| 98#(EgD/+4v-yҪJ3tjĘ x,2:oq< )f,GeH0ds{"MN:Wx_-djW/$C"14")THRXY 2Q"Nږf NѬ;mQ֮u$!$"6pPacvyK7*V q8˘hۤƎ&/̎mO -c֫'9*kNG81$ȁ6^s"_f͆w<ە[8ӰՈRmnaAhICotHҽcFFE&yFX&_ aZ:1.8"p1cr:!>=4y-TJ}ֽZ`'˱B,ФMŅ ~X"/.\`Z$b&ՃHR|c [TWP44ZܲzPp\PQJ΅TOA+6*(t5ʑ92@FC9[WZl2T.6 yW/k>0.RdYjh=7Wmp='L ˣTRUs+y”Jcg &xCclaz{~w 9Oa\;@`,F1D̶ zWwV;l{_OȁkU ` T/a]rq0 6$(i+aE^9 (GΧ{w=IDhbKN1%50F-H֚%ua| $)XfZ& '>ga$RO`rx+äUdQs+ZA>ZTh(1:x<`mi#[C w:P`*T0LLŐ;M.ttI?H J!vdC`)Q0]{^ͽIRNP)IUZXe ͒#t\4{`;B5DiA!xR eEJFc/JBbmwܦB.$n jX"1V4rlzzx±\cAbWu'Րh^k\2/j#Z\)Bi㜓C)B,Y[->?Gsef^T俗H}+>}Nyx)?t>K]Rԅޅ n~q]C@?M8 D,bl"\EgUo4;w'Oc/zwq'r+󰸛քa6gf[,y6 B }% ԁ{RuC8T;# '|:NðA&v;׏l= zjZbrr1K_4nKgyrh/gMYgM 83I $a_dPy @^H9 J,u+rD ] h?ĸXd!,0!ǚ"̨IM썉Q"pE,H"8%.%.Bt{*]QT CJ&SeFkO3g)%[޻Hdlڊnֿs~6S­v6S3귚rl/]ϫ[pn{^]Ȼ27osj@|tB5:Kw=Ej<%.#J$Kwc²MEͩqT!0>`1kگn@zի?ҹ9 ?=lIB'_ޝ_VÛ $]LHT"cRpk­U V)Zí94ʭVh1'v\ֆ|E8a$_ %DõJ>0eMg%qyE3`S^{otM.nrf[/os}T9_7}@ PE`^`jq vr!:`I#lr@XpjF;Zp*IC.ܵEuT.:Aq=1l:+yr^:ox)z@d&mx ) /x0Loƺ~Q/KOxDŽlRgcgaB^R}|!kυNhxQN{am^{e =jۂ )Ƌu7ӳc|ѣf2(Ƶ>sƓїg3ਦDұ{/stFw&N!S*9jJւI9WMݜ{CՎw<9ϨyKL]S>y"+!TϨf /\J>3pϜ|tBb& RxY9,xgAۯ>k,*@.Ȉ)tCÆ[dڱPϋ:YWY'Bp5l/jd}K~Ibd-X<')buX`7SAYG4'N$;fE܃(r,9-ɯ^ G.ۙjZ)݁nGՑ}Ial0"B9&{."R#( $$cÀINg}rSK xgl+!DSC!!ӆ sƑli@ JH (%b"0%Rtv%L|BhF9`P©' BȐ*`ADnA~C`T6IЄY{e'˶η- fPHn'bXE8n`"p WB u3ƧzqZyи<ńaQq8;aTֽ~/݉Ԉ]D:?|_: +JtH>:і3r3嶄D.)C=xTʓ%Dq?߆B%,3f64~ؠQ\ 𸛻ڡqo0 %?v胚woԏdYC8 WI.'?~ s{{q>K/>A;0&\aG lrz`G^飍\࿈=BtB ]-;.J u 5A]P +c(U{Jj{`QM1XT!4i'a[7h*,bR3}(håkib$0ouQ8İ%dm/*Ps$kGDų +us(uU(]zSlhAH* gO)]/R~9c S17w|8:aD훷oRХ}ek#1TD0AsЉ6ר.?} 8Cv /[shk'e[an㖄M6ɔֺg⢔ʽWo,pS_R:*bYrw02BSDtS(R)P 6!ajzCӝժdYtQHəY>@] X8Tk8"g=jynn$Pu3В9"h̵GZdk-chH,1乒XpW OlD;s,RvE5\3*Kx98E3Ӝ}8r(Gp*|98è kǽM3(x͇cj2ॹ?CC粀iE:{Mٮx H\yusPnƆd'f-݈} cQV :a\#X71,~i,ʄLka«Gљ /zI5택KL :3>f> /5 1Q}OSQ+t`_dLȩ/TBJTq PX(<9გY/,̉}vzFV&UjeRV&UjeR[44:{7TcP ( ^sY ra6 /$N4B{+^[Ƭ8"(-0Z)[?TLΠEyn3{bAYIJ<8Oo0 d'ZܐrDA?IA ʂIbѾ !`[i:G,G]FR&(N&$BX*ѝLd} !M|\5%87ՅՃp\i`7!D]\/xM_pl!*׏~Fzs* Š/#_] h6_]sC z+˿0CB!|ߣhLި!7LQbLWx7Xj+ZwC  WooݠmԴ΄S6ӵbK"8+*v-B^U\|b k8-%˴55RGT;eœCJ {%' +ٳbC)al`!b`l V1XDFX[[[7("Zpb\ܻR@98T<uJWKBEZřPZ#J7(VqTcHŸ֢(C/4)mH7lw]y)2G(" aԺԭ]|xyu}og㑻^=:l*z_sI8Sdb96Q*9G]iہr3l$TR']^RD)g'u&E6xeԈC24ދвz wՊ:91p @LV%C:];OLt?V +$ eؕ ;: ʼ޲nc **~*:Rwٴ{\ 0 ٳQE Za4e d͠5v\/wjt1b%\)c|KIcSDLfVp/kb'{$ p?<\L-*/ZvMujфrVqOۙZXCa岒:|rg#s^:Ds"ÁJЛDO2P;Ƣ zTp͚ G΢VbFiR`SLmw$(>I#ƽ]؍= wZΙHH%mR!)1 V9eBp:A5V|n}f;@yc˥m'Xs15旒ʲ]ΩhjK'Y%~75tT2{Dt@;'zlW|\L&g>y4S{ {쒣,>P 5Nd ToVfPj4PZv F:yQ4Fr N5#gi⏾6Gvƣmͧl1w܏P h^Fv+On_4"\˫/3ۛxcńY࿫F\Ff:)=,uՋѣ,Fw1'-\c0#@Cgy3p6ύe.:zp3kI,.[-Gy2'5uJd]e6BLmxLwn\11 q3;i-(FUv ~;Iv-s4gy'zT[l|z_|}r2p+E:@#vtU|{ķa) "IJ|5 )sIQ_|y*0<F6B/h#Ԗ*"sfݐIY}{ok^{~ 6@9Sx␝JF-Z')@;n9w02TRrԁF{\4|1DI+-J[L(S̎W?-ww%r:9'J#},"7[o蹶GîH1ƋWnq{>.rUvȐ!x9g?k{p=Ms_|3BN?Y e}/vݶ9bԅ갣V< څ͹Ӂlŷ"7vb?-1fRl?~vQՎN jf1^_}r.֓;רHCܵ`{yYsYCécD3J&Gi)8Z`} g]]|?';guzy SV޿,q.<Ά*e.=[A^>m_wrmpji!llNҟe廖Qnke5kZR7;-DLJu;- m)/c I1NŴ 4Q㖢;xțrJh4 %GJsLo+SPiXm5J^Daז$"K68aTtNL yƏ^H&((PsV"Z.LE)\^npai$Bvlz珯ͯ {daZ[K#idƑNҢ-=R*ִNl/EVL @'L6!2C!آPi;g JsتGxj?7n!IRc-H.=^X蘷Y蓙r!!M%1Cyઑ3Vb[NyVN$(htMwN|rs\(v;v!lMFCEЉ8S:;Z&; 0W'q゛!ᨖx3y$:Ab̔ĩOQx1ONVt|*8`}"8s2򞃭)pUn╛ɐLZyB`(QyMTBhP6V MUW=d]*Y׭ 0l\mF{M Jg :9^#\?>'k +x!hwN9h'Yy 7RvlU7|6Z@U {L#9+&n~ c Lz9k +pٝ) ]No^)` yǞJ4Jbw }'nNɐP,N}6^^ʲjȮ]4{IƑr[p =]RpPIyNy_,˜Tcrړ3RkSqS/[? {̛4c̛y>A#ى^A5UMNJ_ ~ |,\~F Ύd\M3ԛb< O!oش%Rf5ZҮCzF˔M>8lhE'C ,^[)>t;,]Jy |d@˜al x3n+~Ԯ5] O]>n/; 4@03I8 mH*ExAa 4ѡ͚-oUr8ۖ45Bk=v>% mKߝwU/,=E Rghe?#yܗuY 43k!y,؝o'MR!p&ܙPH Dh#yx 0LZC[K*c;d&*-9PxI" qD_xNy((2`N42a1†–Y)elCEel{86\GQ!-m"B"2\ ۉ+P`@ N[00a}OS!jiXkc`Ӈ奶s"{| Xj|!U7SAl.׶4UX`_ #$Blf @ԡ֣kF+K*J)vV4 ^ -T:\Vªēp;mjc]b/&cӘB#*er{ޡ֊*J *BY2fu=C(9,ܵz&"Oƹ O$@ zz" m+w]5!ȭ AuU@cAPicNB)jWE-]T?*wu'TO%RUA <k ںxfuPj?]XK4ܭWZ A@Q>Xd Tqz_4}̕Sq+rm7Q/N7N r֐K'^I=ϞXL}g'RVH zV IF@kWlo^7U=n~^`g#(=0V[v:1 q[_>)]yeQeJȯL}k,aR浪Rך =S!)cH:ږtEi&}WVy8o[ AkI>)hsd\Í;f,>7ȱ]BXij8"з^7egߚf4B?%Y}T;q=r]^r#g2w=G]k}T!E":ʲ\ jX{j?ى{ c6\lpwI8G@>\A<,Sٷ58DP"X#vSb <])#Μ2 &俺I\6 }陋qOGlg[ >nɧ[uO|1?N @ -)x*Ogrtǹ! L TF>neyH|Qiab}av>ݑ_G^Zb2 <[V,Da*`{@'v(ûzzG{aXapާw'w 9;y5Nft|r~jJK|_.OҘ,"{vڳ`x|_xq_WWo߼}}jYrk<~vq?w|{_\qN^̫M_6nO f^ 3Z/jp<{N;Foikrc;p5ޝj85'Un?[7;ysW/NuFϿ:}mm2|6dl4`2R#'94c7NKW;+kmzlHjv>|KW٩uJ7&t6.ҷ,ZM-X^ enSwzE6=azF(tq(hыAlfޚf0J4'-̈́F!v!ȫp~_fj\TE6, 5Q`8 !ydggWּJ^(`tjfOHHlƲ^GُBAr=w=n:3+<_Oڱ` yz$}kөjw F4/myr3Nuff52+?߿}uNM[[ݢLusfHK d:r\C!59`5z3ArR42Nm 0;b8 k 0H ;YcAel6*Y+{_l?km:IC= 8B8qhju}@)!<~jakmg Țʽ>wx;b`ZY8lCZk0hǁ:\;rEc$e_ui)Kצ'9"`ǬfƕP)- vx0̪@NU-uB17SϜu]t:ݤmʽ >e \9J:Cߢ\ՎT笷1 dKTAHEU(UPR*Xı20&Q\ 0B)8,3 2rc9eڀ-o$fZdTU9f @æ/mJ=NzlR ik;` fHX]5J%9N-9:*ǵ.9x#Y,YrnXpKF?1z.)jjŚ=MaM铥iCɊSDc$:>F4"8"]EMɾdhɬFNxVYX%ߞAGVdcxhr>˒늰>8??Rzhe {PRjAc$.E_`3LYq) F=Kw>F"g{Z]=`gPl_R]=U)ٹw,M_'O <&0CCDVN>ŖX" Z<ϷtåB01T\!B̍LcH1(R ({V!1IVrC2veUZ$ĝhKsHw[ashq{)}nCHC H2!dvDϚs#nr;j} -n?:͐;ețd.=y݃_]33XܲK![g̶-"њ3{貋wȜƖZ~-IͰVGyκ %=3z Wj u/g&F-$4 -j|Fy5uF31QN'h*;Aݩkn}Z^ma阶+5;um-}mQJx8c 8Vu4"!2`U 8TX+RlB(W.m>9` V@J0޶4x[P"-.ʲSo߷Eܺ[/mvJSGfH0qh-7m?@!7E@,iØ(Ap L'N\#,eǯ8vw2q•Ny϶ܣ&5ʁ'ElG**BXʎGkZaSQLI/^ڻK#XBZJHPFA@˵@vj&M{Nqe j ϝt(&hsSaE!bF"$Dp@SCa*ZU&~̯>e{,X ]@0 #;C˛)OKdW'gY k|z?zW9D"Q"oixa⠻MD[ lb^XotiFIHL|7GgK{za-K)1s]"B-æ`}֠͹DZn lIC`#]&[@\f=N 5!O9Dq!)nGN^wz 9f=ahs.0ȱUFo[d1HxD D4$"k[GcAFes(rE{c%i8ST BU~7&C0Qsn'q2v/9Ђ P5gюkut*'ɚǶ>wܺc(*`##3vgx`jʮvtZFWrLIe{owrD@Gjc["$vd5BUd(6z񺸽ŴE3V3hb$YA$hMn:nT[ꖶjjSwPmPS*ѷ=6[D@^SQB}խru+Q&!ŪV!L8$T|eV  :z G[YaYQLH yuh.u6z)!ST\UfłP Zb, DL6X<ٻ6r,e[2#$݃nN=0xB˒#Xa*I% :XA>~, E:InQ*;cLj%OK"zdgȕQ1 ( s&`He 4VZw yˉO>f ΐgG't( 4im,σp队ԗ[f3.2S?]vbĩ/B^#u5+c3NPs5 Wj'%늫]Dwؚ^W%H!oTNFe,$Ӯ2 !9L*N10+ ~bJdZ^&' zuKpL-fS-Ϣ<`X,[bBOɦN@2r/<2FbRk*mByM- FCdb UԫuUo׳ mqFv&_O|a0FVm]R|G㇇/z"4ؔ$`).Y\[nEcNH8+SZ)Vl2_QRP,^y!l; smykiu:pSS (X)B!rrI@Ǜy߯n߈ Ff($lD.A pV"Nq{-0Fn-0X.khR'{(RPHǎ5i[N0\<6*+qyߦR;?g^K>{f}эάKT lQ'Ks22H Qlk)FfTe֩XN븏'@\ZP{GC|t7] MSPԭm2;UfU Z;,"3"WD B$(9-P!]Z/D#{s`ku-VOSP][=:ʕݒ} Dv%Jr K#}ˤ5 ]2]ex3k(O03x>TR`jn&^A `U8 SgWxz#Wƅ˕w"O~.p> |v?]]5 MՂ./!#)m0ao{(F*?NϧiG8v]5XS94r0,[='_RZ,`6FGRm!VRqueg $1D@s>JGQ|kE@ D__G30)wfdntuQ0x{@]űTdPїLhtJpTB w48)I; Q$|qp(cwcb.k?{'kh Sf]B)b? 1;c%w5=h*~%?μ .OMZ8 >%eJ-Q9n3Kߒ[X3,fO)k&0Ӎ3y4 X7jkFG ֽ*A.ىlͬ]üO4$ UY'#Z6!h>! Qu۫5|jvTTsLZZ|QZnڟ9Hv??]kV.iZE}ִ";fzkfnp|3](|]U[_l3q]rNr$/xvfӴeX$ t^-/FqTg9@t~Qn|IKUnHk'֯'-)naoX B yq p։hs<ބmz"qZ9ŧ.h!aP )VI$Ezu y#k0[et"! >UaV'kAOA v .?՞}~>bmLփJS1]$mɚPUki\Y1p[1/ha5:]`L j+Ƹ` |`N,'αN#&*ioõUK-"{&JOY(1|1_fK5$<ָ@x^AXtKJU ƾt^D cR`R %kˇӏ[}lw?0yfdLڱ^5VHe.ܭj?$!}"tp3+"SdR?LZ!Uꀐvm")щ%h@RήUƌƮL*;U^W4Tz괉 "9ŬhT\JrC+a>crciFq0ˉ6S|jkkPʺf`Vv~k`s܅2 qr/ 3b?r'ئbջ5X+XMz?pO*nv[tK~+ Ruu3ӀѪM \GPnuPI.>ݤ%ڳporp%7@Kh#KM!ukA4}Fvm)lWں5'ɴnmhwMtR=mV֭)9u%L7nN?uk^̔iА\EFuLtº5 Gu>u; ȔWmݚfZ64;WF:Eع/*$1ok4ok4m>F-[ sIUH#rܯ͙Żgꇷ<-Ohu&hLP. 8Q8Z=rOC,q0cˬVPU</_%d G 4x_lq\!-3^Z@bf s3rVKG*<9͖<)8Zс߃l, ]\Jߛ" e;?GT=(Y|1ʪRe3:/j|Sl$ ,c h6ٻzwnډ]>t);Ϡg8e/ wT0y([ ´r 6JQTw3Q9x&oe˨~VV Qnηmdj[g\Xbvk[z:c[ #L1-%j5LUOUxݨ̐9@dWiLJXx NY =2pB1ZJ*$B2!2~$Rn)%nfhF.>D3Ί:8Jl"0'S}FXX, kN˲ĸЃ v(}_q# . `:"G!ʰ-eېlԀC./%DkɃÜmP@*c4ADZ{$:f+`.@ʱ}*vEQynsEzyN X;)Fa鸍10Bc#` xցo6|U&*Y~=.sZdUIgxr[\t_M{3b&Á\>2M.Ͼ{`VGq9#`DGɵJQPBg4E ~9 c OBJKH[ZrcZv-Rv$ ISK߼u0"M9FCLc&hh*Wr7iW7@-k/X5B%.8@,f8F/pa8D)Ux[L fšLP("Eh#% yM0w-B>Xh 1>Ab$SaS L{q; f _+.u"'o3, G.=?_{ _)ALJvh^I)}|G㇇/+{O0ݴl$ <> Z,,QoK2Ζ4/ä]}E9D`V]Da C3NO ՑS-lO:EohmjH4 $} !bޟfn?f7F䙘(\L}pU:sps%%$R-DhHAHF.(H[BTǝ1i>ZJ,N?!GM6e#c"Vj588N -"gB~psG&S@B3H pxa8Ag_Lj!M59{ A+Bkm9~ `Ep7XYrI٦,JjQcmQdWU꯬7[BTG{:3 Wd?aM%n_L [3.9 '@9 2%H9… CJ0.+RtxIh_Ӳ='о&ḿaCIlǗDgĞs{OՂ@;̰1 Grcs10! X A0͘crMV7w;F|Qq/f+ia=0mun{QSA:nD Bt4db 2;GOfy  HVGP̴s.!)f㶌&gyOFhicwf]bS*jTrD="ީ{W)h{hHڎw*V!C*ԄS'lpbvO ,1m(/Q/dٵ+W#|G%OnSlHԴ DHER̳FCm0P0A-䥽f# ❎i޽Ӗ=(L1WCwf<*m^J uX%}mc}Yg4\YnAV0]cΈ*'KIV:i<ܛ)y@0z2K3Pcq*=x8{_N˧Kn>{)V+E'VDq*WEsG @>(gxVG>ảr}ҁ4ՉRlKcO : S%AVnYyX*E_<,ϲd,E1/^ $Et!.;Nj7JYu(!1F}Zd(e R zcF4<ؒ,DW28v4+A ^^(dJAEe+Bqߒt5$j$ll9c]COۭ,qT?8_(Zͳ9W4I o@UX#jSqUBO%GKRs-*qi˷#Вu8-B\q]`isBzl:}rĚL$ha:4m%'VՃfGz|LZ\ݻhF1%Y Ds~m"!rpʷ(OZa|*'@R9Z=fQ58נW_jr]}b&Q.]`a'XP=;?m;.3 @Cp먐i"O^oM¶: pXbv C=1/dD*sps:v{t@=jMv}wP #وE#6bF\c\`>s{mghSqiM =6D"Z:=j+˚;Sw{-Di=E "F剎3j4lk&:#3&T,ؠq"T:qI$wD9f83 -sau(ka1>ɰ1RKJڦ~/B.:!$,hp$ ^[B2ҭgk- t~[(*8>*/K읍O<6IpzkQE%o+TxXk[[@N){H-zc_~9]fzܴZ@=3cy ۱†"%OodR/)GWښz2cܴI:q4$͇çI' 'бӳEKn:Etrv<P'TmZ{CqJWgI3e!3֐$zI9gsbDKKU .ս/Wҋ8XC7nB\؏c ۊtg]X u1 |0LF` b4Wq50^^NFeID)x_rorjz!-ЇR$/*sOJ}v1źܗ :~ۭ͑IL BV` WG\o瘉ļ5e2jW W֠ geHvvv[<;±]qvs|| -㔙1c#K<yǤ'RL 4ZeHS)؀/7b8 cjq}z3Ysb `L+HoELjgW$+" 'ݶ҆ZiR[+}0A mOSK{=6p$ uG&8 zsaf2qނ7[D&'/sq;hdvмBRC_yQݽڣ)dqZj~OO\:5Aa +ij1njY!LFmEyJ 33<At\;EX{Y ا|\5#fX T5cmCpW Ls)X5:g$&Mjc4"XXbb`\*eBdjHNBC| M8+.\2d15vQUt w6OJx%%$P-D.b{8!s>A`5p>BP-9[Zh(>A@IB"Xi<#fVL)gx5DP1[|U9~!lx2O=i-n}̍kI!h I6LD8AM C=W(#s3mw QpBbiad^ׇ `5%8cLlϨDԥʦűI&% YaaL(sy),(LSS>P8[o\jA <_ }UgBZB% a @E)L;Ib3ÊӰN6O!,K0ƗHy0|YB;8?f ?ބ\_=rX]Wx|X\h,>@>|EE~s3jcDfB~~bX5i'r#<z+gb˿}Lwfk~ZMp}J60d5o".Fzg+n7Z%blpvo=/c\,FfSa[ QpTbVzq!"_X>IC1X<紝; 1H9fZv, *͂"d[)l"Xxa'Ldک;jEiT6AIu `zDky_ΜsBz]qw<%B˷ZՀZ"'\Y`7Dl1}`;rvOw4 t"һt$u<&w,^ӡMk:S bmF!ˢ=1zR/]2Ao_G1l"Ra Ta'ja&mU!n670u Q / L*,gSu@x.i]:eN9HHjAc3XƄH"aVZQ7|c]$lNhg rJDޝ@iŇ)Tlm i- ~YL Մas{Lް)~_u)nbaaqDփǹ#6lS7l8焳a >Ux U:=YX)]XKvGeZd5{7:bB6^$%a;dQr56 qIVYaPio Q$'K%2A7hd=PXxL2+Vp:6. k %f&iƸ3X-4haM0S DjD ԠWP3zZ [S;d!w#؄<]E'XLSyI/.N58}j/k_<-=zYqЛ/a_TYSd!DT r Ï9 ֘SZ^'"R\ޣi[Qv $Ѻmb/G[vpJ<}OCD=cNĎCebĈnEu@oU~F+*]1gv'zB'rHo,dqƵ!=qnihXJ{357Dr&r0qSBcKv3ݙe;MfuSx2L?0NR40gl=B:$zH hyȵg#&4> +K]jŦ XaC-{9kʮ.qFm *V~C YJ;3ΪyIInRnu0n^60OZEwCYۂ6ٻFndW.ZG`ً 2ۉ-;({ǔ$p 0T}2##; f /)"Jy'nmcމaNsF0nXto\50}Fu+:5]_䁭_4<0 yS՜ |^K5n f}F֣NWfP=Z/kR̐7>E)0X7)$ukAi&֭GD)􏮭[n)f$qu к f}F֣NWg"][ɑ- ySt9%܉xW*,WM?^|3[tBB9\8/7ˍ,tCNa<9 fwN$簨ߣ82 5erYu9(7䉺'IaD dF pS=$ib DϞrPs{>+.g|`g:/ŽWEVYU:q͝5wܹ_s{Tռp؂Ҷ" BBUunXYۺъ@ gyg|<)eζp` oC"ZN*[?|ih6X5Z >J[UP <71J( d` TD Z+cDzZJQvfhመp\s+7RRͭર̪JU2E iRrU*un) Zrs4✩b#)`"j/sWO(o$|@P :5\@Nʭ,<'R;TFEQq$܁]՜FnXuX܆BZ: #g3@2FgLp]3kk#T弶LY˜IwqߑԢ-%()hE YEPH7 Fm=aƷ_򬨮?F?+m6udF,M(vri؈ 2^ $7ĘGJGJ*P4G{ :l<ycDr>ouLٕD+WCĔ3_ hEC\jqb6:&WP?3} >g60Ob$@F+ sqF^IdfwrP8IxGۚз' اJl-?\egFa!s =DH;i TQZR,- a,U5K-JJa73RjsxKnv-^j,NDqDt MO+҅pI4Xo|#e(i<}ޡV #H:quS}|'Y*/:߾;&j|;fyǿp |wgt٤!H޷{wv˿غoP͗3#/`z(>nP A*A7,6sM1d\B@TC*B}ܭnuws[UXZ8-\uRPCEu@4VI (#*7 3!.BmVEm& W_/o1KFh(8Z fs s,MLKaX6ͳ[)nmg[K6$ $p NF EM]O zq?qU3w\Ss6n@/饽'~vjc3߂ʯf º;<93~Z-FH #Bř}M pwsy1ޭ >97|[~ʳg}~g;sb]U|8j3rP؈oѢʹ gM>]5Te>A4ÿXӚHQPwg7BtOpWe_~0RɊܖnG'Ȥeҹ Uنg,Hf)7,+{pti Yz 6[A^26 x3S:#0∹*Jnyy̘!Od9ϫfS; 5S{`>I0o:rGGB:[q`43kU h`0W@l*u[`:Qz2WI /^: 4׸ Io\6r49󀍜+(zeia輄t;A6J5~v*H:>]A9bj}%An.LV:BFrI{0wqjŏ% IFw>{F4Ia:u[m"EgmTTAaMe2#Ź.Ժ[ پnYpM"gE~ 8(s Y+bW=X=.vV RU )/AnBSW /*0c`N${7xqAÝLM3q- ODM>-sό2eI&xJF`+|tv$RY:[eo![_j LN_%T/$2Rv߽X )B WmK& YNOȼ%x &b$1hrǀ-Bcڴ`k\*rPx"Xǭ٪48dy++3{8&x[ζSnhʎYW$%>'MO@N=HN2>wo!^^` Qvޢ1f"\v~UhFJ ͪ,Lj=V6Y[isv>! -RW˕W+|Zby9WN3CR|`^sXAG7߾=k? P4x@x&8XRrd:[|,p. @N٪,JNILgF״2vTO'z}:8/ mR"Xc~J '˦Ú1q>7L3gS%Wr8+EhՃxﮀs<~gPĒVrĻOj̇zcU-X2^X8p V&3KȜL4s\1".i%k9GQK«x]!GnoNzvق5`3~*Rqٛq킿1&EZ5qX1 *Ѝc džA9O7S~eRXKW ;¾}ݻ%$TXő$ĬVeg$@>X*e*T`E%=l5M+ ST"$v"H\`6i w*b RV0kш'vA:A$4^yWn[]_N)^YlM5?X {n/7n5 ^JD79l1[6h+/虈@0D;'Pa 3a,+$jaDq])(LBy ` aЋ#ΕgJ!\v PRt4&2??’uùLq{DFW~+䧅d:c24ZTl~/%*$MY9ee7d&ԂgINts{9;;)rN)Nng׷oҌ)gC͐ oJMit7~.N/'Ggg~=yC촌f0M{KY2]4rDdXtF7|`n4#.(!Jӛ/{)GYM2vT28O5Cld+=:rc~i`bC#i}`~W#a騚^kg#nC./'#h:ꜽHQKb!4?tHgcpt~jqtyYVQn{J4b/_A qݏ SHͨ!{?.*[gaHg ,["LNpk;Fx|Cv8)\"HeA8)ה6S>!{uYм=\02: ,OFC2;oL&|exCVptLj\2@\NVNǣ-0{gϫ™{ ܖFJNiY!;4Ȃm0YmDR, %ň8@kڣy`'VsIZQc%3 -eB`Υ &L2Fz M +j(vV133˨4YM453sTsS&D'5}pl}k>zgKWwJ:Dd )!8k|",5wE1sT&"V-V5 e.J-NAzӀkJZ!%ʜ~"(nD.I߶ƺZ)\a8k˂rH$΀A끜Qn!ޥChoX+\0CE4@j.amN2IϦ:Bgs&`}8CRRneeĥm&1enoekþm.PieJM&n0<3kny`5˓â K],7&3MGSiaXЊ32F3n'xg[4fZi4wwG787 2f7;7RP{fӠ,Xsv1JuF+]lL$y9J"6iP$H'ʚEfKCń- w!,,C1(aQ P~lM`IPt6%rz64 ,;Zrt8MVD֠ٳ!y$  m\]+SN6z6t;]< rHdD@r:4+'){se)z7R.$C)e9`f}qrz4ߜi_)\-<Bʔ6RR>{^gKy|n'0j2ބIIw(sBl?Ͽݿoz폔`Ɠ(:w1ϱ0x3OJI8>@70ŀWj`k:bEMjs^q6y9o4׾4GeY/N lY"{e}Ei_z].3[I w@Զe2 :0/#w91DKۛzn._AE>bZRbBK1aHJo=JyC?YcIT@bEB{M&퍲A"vl,(0m/$a^/" F\+;>+P) ?~' |']֗OJ3<]`vlb-&S%)+|1% ^Lϛ}ss0mr&*%.9|=U9~gE-ӹE匌L$F^K )(ѥH٭GU{g T`}'~zh;lBrBѹD ڟCjXKB˹.<&h!/u;"kO@.u[^)3#1TKqՐsF펍< Ӭ=VW5qNbZJ'Eg&Afb3b&T֡҆&+1h`s4I\r d1[Imo(D-%m#X!mG%66󩴣&7_++_=׺r* /HzinQ]FjZwv9-pmSm;uګv햝[B1"vuȈ0>vsrӷi;:z-.r8v<^wG7@7z>(rzkpd7<[Fpí_R'#UCD o{ojP >Jt YBdUHBW [D E4byCy@mAB,x'_/FugmDz{\ Wރ`W{>mռއԆocA͆a͝!wqnrfؠj^57}.=b+=YbT yYP6CDQ@EtDRV3q̷'SVgOY+z"~퇛8ö4k4I¨e=%z3*2e=skДXL&8e Jn2bOVK-j= ZQ/6huzA{KV{@>p ^ٕn 2cp-\`ߺ`q2ZZXF69XiRX͇yr?N)>]7\oާÎWppKwY~PF7-ٻ8ndWz;g%^dZ H|^YĈ}9 ^%%$̌-Όľʹl9-{z*bW-9M:VZHIk@y+/MWs[_%ZIVK^-=|~p5=<z⸧=Y 6;2qbcW\?jgEʛYyc(ݽA]Vtp `0(\dck?0X#9 )-<čv"U\>ȭց':>RGFZ(d'%σ{ٚXP2۔hz[wAɣ}4ZNt2g ,쀯1LL?~̀d_?fW7(oŽa-Pĩr81y@BꢠzUWnZa9suA;Uk]5m.;AZpj. $gcZ9<2=,{]4{'/u=q)vIMk_U0GA8œֳ&=LȔ ~aJd=|fQ$ tȴorfm-LO?,e'izir$dTK~"|`IKK{*aG~} yȜ;qWKOMo+ϗA~ U`DrJ&TCTK)"fZoAnVv?-3A7Z_ΨuNiIP8'+}8* ah\Qq8aIc;ZCz3=O"J'$0 d"JʓX$bWFMF8zhmG؅! ‚3^FZt'*yEX3E؊'cmA"CW4`p~-@u@ƍeXIKÉJ8+6!)9U!р 2>y#4HI)PQ2Rr,+Z1P~fQYmಠC5PS׋C+],ԞpZ+̹_PbJtG^bz"&sNæy1g-|H^#~~n@_o6ԉ ַ^њ_,'c\CCzͱ@T7BjFmm[}~ ~`-uKF{ҷȚ1?LNQD砵ԔVɾ DՒD4Sa5G+mA.Ph"gz LX:H3X0 cIBɁw6y()m&5AH°ZzZ v2cB?Or2?,rg'u{e8>Mcj99NgՍKAT+V|$e/þڈb@fn9?2!|տ`<Lp2)n,כ?n2I3rٛ9/f8V'8#>q>ϯOoW3DS7Y?哱+a\"E.D)M\kkhhw/sHU:* 7UC?MPszߗ TUjW z3mƊ +ѡHpU;eOhء[z](>% (גװn'@;5Ixu}mK\X)%)ˌ葻Ԇ;kd:Z܋N&ӻ?_]8 g;=rgGe)%:[=}|Qe/_p/ϔ$ǶRZ\{%En$X;XZm2&]`آ4+?(uyQʇ2Xh㎣DZu0sy=3 HbBCp>Zb-xq>bh3N y#:u;}WvwWG+\ǭCZmG}tf*o!*#6qk㽌 LTK ٕP?$Kz?o+W3-˷7p?FNHz{u3{{]iYS˖38.=!C//U( D`JƄ;ԓI`\cUkC[ۖk eIlXIC5(Sd`Z {=@\!R N(`T*MAh*ŐMP II(#i4 nQT*'6 ˅xTtyVr$)Sd4HA A Z$zO?z{kLA9yؚ"EW N{|`|#qRsZυxYd8rl`KM`$هZ=!nz풱:s 5FRw.o/DkJfJs^F{Nwq{А6]D݇2jEGWFNʨC]嗵$a,Q+@ އѡzG鞡n '\K,gUZq}=R,T6E8@n]dRN,i2)s@g9tTa yTqDUMcQES3aDAT XMVPπ{Jlt'[/ZhR P4pWVE.RDNΦܝ֒JnWFJ>h0EH-bV;;8wC"FecDZJ=Bw8Ö5{TP\ʶzo_ݺʷ/m)F_l 4 %cO-'QS@QSB{)AWfeQ숵c0Z(|C~ˣiqn?t@eIjaPg;BmXm˨FgHtɅxu7)Z+SJskM2JDy㕔8%1.Qn@e}FrRsC(9&bhyUQ@sƼul4IFT!`!R:'Rl]=^۫ ҕ>LgK'mIWCL5Zvuc{#U\0jE{ti%xE@b-+vҢhBǜ\展v wXW1/t& @o0qw; d~D+NrF:Z:o Fu$u1jrb bfKT(:%y+))QY;;/y֪s`\X5W΁Ld9aOV)+)+#T/Z$FQ>kZsׅ[ppk"+p!IH#v<(a[N7T'p€H; xmA4&"Na!DH[*5ƪPc * hN8&s8wѬ)k㪷|q|O:7On!a9wߤIwϗPwݏ<}+Ns/!^5N&j:? !-- on,. 5oO>JOFɿ_T|wSP *cՈe'< fvJrK!fܖ",9tV5%L2![#n)Tn /4Wt\x7: < #Zeb_.K=n`TK"P&hz"$k:^- 䔛N_c^D-o3v+^,DyWwgg/~^tY5;ɨ̫1_.~aէ|5.m=v~%S*]w97 7dP} a( w{`*@ﶁ/z`QF1>}hzjd)bv9EYJNH.YRU칡Ede4NTfwMYj6 QhdFkK;#4jJr>ݧGX+0| QzwޛYTDa~} h!n 7UB mTDu7rȹ,\J%聫6Ce/[(\N-ri $ NgQ pS<"&{s_bx/}.d84ѭtv+Y֣Xr.2PQ{RQX$OI'}Q|?" C x=Xpc> w5|hH:j̍}Æ{Ɇ:Cކp+if7@ #"+Vvgp|M\va]#l5%d|/K#ҬTStcpn>Ӣo9^LU9e{,f\^L+sWiE2\.ᲑF@lwMF^>|~p=<;]!}JRBIC-h%0.m{H @P}DRoCF0zg=:(8LwgXrз$m6⃡ 黻ǃa͇S4*ͫOvG][Ϧ1\y\KnFM ޺LDWVqG8+miZdMWBg,}N>m$7E2n.f}r{!p$+Ie-Ӗd/jfOdۻrqiut!.܄<q| k}oY1BasUA1=P5 p;xX \YKZ.AKuɈ5h5pD2GI+z]5ѿ̞fRw.#mHΥ@6N뮝*TQ˲^sS= }Dzg No2!pKEp0Ua֝w]>Ȼmd=]v)oƄܑ-uv'F$D' 5FNnms.%½] a}iB5dQѹU XΪgVGj[ńϾ9.,2̈́Uo)e7R  f #1d$ K2[qj"ᔮQ:* F7uނ]W|C&>Ү\k$eUdMXk_ I%kä`'+>=HfIFΥT R5Uhk*6t=# On!_ ,BIɆ-ܷtoo 뱻21鋯WaUǥP kT[{:LFŚ>\yW&d8$[kӴͫAq`˵lc~ LjP4(lr.W{35X`1#m MTD& )VhI7  laNx{xQ(JO½$abDQ0NFwa-$&Vsz,nڐ^ ds\0-h.mymbiipRCA>^/q Bjfރ$Ω M +T3G7 e:Ig5l 1&k>ʏOz; 9O(bv?ϋ@٠L'ٯӅ\ ׃=^d85ПpM Ą39(JW%x)/ (Ɲ֖a9,򓛇hx.aQ0 Eorqށ0FLWGQy3g( *=V[6 ]ٲP"PViQaf0D HuҋW #Cs'VNȩkZńSOj`&1ﵰaKH9$;?BwX.t5t)*r{t#Sޔq{?.. l.# zK<輐)RS*J$ՄZ伨(rBi3n;g~J5LPw>HNb+/`̩Bso '83 |? V3T܈a_^Bw]i>|oG[k!*Li\ѷ m%UɺC> ZFw2wm \c1ּ%*o-j4z`JR[@&(ziBqt0U&y'6m~;%`Xtґ#9lH1 xJ}Yj^0$G+ΤM$xўY#uy_h&%+BKF9[xEnsTc\FU_{yx2$û<_QhLBĔ)dI- cV2⼵b ZìX%)A-aJK8%cC/^8p qZ'zG\#Z 6u%uQ_(,GXZLAODES75PN^&Q]U ,2Xm12>0M=Oܼ˴l8 lh?M'w zl:&|7E0H:%S֏P5JiY,>`IL`Ę `F;JaKg,3p^jE'3;x-{7˰Yg?| d|3Hh|v#ͤV2X]~&"w˸\"fV(Cn +OBob𿿜m]0\_D!<"("(%BIN# NhRRF[KEX: 1%@3 S+]~ŭߔxؠO9 F}vs˗}% a6Jz"($IIEzvi#šƤœvFs`w>Ƌ/n"ݍ[i#q=+^ r"'4Y?E#lWB:|uamy) e(+^HGsH {.-}k7}DZ"e8^CfD@ؔ2yåqX2#l3w[qdogrZ['d=L5[U˘4tvgV6yJwΥYv!ڝJR$0[˧;FѽD9ARЎXxUXx],<I3[ u9Z!)"B*~MD;  =BîUL<&5\kmb±<>NVQP(ju^wմNё}*$FZR)9* DS1փ"2ջ tD}C '} % \Q"jvz5IfD?c_tͯcd:j+Ǵ`TXB)z1P"tS, !*$k$PS(Q#X{rʬ`AH1\ri(IZz}TBQTܝ[q܏VZ0w쏧$'T{}]._Z\4Mћi2EU#3"Wi61/⩽Xv崊Y1_}SUx9qů5f" vW|[%Sݷpx, -I #^#ԞzIK4䍫h-l2ex[5ͽ[^],ݭ+*v3[{v y*ZK>NkA6҉?1 %oxeh@Ѷ~Y]BLIbj@-D)3m`~&MAERhR,]ɈZRf/)nLह I ѫ4abAN`WP*Xu : Z7{<鍾.f_l.KJYF7 Eem*DB%kW.V%R͵5saxַwzlk͚ ʞ~+m~S$}9i- J{ $*9:Ii[37j LJ?hG5Ez^#e hnXMOs+=7CB,"ίgE4H֬KtB]/Mq/Ўp6~RnS8W 8"9^Z2ˡ+~]Դ]TH!5G{'eΆUbz,_}BڰC dzD.\&W,a އ\Z^I ';I.4$T%mIO jBC޸S $EyYSf=ymUN=2EMAs s!Z31Se@?vahMOθM z`f{ |#s&Zx,9BXbx!kAOPпQ1##TZ1}J P ݈U8o%Psmؘ5F5AP.(jYOWeI$V-(I?2{Mh˕ba*IڟcGn:x5Ҕ{*n>_IR1 0 "%P` 4Ѕnvw=_`Г|s{D$oCWA<%OglgCi:E9][X7;|7Eꑵb$în$U#  Ђ'X(G`?Sopk) |IE%aEb@} #4xnaZ3|>b0~ri] bp3{2.E}-=QG6QdE6Qdl)aj}f.cb4`XMkVU#~|% ) G˜ 8J,xv 8U*|8Dgʚ8_Aevdc4 ӱ1rXQ5HjcYM`5@CC(QeUfeU֗)%Yr!at5>Gր?nTة`i}_RQU\?RTuF>VchW,W׉:\]'*Uڦ hG?XdJ5ip9Z^ b=6@1 /=}%Wk2Hp^(̽Vkah_0c/p$I6qӳrt$4>Coʟӆ$I]~f~W3ŲQVѵeg ^m;/÷sZh Y!,]}tЌa_`8\oT{R77/Wm~7b(gJ@8Q0'I[_h3p;XQ9{F4, i;o#݆)Ye{B9 mg=v1^QnFp_΄=]u[)^G[WVn竐3Z(}5Vrx:D#T1?d Bwyi9?/8u+%rKw00I *1;$T&(1[ݶziFO fb.^O. cCHuqJD,Q:4k"BIqCզkќ`6`fj^ooB1j xUEg!R`2* _qi^5z"q` T@SRmDeEJS 0A%N `=XD0$E `N9;x1k[e$"k$gGƂ``IH! X^bm96N }Qi<08(5 LUτK?D7hG$ڐZA26 s>j ? tfx fx #CT=SŕT}Oޯ. MVby,|z{//IҾLJ՗ޑR^|E]}|l= fA7XUNoR??Og_a2ż;KEkeͿ[K~~ ww|pli63Rxy7s/+`A9[|nt6QKqDvj?dp5Rp$AZm&POmkS b\.d:V|=KT*jebHQ]{qҨXrX>R;,#xxz.T||OMRq. ULuカy-!q+㣌{>֐nk;>Ts> "#ǢT& !Y Dr*UrMKʀ֜Y܏ZV*SP`?B;~V"¸{-|q\ԫL6_Hۨw#1]7_aʴjfŭ'ƅi7f^,tVܱI "D *ӤQ*|6mx6~) SoKrޮ.q >+Tw#<֕$h[jI0N@ט0Y:6&2-lNI,DpyҖo&ALN1碓X.l +r<V2Ab9X @SG& Tw~Jǃ!%(Ⱦxp@A1A IE7Na( w<*8niLI9 C;03dK,zywVW2+Vc4_䬭!R0ta%M0͗R y4AIRǺ1jEuW  *- 0<",ᚥ߭e Zԥ3vS G-0 -Cx?O=q{`fcme:h쯍n}'!\;evǩe0zN`Ugc6~K$*che6Q1Ͽ~1[k}v1| L̗ /0RSF9Y=|}.ݦ]YW\8mj:.oY>̠˅ɰ?`Z?-ᵲ^RcVqq( thR[GP9q4Pc[3)jHߧ#]Q\MirG[%89-aQBAH^#T|ʛQ30+^tIpK>Pi Yas7ު_NwsCN7A.R[ϳm\N!Uγ;`h ã[ϼ@RA\iwb;vمʲP]79hLEwa f4bn'qILezeO,%5YûH~IjiKp)ީ'zcGa X8lY8ObYGyU;w:nR)#:( =_ ƓkmC19d/H=z H,mS{_wSso)89Qo/p݁@ͱs|NQJa >.{ m,!G`{>\3Px< ;MlLQ U'BF,G sb.x`ˀFvRØJ3( G S2LN0QT!(& f:WZIcz"1Z[Ƥ,f &CC؆pb@xIU\rPTG5K"X[`I+Ԑ͸jC&T'ѯ6EWAe@&Ǯ\ ݼbGŴB8u9x)VI9}M%qv|%UoQ7ыG1m-i%fnLr7,9G~|GC3-+ڎZL!JxŇvfNhL:@v)֓T ձåHTrq:|!Y)$: h7cka؈‚ 2K&1]HqT9j WbvT֒ 0Q La ͈/Ն[湎l{uk̞ Щ^/@sQgl4P0L}WO[Ŋr_{-&#㽆R[؅ٖJr:䤱T 7)@ EZ=jwaVٌawDL/w_,&RЎ>UD~,lm.:LBUD$a> P?FE> 3^󣦃&h(INv3IxXI%oA0A봓P>#XS8_\xZV#AjoOP˛B S.:jRkG1`x=,n2"p!` |R8|[CxjI#(=) B)<%E DNMÄ qRF"Q1 aKJG2żJ0( A`@fh1rZ96ko"1!oa2jD{΃c\9l-%${FhQ*Wrݯ:M޹/I)PDeX,HNVx?>xK0xvYYi(?JwgQ ZTWH oߟx:/5d";OqQ!tY~gbfss뇇;x> M6PG\28Մ-?d7qK`LK/2"8>CkWªs4I9 /ayb\c",Ί;{*#g"D !U(IA-)6:x6~)JeO_H[>0܍[Dn0:q|VxFxV?{VgJø<9Ea)R ca(k^hĵD5D:5٦՟k(']::W9+qd_IL4㬆_a/s,sy">g= `6L=M& Jf#PPD*$Ֆ` &P\Gx6J$0I 2Gi{+^t]!< l!T ; Dޖ*c&7olefa/S of7Ly\L?..6W67G"2̬ʃsN,&4`#H4(x5NS0K)R}hZ Zo-ՆΜs/)ݥiX4^0}E t?|{܎K1[,a݌]񍴶9;a17d1^St:-JKCS>ӧ a@MdԲv(TڽL\ܠ&H:rIՙ$$)ZRRݛ2q&Z:)ouK-E gi?t=CsQdqXe1`~l"J0P 6dX^3?d閹JW{o;cA=O>|aM{CŽyơ_>k;e,}='g4P|ËgEg&}v]0&yʬK2{rPQ{Fi ١;E]u"N]r""Ȫzuq9;ܙ u1sѐ-UJ9,E6.f "} WI^4Њ34RZtz)\pԒq g ?⃕1*.b̌fEˈ ( QE2r*,g)D \E*zsvE[Ws5֜%+߾A=bgίFYIU:{׹ !1#VmFKXG%!r`:fά>9Ղ Za[fy驛^i{PՂ IاycWk6_ l; X@!.Dd)u* ǁh%l[?p(o\PTHRp?Vo⭥`'QiӰ1EM`@85P:8ю2G5b`lB#CS4#Dz 4Ȱ)K )8ºatINB#!5Pɯ1zk a)G]B>싍9Us-Fpc ,hJwc5X+dw 4*7A¼%j *={\S$o%B9NؘQ)mXPfApf Cj hSY,ay. {ZrT)l - B0ϴh4 ɄM3m&a)YV<@3:? 9bR*KYT}Qo1]YUC+&$*6kJ}fe)cFOA`q;ʆW cqcо f#|ΕX1H+| )se'%iǪNڒ\UJ-JM9Ss$[㤚˷0")?l\@m"K8]na>;'k!c(hզ Gb6S}_Zyj”^iԁ@] A󙰂4AxieVOlyIP-&TJ ZN[*mgzK5MDj3u6"ɩ5JN s5?w>@>})l!r2PpTL%-soC(WOΫ;W5f7cK9-C z=P;LDK?)OlL AEOo/z<Һ']UWlĭHxt1[d:ʦ.bNLbKN:}-LR)-׊Yt6i|F7 \2GE3P*!YBRvK溹oҽ`'L LGcD -|wJb!D ϋ\=KU˳iUDj(b5LdS1ya&dj-\Gv-w"00""3BḀHzU89øp!s#Lw=%1f-ĮI५'`kB1IQ,'q0eϮn)}|wcהNy :W.LDn_{FJ݇Ɖ RIew}5CzPm}hk@r;Zv%1l&??a0D7rw}=JoF b㎾YɽqǨǽ}uևɮn/@oitp JtQ )RMWuPKx }nɖQUivrL+[ 5BHٍUՒ!ik56b WD_#>߭cuB ҸTâIК^i|& 62WX}9-&xIĿ6!8AЦ?YӞ3ݧ]8 &uv=1lC2?M/>&/cX[IƼ8 uѤ0##DFC ג*n A*ɕAGJsjn:oz6(}b`Ο@ Q{}OSvƝ<ɅJʳgKNJPڋHFJ0WH뛻`ҬxzMߓ\\<8Jj~4On>,IscHOՖ]M@ ~ET\nrMDiܽ/br__M]“臇Rķ3(!eazOzWI^/+A;鷤 =fxKh6Cr4ހzx; i}~lY8۬Иr@_XqNV(UyaYYIRtb:Jw0)M`ۇ&f՝,Р~(tM3Ni=Uc=}(n4TSrǀ#ATVl)Vi A-e%6*|G< ͤ5½2hc/CA#H.ܨ  , hsTha5-3NJ3*3+, Pp*~b הSm7^aTEk^h;;TSMH&>jB5WlvBo}}*7+<8^H㻙kAPJ_".It2mKZ0 ڙ+_UBu1e;LoaNCv]6>hev=~LYfoSJc`xz䟊 ܪ m߿]YrwW{7Ӑ> ""WvL3s5l?0__8;wf6fqYD>҈Tz{tϘKqM1(*%*>}tܼPbJmG;O/vHYDQ.g=LJ r%"K.z!8RHqڛHL`[^L'9q6EĒ'(($xu ?éˉ>MLN`eݴR?U,`;ۡZK׻| GNq"|=~lWXS@MA> Ӏ^$¯|0Mr«gZ8Bnw~? ,nbě@ËDiIʉj%_Rp$KW5U.\t1Udbz0N~ԟiO~Y#bn8䅳hOG ؓnת0Pb:s4nsMbI썦J.8䅳hOI<~SgO19-$ɜt;ʙ.r7,:΢3?mm&EկE,U .ŸB1d,\pPqN,MV\/m%5OѲ@Z#$;PvMj E ^M2ޜlThXQۍ\Y77/h_UO^rY,fWqғjh"H_sRk)0 ]-)j٩H"] HK''dL" TiڕفBC[oH2T`+/uQPlCc<_dpiՕDTx']4Ytr6K:wSEP$hW7CbSa%docM_i̦-uH\s3 ;{qy%0Bܭx`Pz5.>ǔJq27RDGq?Ꝏ~gF[ޭ9j kӦcwg1]^J̋E1һdkvc]{`>2ǘ!ǜ׳>z6r&DUcp/Q3J?JGU.pjj][I e l4Gzڞ=K8; ;?tZ0,GFG O]%#R##=J'ħW$Iى%<L `YQi`s ;*l֛ay~[]AMɁZ!9eda-ah ;ˆSJw$'FD~{M,{{3wl$vsoX'8٫frZZO*iI_<g$ S\Sx-a,l 7]e %,̗dT3Ktt`Hk!C{&vRt.tBvCگ":R Șè'Zq@ll0RIae(CFpC,&-Ŕdc, f@4e @L%PR(ΠTXjU*6Z&O\&SZžd|neaɴ.CSbG[oJz| +%^"5m*❱ 1?+op19 y`ߴm̯KX?ug&ߙ)?0a=}Wl eW; `(@ ?kGyxyx{N~4R8Zxx]pS(ċ osTܳ&k&DpN=MoղzLz.ѓrPeI2G,KXK'L bZI S-MR( ‚00*C E8Q`$:o@,afG0/KeFhfV>XVm PMd:_bڭ*Il4vhE1& )- Ќ9{C56Vmcy('|Ic\ٷ{A=D>ParXNuMhw F%ݻ~oYo9:luaek ;ݜwoZYĂrZ/UU\K7yỏ tb14IH6o#*lVT9NH+tGZ/Iu_`?dQB6O;Z &U;ڍΣJ͚h~|JouXK.tِe#cT)n?9]K HW]g!F& HAa6ڥ 6v,?u]+bWgю1/%rUl[;[j'+>VڨM,Jd' O×G 0r_iRJs~ e-*1L .?ಳ7\ph(7X ffSa?EN)82S!ÑYPP۹,#9j8Mȃw8c'1Vq~<|QO࿤3/~j#٣Qb Of*ћ zTV c$-'V5P_~ws8-!Gѯz{p*JnﻁL"5_51׈Rqb8R(I{{EIZ4 #jUӃsβv( U-h2gTSiWnjSeD=cĖ*wu = #srIibruo|IIUtrqP1i6\XrfQEaZGV:鉻X! _[Xoy%bϪ $Xx z^SP&AnZl&DgߞMtXO٧CB`L5gx՛?b.PTP@ 4_ucMSAmk`C)=:gJn"ovЩQjm@y_]µauzE`$qPEE`}.1osSW CtћuOk_gd}!v~F3z/{ދh^ԍ޺17&^[|뭟Ũ%lvfKm֕[/~2y(C+IU9g_QIP-RSRr:r}{^MؽH3w91 Po7rH (W{mЈiG2 &A{Tr9Qqb_ľ%os5b6E\%EQr_D}Q1QB^FąpV.ebo"#C[|;Q֕^YQ]c&()!|4rÛUR^.ŕ77G,*}z; vHQG7RpiRQ@REb9RE}*$Ƴv?ҁۯ'X Ԛ!V=/Li;DSt$û`i4%RA vc,Sڔ!Vը`wPfq˕Y1-IwIJC8>Pe< h4^h BSR-2U`?pFdI&L\EZ)RD9S*q*40(8'_MFB^BR`o U4qKHjL%ު_ J6H0: Ѿo8ﭧ^`ee.A\MuuZ=.]_ S~nlnWn&'8ǜ̺Ӗ?na m2ns9of`.h`Z?e8v[ [4CRزUfGUU\|lܓol/>C_ ?r>R{*Go2Î6AC*訪RĠ,C8L`h}t~ pWze6Τ]8`=@L}{]=j2ĿG9Cy7a;wOWҳ} 9gaQ+5Xխ YevoxZ宴CU%v @S%dXEIbed"MwĞFJJb|s"9kZB`,YrE2 4FcYƊ6$L" C ג*iT6i_n}8ǯWPȯO?wa,b|q?= Q})9RQwas7>F%2Bɴ d5c8K4Ɲ S@XT^nC4_Ms]"հL$3I$S(I`Kj )Ia-I%y˄I2d)BW=xt3Fh6páq6B]Pr(ש݅4N d/2\ yCMhsYEW> ԛF*Km`p7(%F`ٱ:\̟+Fsɨq=3Q}kj4^tPa^ էJU]%A?a5Rd[-oUXҦlm}ygVN̯0 n=#HY88+K529`l$" ndR S bfHGd9$ߓF`=hS.|I,i}`e` o"A۷7!O{7aÏo̔2?nub53./ȝƒg˴Bˑ86Gl–MMsCl U~6~+/~+lQV3@/u lO^%҂V%.za;_eٟRAx|c /v| {իy]U<{aUGܬ!/Fw[o|8S<?n C޶ 6 \lX~&!xa{ oGGj2: oQ{VFYˊ3S\\떅P/:'jY}p~m?|Iv'>xଫahUcp:2U<ۏ9? ᗎ s_]#϶IX$œgx`,u =g8ZX4&8:OWp:, @;;3FBzc$<0~.3(V2ќg*fY* ǖh\q4)$1Q xS"ӮxN y0S,&OeƘD3[arc`&7(شLآ%/6~%@7F(LI5Ҭ FTԘ4?(5r^jHB1BQ/8FRT~a=!Ml3+1N [lOɆ.L6|q$yœѪDB+(;l{h@9FFB >Br}3F` ͤfL!XMA# Ӷ9E{挏c}aG@5C{3Ι-AP*n7etl}PhxhLy~c]Id -щFv>x* ̺EO4Ժu!!_F#8Fb[,AFlSnч֭ E4F,Lbv1jeÀjv K;ce`ћwEk&μ>6dR߮t=h`<*]J\Su,[uŐ'q=I"Y 'Ex&Jj,6+ka3EEɐ0V eVZL0h RfͤIQhP]- 7x'K4էIjMy6( aRYA ]$τ&y&hph|i`OHr&u *K Fh?A,&b_!v$u{;N8)j_ 1ǯE4Fss d^dZhTKdre=3e"ŵs2J igs  oQTE@T+Bb1ي~]9%Dq5 ELQ"8Fl4 EUwA]$B5~9 뛶hc^Ihz:pZcá/x6iW1s/ޗk0;SO`ar98/L1 w3&[y]]ݵc`wWtњ4PjF,E$n)MXJD U(I9˰UBšbRTWeڞ$-Gp[ωob 9Aj?cܫĶvK>| ,(xt; ? pE,݂_S+2X Ų@}!Z.V `?zQll&_{xt7('<]}  ɗ1Ykp&83Њv7*)t@:acAPŝ67gT&ߺ%tgn O. KR쩼v]E]eĤɚ+;_?|7JL{z%㕁YTw7uxW/ و,%8ae'giɆlU/ UG$2㤒vɚW>tq\cR:QL:t1< #T.#g̋M?͖ i?pM=\܈+cv2[.C̓]T??&J몜[S˻Yfwv0tUKu1oy\b. bXlf+$&O,v\d6΀ o핆#-6P`y>6Lhf¯&UvcK7Vsvƴ1+J;nJ/E'{}o[P}g{_/tTa3|YJ^\*?ڭp{P=8_+v®:#SקZ]Zsu GXe6j\tN[FC(|{x) mܔ:Cj ;/]+#ir0Z%i0);ڿUyCOvtZWWyr;X8x͓gwngA kVQ:Ѡ}n tj_$-CD ޻ yλ ~TDBޠm}+1"-ެa+-knjV7l7%ɳኗgj2AUcxB5-)Mp碝-f9a}17LO^ \ePz*-ľ) _\W|}Sw3~zEȸZnG έ-c'n0>8>ŤD8o)B2:*.yu4rx040J RRd=zpVVQK&ZQyKAw&Z4G[w;63+*APA#MW9n  &0U=-I%R<>u[*T0sT _TZ 29)w5=rH n󫊤e`@d=3QwI>ylkSbW\q?blxn#^L% vO0,N9,;,IٟåR, v_?>h7Bj&sb+a3`2-z9a脤R6.~}z'tp,1jc8_A:^4zHr&:ʦ`qwSX bLg7R|{Vf:6/һ尐 716l;}fXwAFwoW,%^wrXȅ(2j=Â1wAFw~|uZwa!nm @+ "LC>N\b)-qݞLZ4`A\DG]ͮF)ƈ9.z ]Nc`FJN(*\9J/췲ŰΙ(FJa8 q1&;]fVMylyF2>rZn0'%RNNITfGM׮r4ESٯål5 {Um\Yz +]s*kh.`R%q n!'d!H/_jB (1¾r7r:*늲Xͩ.Uy{jm_i[X- xg|)s|È({.pNrs6nDH({nOL+zԵ*;lLB^.ƛ؜R`_PzB׈܄WTۋp,]쁓6\e'!FjVQFh:3`h˔5Y* ^,t/sn 0w%3YJF* IV3Q~M 䝾fqG=Dܪ?fcVm^r14_Z_ZkQo Gq jås:dh 4h^f5gdF!T6?q[ۛwm֦ k}#DgNKRu7ħ i%mu}$!>=>53n{aKfkkjqBmA㧅@(G^{oun}u&Mr&7;6qp,«.G_bBxsu C #~ >$'3;>7@_~ 憡c S9\Mŧӥ9/h |ntW}h3>~VO9QHh[Q l ݁o,r8/IJW/`ٹ, 7"V83=\i3=\Çᶾ~e(fTyK;Y\Q3JHLOWҮhd0x=8Kړ[z![R\On?4OMOgɷNO%9?P2 9Vs}dAH4vXZOsrN'mc?އ'fS89>4Ll~Lg/CϬ6i-/SwӫOc2A{vPOO~1fvI80q/VwYf.DºLǧ?A4Nu2xly9ԕri-늬Fm3zU-_;BdbLv'P5Ns[#m +<!u|zlʧ:,kLæINx8[14,UrKR @!0%Ejlږ2(*4t]dtn/G3MKۥ(G>Jw;)a;2~Wo4njex{*ë6J^FB2Ps"ЩM:0@)ܼ0㪬PHT_%R`C~?,.È1jmK>BVH#fђTWFöح lc|n~]z#d[E,s,dl341HøkWEiVZ־6AIyko,Lȍ;8%J@ `tJj3PK 9>mj"̔K2{8jJ"Dg6jG40$g={}q&z&M;=(&.8+<$%ygD1FFixgkFmk[ s|`O{|= `ad]~7+GF!} jg{"jqm+=tD5KS|hKXxK}bD:ϥy"VzMB@;{^"E¸gZ&pҬ"A/t{tȧ+WK'uKzxҮIjS27"oF- }=G[+UI 荆B_sԶ6b 6 m!./5luY k^Y D"{xr;,߼bOZ{xse[ dNE$Ђ'㦅AS w߯e-2|˻w[B"NFD$bIv)h fRT"3,eDR*!MR vTG3'7N)W￴X;ӳxTSY<9gbətI 93,09r&zajM&_yzzrE1FR·Ґ2I2Vʌk0vRڣh  5%(LrOS.1*rcTDKyN7ky&(rE;gL('6}~܎X-l{l?|>uo?G Ad䚴:28 J.9FҖEK1ˡybYƯLuUdbҢodb9xU .L/Ak3̈́H(6Aږ2PR o08QArb~ˆ4xłfHtp:ri tS Nxp:b3 ..d+E\H@ /UsݦvsdZ}#d­C:s+C5mKZ_^}ev残e$r,NBʬRTJ8#YO^{x *!F8oh<5q)+וbhlU56VfGBo 5Cu dkh:5H;~yy9o3WVA?y<~=z>ɧ/d9T18Nh؉}K8/٫tVe9(}_8gi&J|DZ:T5ׂ5#V)^>e&h$(_t `ஊFPֿ>!z0pĢ_#gF0'~6z? Oٝ/I)l{ >}"H-o_y(=G3C 3H39gH DˀF܈pG-DrHgZbgbj0,.J<+hyDnB1~ֳ<w1]VbU3J؃±X|t^ܬ ju4iNQ8[S0K?}g?W1%_žǯyX$e^v|97M.Z}qK2/G!t=>]$`'d r}WX-E[}$;XԲAQRKT] j,) 2 }NDuR3>֋:8Ji*3z+'aٮaE͏7%_A, Ϳ ߠp~Vܻ>N:l*˺Vuul8ʌLz:2-Ob'-\NK &Zċ+ ϋJXtƤg\*ϋ5^Ni P"N7qj*>ߓ $xƙ3J+x-;%3#ճɝ; в|"8}[ʕA`4\#Mxt&4erDhxE.^,9|הFh㔘ךFȶM,yY+AQٲ)V0EwP*D%^K[25Z~Wm!jc!L9!mj3cB'\281O686>;^} /$aTXYEI8arG`fwѯZ"Zfl1~ѭO`~O<[4[#$8AAӓADOuhyuD6;yꣳr;/q z97jQ#\ %ªm{qN[~|-\qlQaϪQ8XMrRZF/Yt'1p$g|1øv 2W5R&͐יaR{LJr(kNpq|;8+@h Ŏ3dK'ڀjv 3Y@!czbq:AH+eTdQ ADSITj)$DTG պ+f-10إRQYSiZx# 1ABtnL 3HcCqՔ 2Wt; 1v#D4^h2yya'e'^_\~1A?\>2<ɑ_.}(i9Vh&9?eUU.LK+C!LK 8/Bf)RWVe׽NN4cZ$6$X{P@Ȉ@A@1* TQ ıO>gc4 =M>t?R1 û?]>mNAגVVR]Zs C8Ȑb@2kŸ͗YmM>ar){ls )S\Ɇ҂O+GRS,k|ϋC,Ib\Kb(Tn 0 KA ?ۃ>$r/r5BVB:Å'ŸuǷqOI熣bKnj*1" zJ^\~E ?봓z[ VZP6,8Z Ωn%7'UzӄV­ik ^/o#ֵÖ潥/jUM1vnxE;"˪ʴ|{YUtqjxpTp/⩣1劧JCRLd2cDyfOüiG"-c^㱫R|BN6ݚ"NK0ϸ:{q͹{!$9]_Łf6[ӓ7 6-GʹX%+Ni/K5Fc6F` q*[Y"`|F%X#JtQoK+&wtU *b5]24uz1R"Ͷ\r%.i#|X|؟=<NNᾄ N2לXij X殖aTqfXId@ VhŘ/jң 92/=Xix5 axQUM9'׃(@u);exa P?{n |4y(rro )o%IJ@`n9HP3%xWybL&™yg<{성Pa/{Y S̫g`1ߏaͤ1j0cQ4YOkfN[$M#&!k7H%HgPy|/^kÛ}Jž"'x֛]*hqY&rqxI:O0B3y)2i㐳r !'8YKFX,As5 ʸ_].D0,XCgX#p*pI37a}g8sb?gwY֤& 3".2`TBgÌʺ[dJA4Œ-O ъ^6"y41wg:j 1qJM yO,zJ1JQܺ `B5aS@؆bA<* L{,$x1t9PZxN<FF90YM=&Bium`w yZ{Up TA b&U)Clt}nx_hb1[+ib9cT`-7ʙ\36,֛AHod(ĹљL N1<NS(-#Dpk RjdƘ3\aPOHU at&ksؐ!X +2gB!Fr(hvrQne}7M&/#qup݊|![H7b5b?7 M; q!wmv%Ζ>Rh&zvG8H2}KzS(;W0} \]k)PӝNNiF.pH)c[%GF'WZrrSHl}h7C{HKBUĚԬbK]iA&EǼ!oNN$,E;ň7|PAk$P89 ]B.VGN<ś{x&{P "ᴟ |w:'/pgwBB|fg`YT!(ֱdg tkY8ZF E/gç?'O@6X]2rb䑍 m2p kM75w|.%rM@Q,@©u=Vw'_ತx\68ٔyerQ#Ø&upW=|nƯP(Lag.B&I,I7n{_n<>y3=} /3֪+2!9R˛k6=X*Vs3Λ+EYftlӥ᧯:]E@ٷsqx=vcs>9&zq5{"mk%t 'Ɗ@rY`,xGcr%KF#uUcqz1O,q4kEKT߬S9ws$J)5aSKB%N{ nF}Z':CP0T.}gȏ y g)qH!_\b2OK~觥X&=#)4GpbM֋Ūb=DEAsM*lUBw!Ra4k+ݲeW$n^-JV2K{gu `:J(u<[< ۻ.^KD ,lh^w}6i& a9>[$lVH[.ť&EWsHn"ѭ]5K/}?_qWW(O DpVD 9{mȲ {_f3n} %;^ęGH9rpA˦Fi@$W?ZȎ >~v^+ו ߉ZYí35ds)+ǽ?I^>{|ߔ >a| KE0ץ󮱤Tqs_N/;.Xsuw?Ց _LLǙXVֻ9aӼb~]667EhBeln9Y!T\P8@:*ߘOi5P_CM1pz 4JVN;ymsҭ(ujR͙;e e(+r\i'pm%B)=B;*kDɡHX+ @. -l &䴣s)0փC^kcp3`'9hl `Q+4w-QwD( acR,F#€5z9cDc"pN! qsWݷZk=ds쫄H2dC귂b=꡷»ySu'mssxo+˹=X-Q]R]OQbK$!yàr (G%ֈ>@L8,>؅ U~MҚ׺D#L#Yro$?ٰ| gPu?e:H*Qy9[ӴyQFK9 }oݭzr㇛w`Ϳ_,V}ف[n7bh\>>S%:G~= 'O 8 0FQ.\ns:V)k_`>/rgtdC[rw1_ґ ̝ rTO3s/KȠnPo!vׁ8[玧kL36y%p"47OWRhLی@mЊ>?>} d 8>p]ozx5 ]{aW-Ï5{}{U;0`q'z { qq>k!OU7Y`mN.:Ep d^)_yd!pj'wqۇCZZ5M=h|Ro\ME"7 \[E¨I//i7Gs:p{#}P[Vm|"k 8¨^eQ'ݭE^$ ELiݸQrD֕!u6x(nݧkjhjzŹPY9֕!u6^" guF4TŐ\DdʔREmzy^Ϧ ]2~ίs:.[_Br.k"y[W2D0AIay~CV؁0/Q41';ܪncN;()Bh~.{T\'™zcְjjԖ5']BJSh~ebmcI6Πgg X5ёP 0@%d7i$i.95`E)O>axʓ8ڡ G}s4>6/HZ1x}gpfhlEse*T,괫hI tJ$:R3aG%k[!֚ZN5o'hE&KЋt6=h^(nwpF G>}|pR<[4Oo/sf'+]"LT~zOH (.?Nn>}ݏʒvݸȅޡ$:*= ? RՅI$ weҨ0`(IGWnユ[EIyv8 [XtliĔIsA ;"™hC#[< o,hoiRw@Z8Ֆu0wcbV+cM(^6N`^ɝw7mYGn::7 EXEԻ: 72]% "gߟ'H J*Rcwml>m`5˂6-0okFǵ >ٜ"gHhYϸ;H,zr86fH5~Efd >%{]}<"sj4_8:ڂBQ$^ꉑV*3%y@VbN'0hXjO  VLd$H~HT+DYTB--I[ إݺ(3Fʡ#8Nj^g}@+c4F+g4؇FyFHWgR ˺0`YqhAHm8JZ TNUw)}1rŢq.nh >9VjA[ 3][;*腢L}PiG7G#z-]'Y9XM6Rcv? >WNן<~:FO(-yO%?cx bĈRpϥL;~B|; k=J֣ ܌ܣzhb\&BjK& isFlyȤ!0Jv<@h[m"vl}5wT)A zJ - WwLs>_Zio틇DVkR hʝoea]Tݪ y̳%)lPi4:S<[aR wN8n0)&hd؜eF>1ePkKgk0LJ .R0)&!κ ";!1ϛj^$%nߙYc }K^;h 1s3j,: A,(v O5NO[MD,y7% t90G-[tE,Fb kƛKwa$1Rbӵ-5(h*LT& X֜)5j6dl2 B!@5 g#p3va$$H8c&`h% kvd2$]eźM.PV!S_2Tɢ̭i”EF*,#GHG51֒qZPcշ2g@Lј"3U-8FZ%G9_Xʂ,dVmI`OSV7u&Q -u5#RIlG9` `Y| ăI~|mD nDi4F 04"nQVo9aiڲqHklE׆넜_:=aga꿮:l xoWۉڪ _(хVLPe@]伹mRfx'D ҿF'3cu ĠT@K*@TEXG<\B>ޣʒ`Uk0bl",TI%$;;4yh~C=dz+!GF+A  X'|$rÒȇycdlS%ʱtemIʥeQI| 7\4Evn yB T̼yq"%Rp`^@iD2!Z*s["!+@ |Δhfq|DVki\G>d]|:6_L_?q1Ukf2m j`tCAR;p`$%+.T'l0wGIWk$d]"ŁוcV(]acK:`FPR fRu47g)=I) ԥu `LtQWry1 $uL?@)P;S W+3'(/KY JS22MY'HT$dL6ꕷvgK-kǒo>NQcQ3/T/C.ZضDiQ$X LD.h.YZWlv w)E^L ]IDIZjJD^JYn؏| @hiESY: FaAJ`Z bCY2wͺh%E3NBoβg=]Ӻ#QOdZFxl[;J^q\杻j=2P7uЋL}QigB/: yt{ԧa]c7}AcۋmSsދh [ZR 6G# yޛ(lJC"'$_TΖf׺_PE}ER𥘒b(rC b .{K,ۖɨ]]P&4>ͮBp^2"m1jg X CΰtWtm3 5qqtY79/`v1OiBd=8Tۨm#jˋ"3J"G "Z1f [b+Trb2@h)eƼsQ0VMD %&†9dIVeؑP=)IKHdQ3mR@XbkCql&$𞹟V(i0+I6:6@ZB5E0 ,Rf6ԄqM/)T֠L^8 %0'"#fX@k13*q@ڪ!eJUf&rh-&Ȏa2Ix^ 3fbOS-[v?AvSgpREU;찇fB@BF5ؤi6 v`TYJQ`e&AIy 9RѳюN&@K2u^($,?8#Cf_EZ20Q!$KuRϖTIF|d6J^ƙΌQ)D˧d]v+y%V Lbg\5jqu3P#m"j6ŜC "J9du@Xy@JXԨjj68QC2k,9gX_<;~DdflhTN$c)*6kcۥX/jT$E -vQ鞎2~ޱk);P-Rj ?RcY*roYK.s`C +-e{]I9tyoWR{ [[=]6}k%&6xcf9 K|~ս5 T{_y?<4oL"xaH]]`zdy74}Gsȫ Fa#n-yc#%`P!@]jj?:pEsIҼy袻y$ |rA)~stYu Qͮ8\1zGsir*wH>!f1a9s}`1)gU*(%^JtHG]0Ecɺd"!0Gٚ(36*G8h# q`H=[_-<&u}_l 7{UR|L [UfvM~cqYj}6JfsuO5}w?Sf-[~@)ws- Α}hPWb\ud yPt~6HͦE`3RR(tCCϙ )ki&`;sV|XG_ u^skp~bܴG~s<n&GfNGH 8wF]l;E(ING'Z6S' ϩn;>rynj/ (w<Yi:)rp<)mmkS"뒴(8P y ɽi&>Xap#lY>AǖR8h #lq==EdGL <-#'Qt*zܻdz=Q<-$Ξ-o)'eׯLJu%맞[blgO7R7an kN jdNϷL;?]9u7)#K$#M\9h AK!@(]LY)>?Mя,~T^GC@z -APB}: s8K }%ބ|Y//7+_'l :~.tIz-kgOB=Ύ/=IBkB?ԛo.y0Ppkb~줜֛k!]vH'otNԫNzKNAsjxY5Ǫ=bwgz 7̓~`ZajmOبƲ^t[G&};X5zN:Ds߾6޹yu1VۣџSS;фbn"#Si3:[MUGHf8 ݇6j3Os4pXkOhe7!n1Pk isF7 ܗn, (:8)wN܅ZZԞ-#0qT`ߜ= |NY8\4!ϝJ#=ge7o^u+ե0ݟqHVYs1TL'@1Ͼyv Np9 8J/r~/oi\m怕{I58`>hR"ǙP0Ý+\| dP]Жm)YyLīuDݎw| QLR8#yx4L9p>I6d  "B 툋M;`WL" 1oZt%݅^$KYH‹Z_^' 9cR㽩ttdv|Sg$TulV-ᄖ:*9)Q$\ R{+ Big cUq4 -Fp9M#+OcגA:*d(3=LƖSCzP"cNƖZAK- ? X1}C'6өԾnŔdh>PŒEkg(3P4f6B^S&f)eSbZxymBHM sdlu3Y3PfbK5 IZHtCMR=Y{^"JJaaqn$2䅁%t,ƒ` PekV9|%:dQXZBMt˩mcMt3ՈZ ؕMrΚ ;Kya⊮.L:C$J4l-CRbtvҒIFFSx\l.^7 4bX}gV++;Kl}:-vαtd (Y/mah9 ~G(?/A6u!vGÝ+-U7m˯gɝg'̿eIgX' ;-׏|5/0}Gr>,)"y :m{죫ljXs {Kddl0uZy;&Ƃ23SbvoEIhK`_W!aīFwRn3Ӧi[,wdf zx3g{Ktٗo.7tl!`U ]:m3sх %"OFɺAB w%;JiCN7x۳M/D\V㉫ɳY^/<~ bA0prOVm<88'6V6N͓joJ9p [hh="Y{ .Ok88{Q^[u-:jPId-PO8 vQ?`PC NK C;^2eHwqk0g[;gC@6ڜ[&xǗ ,'}0%lNUBJwF%H'cۦjF#l|}FXAyPk>&W$BCB[N)fN8C@KnЦ0hY֓T\aقy- ˳6R]s ؃MX޸gVK^5/?C/`bj 0_tvN(L(<(B\iCI5x[?執ög[rukt67җj\ dW%I7isFm~=i \Gm^Nt3zFW-^;ϲBv99 ]OFIo:=[t!t@;vcM{t2@NJ T3,ېġwZm\ gCj%mLFo!!=b9vrՉ{"xGiTq:kqa3s}!W1=j-#gPu"YdWbI1-yB'j,DIn1 ѵ!ECBCt: D.|H];58E AYFyģ] !DH4N;{@)'ֳe(8!EtVޱ>YTyMc.wq5Lv|Az@ M@*T6*%# ^p5-h.rBA!+M"aM_x!NwgMvnߐ`#=51@;>2%8yv)h4(hݎFf1v+P !!s#SMT]ڭqQA}Gvpyj`2ֿMvBB~"#SPĜ7=$+yN>j~ݒڿۋJNVwHY w@} )Έ[EOWHق[jc+M )7OO` UHقF&Ol,cRJ9 jY׆rzBʮa/8l#~砭;mݒ?;w.KfMzv!\MjQUBwL|,Z k)NO%JV80c]2XYdC棭|Up=bVv #G4S 5٫#R3P[KJ" GI^^"]p]A1V#ĸ`4z&PR 1 E#yOVЇ9|@ (,QeB>sP#5#[ӻN)䁆@L<wzfE00cSZ2fAM (.A ,xgE`F1.fYna,ap k ʡ* 9(Ť`Cʼ ڒa"ʂRP2T") ZE?i Vy th9*Sa/*TW69( pȗ0򆳡 'A2&J+;RO#_! +xD" U5r "·l A; h zMBLENUM蜠0=5>DεI_UkI=ZYOח76T>WU˛W˃\,U@Nzϧ7{x'ɡ0GH~H%%VZ> d)X*ܕm(({QF+"+BůOE?S&9k丆&uV|7C wcӥNX]8kZ6Oo!\m5Då5-0iiQjG@J S} !ҰܥS)q;'&S${נod)pTR_P48q` p րTrŹ^R`Xʼ#.d-٪IuпS)ݴJ_ `9H'AV~x. KnBfR2i'_ܾݦKk{Fյ̿y"ggkPR܃%4"b|{'|mnO?97C\e`+Pt y7?5v@(st{k'./\\G[oȠ&QTiY%@Uܠ@jdzPQoPO]op +r:~*z]~RY u3"=ul%{yusT<5oqLq"HZ~PL78dfHST f7;8 F{4Js%O3 PsZoa`)yW5bZɨ/D?CBD3 9ɭSʃZק*JM  !(EpnhA8q D B*${PYۜ$#Jя>\>= 2,qZ?ǻ9h?X!> VN_\E|&+«ٛ6Kz˧n{?3<ӳ?0!P`gKZi{g\P fJf˂^_Zř˜=T[J`gOvX薼{5 A$MЯIQgm} &A% |&QABNDo<#s xӉ7J3CI! :g)nlX!lHn?\+Bơ%2YjOr)Rp 8Ҩ%LD'YjIPt*u^z%U@aҗ*tDbB[U^:"@7)Dx{[4`(F$wV^m"S0N=iAeQgm6v.mmkvҙo~oی2o³&}{(ptb?}3~a-,MuJ-<3}S4u;ɂC8߅&M((́ _ KLI_2Þ&^t .{Ox ,YIBgE TUPtEde(+cl\k [[^>A@KvbLtTj>1ɏi-DW){Fk君HNZ#[ԯKugu "JV&?_4d4N Fa+OZZkY4 v༐~MI0Q?ۛm[/z0IEOQRW&w [4zKV^qi}FaEz,8`Md2fUhN@#yfʚnlzYʫ䅮6UpWmDjd'T.W9|BЪ oQ7ُuM{҂>uo|J]A AN]%Q\t]`]#I֎&d[f| `!r} vV®~](ac.|.N-8G@jHɺF7 IkbJbMZ(a2I%Kx!Ntk,f" :oA $l;4EW9 A:#39:K+mg|fSt*RH>kRض!u}9 GHI6iec[c}s{F,3$­R*뉻zEyy%`!2]}lz6ksDտ97zRz1ּ^ &|K]7.ep2jhܪ瞩Z55xkY.%nCF^hiPJ:Xf.멛97ٱF7nloD ,Wtc(Y=TQיU)̄Q^@6 U ƕ\P|䁡&qxG%cE3 Z;+E%/&O}Q>:}*JjCU;#ZCi[.u!ڈL=;XU^LswgӅeUN+ؔvl]meaG>,q@6ٱ.ؾoŪg:~99*5W={d`ZKQ2:9tp,ڍ^dW{Li-D`}6Z44Xи xPeSK Lj"AYa/J# ٴȲ )Ֆx^dgÇoط~>!;YB4U B_%γLv.xą||Igqtip7{us{{oL쳸+}n;A_7߿i߼{ȾM~f]^bo!;6s="14 VeI35dbz{)"K~daQl#eѣ$$ncVZFtRjj={#6)"qFyGU&-]8ykGYȺ(HFYN֧@ArZ}} X5)J'*8e_Ùā #RD aMpYX߈q1 ן1Ѹb.,ڛsoЮ^91S~}( x.^ى1J &GX3c9hBW+kB b君X'I-NR.fTZIE Dg5+ख़"dJD(x}/UWy`l嘍4ʗvdeuxok|$=9p2;vyT+kБdeʒt 99B)^`F^Q:VF+!X[+"- t>/ XDz< i`#5k<h0VmydCK>|ni_JE'娐?aAQ&gĮeW˫rugA:l=/՗Ή!ZW,(W7a2Ɗ3D)kEd /&l"Z]j  Hdɘ.&@Cm@LL:"&= DT.Z[< ή6ސUK1Z߄d;ʩ'c1D3ʮm\K"d 8Q`%#dbɨd[MB}_Y6߲\ȒP 4GMl7 uY),edBkrt'ǐg1U,40ɗZD^S~T%"m&R(ܽOm~wuiRV) ˶t{hh 9ˡ۟~LQͿn?c3h1B L" " O|d_zwsRMx8JhdUxaّIe&$ïך  669BՍ*$Ԓ8v3n >wU\T ,btFX͝bs6*[SJF.ؔ l* >ɒ-} ~uB1֛qȳ8lB;}$h71+45,nxQ&dycΩpW,мAd}cR.F@DLv"1uZ0yukU;bLy*˞&'lݳPڭu9'''V()̞]rl o7:KQhYSuyS9/cNсpqBhpp9>M uaUL/*4b? T4"+WT^e1 6Mfx}v΃Xy{5>W=\yc ΃:Dr/:Y*h#PװL3`jڍQ./ۯb9Fa/W}wYq\(0,  hߪL۴Q,bY%ox. +u=Ellt[bեmђM}  GQ)R)byRwǠ1mlH -Q6RTaF{-ɢOK URkq8GI] Y R` .\+>:c QH6FEwUR's((tA}*U*E`):h[69 VJ;z!7TLÍ3Lzި~%L6a  րF-B722kj}oMqmwGV<pnVv$AG#Y7צ10Lj-aL< HMGxNr:y\:ʋZ88eťZŶ /Ɋ[+/(k:RNLȫ _:ĨǍ!ŻhJPvQdAƼ'vfvCtpbWG*O7^!tz:ϹFM9^02FDۅsi NVܼ׉S[1s58pX>>1u!f}sE\ 5flvI) yaXOpMTn=۴}aYgv51q~էoa/Ol1mFyX i^*jYQ+;":G5-Drt;ޢ=:1V"Ì6m*9LD!y-'VZO\~& p_ώcFpNBÞf`KF"34Y~%h.gR2ml!p;?2L`J6~K,ef-q)Fdr s>#)tTŴ;z)kVH|!2MXFD젿nې~Kv{$ p1, 48_g4tl\gX k6ON'}+FO&QۣdԦ023Ά龒G=:zF-7V#h$)$ Yxt'$\u\F0öQSJc>p\R,ʣ"yWH%2Bs ^P,v_?|P MmQ^ 3dnq9 9~K'/һ{Y͖ȓ.R34Yy傑9\X-%hJ4STq~|\Er=|L~j Tv@)9ܸGmn&3 MD"h> 趐y1fDf1A&KT0j#+닟ȯVn&Wz< QuSzpy\[8c.}[ g6ICO.8B_7bgq AJ&{ Xط T^:# YK}qe0t_1&tr .#:ӻr0`Wph񖌖ꈟCcQ @[5c {ʟyQů[ ֶ=ŻBYQV<.Л̧=l5l>yrl|[m SSé/tjԞ{t(CeEQ2qpXj+C96L8.u㫼{5:c@|[7ñ,. nEXUrL_߫KϑJ:HΕU! k:R˝P-Y׹56am  Aa5W-8G)ss9I2`:II*JE*lrт[$y}nv.%"5T!VJNhAhbzi15}`^{Y 4-fq2Q;> CBx{=[;uHW<q RDՙg3i|<#NjJ` l| mQ*p|vUTR1,gx8 }fRz !:'Y-,9gR9,g"ȡ\`\qJ;\&}@) &j%Hΰf& F4(NV˔eQFJCqf]R)`W!4[퉢]FSEn'EzL XhO x}~,ТWNV\T[RhtTNX {@DXE|&ZYs>@T̫ Ңĥ dY ښ{aDt GI!'+#b]Y8JnrK u *_Y}^V< Yug-՟ !L}ilfDᤔC©ڧ)]z=@>Ͱ@is4&( k`KP7m: o6ܲ;N o/T6qM E5puu`޽jdM"DɆ`_ / m\ 6RY&_l*̯sOČ#dɅ߬,+DZ[PuH.Z|}6m(}y{ĴљG}ˑzUƘ^~](!ލ[a^66r`@͛ 302lb$C|B/;g%4'Ygp%!DPA1GsA{Di8ut}%p|^n/WtEn~+e%CyowyH {5.-HxѧM HB{Phm?IjϽS%UG\3=_}" D[~_}f̷ߩ8>2Y}.>NH!H)7ٙU8r ߣ(h"P] ;#α]u\}CږcPLJN;t`vz](J16 15S?Q3lS!;Ɠ=f]m1xt^`1RTTB4ư0p\;@ [ <mQo==:+ţˣoDm1|pR|{qN@9- hkzPIĆ73qt<S޶daˑ㹋cVjb>tL R{cyqc0[) Ü)B|Cn**dpmcV[l]A|\MSh4v)6`@ڤZ ɤ) )fr),Wզ2O XIʢ2df 3idsSL̩JsNǘ %S}"Ƙ Xa[%52ȑIR1"y԰՝=R©‚Wπa[4z[ [VYiД J7;/G#{ظ;ߴ|y %耝./k{@=xګɟf᪾p쾛9j`ߨ77Y:lj7y_{,c̓li]7`2oGY4-b̰Je"Sa&l;Ҵ%K؜`jCr! Ɍv Y#CMn-p M%r*)wB99XH*,yM*,,S+&hvSM1_SHT`Lgl3s뗼6n-e$>xdB]Iq6p/?snNCf".5b/Gڛ\3O_A`ޒ (DĬkD*(Cqŭ67k5 ?WsИMqrWW}oW  ű䍅w3V@b<3E(?x ^\j\NgKeFXU+Q A*q/Βnߚ0P1;)N"$i|G]cF2#L8MRATp#Z6e~> %b8~s>՜2$QF3(t`Ur΅Hˤչ%iNRͥ*W @4;5XAq ؃DbQ`ᚰ4yf,. bGr]-SX SLg]n\~݃ LK77_|]>} AݷOPPާn W!,OɁ?~%l~LzwWp2$]SB-GTd?ݦ6ɍaxb_66A[Ä5/Vؔ)M+ }uۭ?'Pz=-hg[SX(Ճ!z JB$=;$&R!is !◫oeTGPكR,z8!BjCc``ܙ I䃏{_W<8,3T龸8 z&vKL9b]9PTG9{"we\2JfH[b?ޚ-0Ҭ*l D{)8uۤe? f2K|bi '4%B)7c_f)76sբ^TUp|ўF)rAV7 Fla#Rf?Ϩ $0*D&C")2dEM6s8ˈ b%.s$y?$wO*[ѯ`~Qdf#"ݓ:LU3+7SD9oGɑv_59N/f$ NbkB4!Yga$ #8Ǔ xDbIlRLvQK*¹o(~Ӭ8n%I9h3cI,uj<d"2L꧶IMKJ-Vgz2-,\331:s(L5ꀂܒJuf6#!Yb*-8_@Xnt*Br>?-?zo+I5Z >NERT"^9R|_x ƘST$XH]4]6WTv/K W͇s*NS\ Zk$E 4 $k\# ln4uq.#"8\ʈ 5%f>3BUN(q`Uk”LHvrڟK?kqsRGaZ$vRbn R(%vo"R rXc,hGc͍bQk9Er,صbKkaay|@$_JׯE>_Xk>[UUǕ^ŸFKKݫX`:/p/ַޣ}?i0iksОx>r 5~% 7ؚ{غ7>5%tIa/!=nԺCWBRmKDM@+s~iZwhN-z*(52r>N/w0v~;a,sTh(a {hHy%u:]D&(}R1>BK ^.^IǫGN^M%rL%Q|%e4v 9 :%'년BQQ,hXUX|`!]DB+{%ǿ9cVBˏuxkr:Azm)+=9hWwDmTcct.CDP|g ֻ]p'WMNA~Y}mN̿k7@K\|"\{& GY\1mX<\]6MSi˥VҒ6)yQE9čf́,\;`z3HGrϫx4P&Fݷw!Zdqt3u;ѣq8eXJ`f/Ď&V6EԪIC `q3Z%tޒA]F!)?)dsfr6vm]c5-n .hvk+QEjI?if3SA;xs#syى&NFƘHBFZX#Dq7g% v؇vf'b9¢~9~p&7-j8-Gůz&3^#'z΁pBNz z%B5ȣOH&"U0q 0ȉg}0Awp{o}87g_;8G?EIMyN??-tO &[̜ sɍ\8K:qjqx57j9[-nNr">|#zѨ7oHK=?mn_nb]_-.-hF4zd^Hݱ7/s6=7lgur fÀ]ʲc$P(qH3SũO\YݒґbEbMЄ`!`GG ۵u=ɻ$ެ&,6P<؈buP3v]io~>ja6[Wݙ+T+bYRlKNbIT[D,A+@:A_ T=xb*NDf^\)f[rZx.*?l +XGˏ5`-3T={&!V G>1xǂJ,Sk2ٺf3/k˃ldV~ɖfhFm9{W.N"Goϴ[ 5WJWX "߀v{'OyNG - SoJcbB)IT0!z@9C 7JL@Xr'!KB՜~ 䋸/(:f_FL'_%>gthxkb`c972σ$<ogV>|-\,Z7f{ :B2Yp󈆈!iMJ4rqk,TV,i]/k7칌l{boZ~zng|kM(Jʟr2!JB;*<|s ~EB kVUPCN?͜͠V"¥5&byl`4 0!4<!#4z<$w }Յ!YL&O}x`rAx/?PڵrT_68j.yxF5+T`fbC/.S;nY',Iޠs4)nZ*.30࿿lS_cZn[}n<wn"pQZ-nvkBBqݗ)yu};Ϣpz˜8ݏrҳ .t~pΎjF퓚ߎGշ.]p򞙈 ,$E1&k NXowϹ Zdx9)=N {uҗ(PI)NrJz!Ǩ~8CW)}R(|ݫ18_^t3fC7!Xik`Z3"S}L@/#D[AZcy/'%1½qZD'}j?޻:89W YLuUtYϹr*a..ѭ_!9ߜB?9d4rUn)ݫ^yQ+lHNX#kt[&qYe_b,[4 * |_}e(Ti:['4r*svO쑍bјx 5^f٦uy Y*3(RrgB&xMC@x;YY6H$zM@?1 DDe=[+gjj~W*, EE0Aŋd.ϼ4yFb2wJb10yj(\l%X6@)d]S(PDHjlaPqUOVD_U)BxKu`JJPفMG!{ x6}^Pk(`bŧc51w>’:[.FwOp.s,عn.77Me3u XML&Q<6Yf6=a::D;m^&Ե/Adsxg`f>+cj5^FfB${F;ullNېoY$exlH?=ÃIHhSAeBK9n6H<V'#d^'j6PG"2"H0FqPwl:^iB-_ldz 3 !%Q$Ѓu$Y4#u=os;ļBRMYroZ$CKv5^}4"|v ލZ!*EK^a']m^Ll A}ihi[*j=zz2g8 `&v;mVUT¦!*~C{Yu!=VBskbw!S.6_N8d^&ԘłMLi/D.wfRHSDԚq TXq,S%)bb~v {9GW[A qf9iqo%i,4A&Qh(ʲ;I Wtj%Mŧ<ٮSqZj^?=HV(Pv3b87(E6d:zt.gXpT<؟o |\j<M>}Ǿ l=xaάJQ`֥Q x|o5&;Xxʣec% KY-Ƿރn[6(&~"JD21&~Q2-rZE|%OJ ͤ^H1Z?%K)y12)ŸRzCRH)$q~>hխZ-^#޹yK!QgդV?9ufq$\AWvo.@eY2h !k,4N K)o{bP[qAz[F1,HĬ!/x'D!Ψ8"(& AR)n>; rv0cź-^1: #![:>bs6XQ/9 8V rƟ`+l6!:}d')ƵH1<Պ ̅*I(3'bp.޿=֋#!=suٻ8n$Wf*V,c}_.:pA`ڕ-e$,g$i' iuba=FrQN2+2Pa02<pTR>ΜD@_9z#T"Y]X?bFc9/ǫ< h+17Uи ploO5J\G,)]à9R9<i$x=sIQ&Aۺv]a{p[NTgZݒٰ֜87麵~.>mm-v|P<&R`3WlXi5gŌT 3aIe.:wv|AI%$D-WZ&Lp L}3`!ytȑ+6A}U0h6E;פٔh6>$h(:FŁ4D={z/1(,rQNuN0>??^墠Bw\}]O͔_4Eӻ9I jhgtCz!0f7^'h EI{l _o~!@F<]o詄dQg׸1?;cNduN9_M֯3i;"%98cq#dyJpXI=NΉLN1eaӬ+ژܶ)Ҵf6?i2fEuP堝yZlUA5@=L8-LOH 90p>퓻 12La'cC ;uPTdLlvw[]9[ϑ?UsOn#z/qZ?!-0f+ppj-ݳ@ -LE𷊞TS'5Ej<*-ܸ|vyňg.byVƻˌ**|e'!)7"vylgU?l/~fC>!'W<<,oܒZ:xhed# xZ͐l?3qe:}0cIt\ŀ).A`I1a'; b1W.W>-x,l%e(6 Wc ~Vn ޲n ޓ/ TsnBpePZf.+ʼ5Va^mQ>7E谵Yۻqd7@, ć¬J?ى831M幥' 8^l㎗!ss1ҵ{tṋhY9ܽ?聬`$q0MX ]KRp ݾ;e e6i/⇛ ql!ǭAekpn?993W% {R~W/>\/>5}nUq @eU-(3OE UEY׮Cpo<^>}]: ׫ooD^Sxܵƒ`[ksU^PVBDUM8ɲy"$g'pYuo6fj͖esIR4q^HrRTvSzkRlJMl.K9rf׭^ʑ+~#5)i.^z~^j"w:HZt=K,\w;P3lJKjX"c~-N>#:p\ltDjL0 \mnD&3jf!yx$4k~ >Bhm7G#Z6!1 K:o?'k!WY— #,j%<%36[D5cuk(RV,sqM)!9TH}8rIn-6okW?Uy~jvJW"$GWT?4rq#~3ub.@I"v2s "tIT /C)\ejrJQv=A;<y&Ri٤,TQ mg͜zKaZTc6cްi"1ssc5 v* 3`* K.dJmsK%YQ%=P*RCɠV-0qŁFjc6z_\VБ@$TC6Rj4.^z^JV $" Rx9z)Gf2zR$'4gFh7ވF4WCX$Oay.m-f.Q-q$%ƵnLpW<ĩXk湸́d;8:dؒ&=!$yZ̫^d~RRbP, ~p`o\H2t[_R!0^HeJ@K,%b?8IZCgpVֱYQ;vaȾR[WHnUi X奄 k#kOYϺ{YWV^XoAF%ҚކeXECQږ}I&9a*u9QQ* 0#Um1"2ѤTA4{A_7!>YN92ˀE;cmAY&eYa5RuYc_dBLu]Ԩ22$Rbwq%ɱ!͔dfJj6f^Cu6ȆX\(m6~9l_U^GsWʌE[㺋gOvw١RK6[?7qnzu`,_t'jBaqw !Ȝn~8h'}^: |aF,v3K ժUż`j"Nwο_+Q7a×pe5U}9_]5ntkf_?S>C7q Y?0TȆbtyqjKJwt>_,-PϿyO :40nB'XD4(R!9kq,'ċkMHWݱ.{nc<`dׅKHBY80#[WǧůMR-O׻%p"ק_Oﯾ0_ Zu+zbX-n\cgmYi.;Wa޶Ft9 \.dVWy KJ 1UQ"3Td̋ˡ* wΕV`G4Q;tYB#­S.gU慖bV(XgrVg@i _ՔefK.RnDТqA7Ee\YDžeX rɝ*1Һ,}q0("8-vI4 x]Mۍ^/ %y)sZw;OϽn XǺH\<8Q{{s+ ?!"ԶXy g*@{+U"<䐛Rga2-DnnޥU# 3vcq 9ې z}4g !^PXL=VRx6NI{4z̃]}IӵMGt烑$|v!XcЍcY}%cUqe#I^m؎6ccYpBd/~l.m"x+n6;cX˫㱶V|q;}y8?Z7!q:F"u$Я}qEӯvQ-xOpw*'4D7اfqؐe+fj,>fo~Meće`Gh~ӣ[qCYĶ؉16^[n>778/G) 쬐Q!2j^#apsFS]C{L ;Nfә67%i:Do@H֡MHfL#ܹaR"]\^I۹7Rk0xfSjZpBRMئv ZZ rA j׿_9V>\4pE_%U**d]uo]W/8~}Q&*~M*$kF(.jƐcrݜeY':S˔#<;w.>l&RoE7gZŰF:qHn˾M-hѭ{_և|.ŧڇVS8Oݺ u~#%r!^ICGL#[OIO1q`BtmXg0xx""u(@t\w?~'5j3.TVhG۝UNx?Jƣ$P`)A7nK/ 2/yXd6m`i JU94}SIɶ Y:C\EWy(Y<.F(0F%_(t,ƊkgX[aO3@$# %G%uXC<3 Qo D O33b~}e_^USaI^̦7ŹWOܮ*%ts kbRBfR"^?5MBgH B @_C˺FK{k".D)dQTlhٹ\\k)r*%oH>IuJ+>=]JšV7dwJ2sSӇ֍hݺb:]ĺ{9NCɬ[:x-sS Z)<~Յg*m.b`-='}&G-}Z<nB0liAQVdoAΝK5yB*©(r,/ !*[E-Ѹ kB ܽ^撖{5jsNn@|z_.~\=dfwNQq2?voKy_"YZ3N/0zY}Ƴꗕuf7;폪^9L\v mw/6[ӯOrn)}Sى$ R;S@K ia]JAEg~FO'RnJ!DgJPJcD.6^!yU)Ho AZg'g+/=%'Go~0j!̚!9?WkXu ڄV;|Z~C^Kg_f{s=1U'%Q^s/VexfIwbjd/bNW c86uO`"KI$=.`0s?߫M&AD?yL `Y/V@s#3A[^]ߖyo<%B^B͸c gBh(m4MV9j`̤~7 9 `A Lb/0n /'>:s^h$-Za nGq,%ˉd)m j,˝ߢKB)4O‰'d F0AbIp!p67qCK VU{IR0dPrGQۧ/}{rYx80v\ʆ]?&V FS .e#UܳP*ۚ 2G|V5G_Jp̾:H'׌ܽ"@ 'S}]/_çr0m哴&:'X׭}g9=.F& ^6䴽|:ȩ=ܹI9aR+1A.OŅޖg %n]/5K<F=*JwF):*]O0JM~caPDu#D@9~$2+,P /P8XBH ʨJé}؝Gepq%yJ+.I|NRSd 7ܤr>NZPDR/YЬ7!4  xx NYvzRGڃPxVmwP;A+o7lxuA@ț yo?-p MlN$e$"`u]Ǚ33߻ɫ(De 2Y]r@tڃnlGizطFƚ1$:YhD*VK#`QXt]Z-s(l vha}V[ F( &#>;}K`)0 -`/ƈ. ʧN)Wl*~>sf3N)PZõfT1x}]/f(7OTf}RW%/~=d,{|zig3IiA+2 I*B2+ X{*L0R2@(#{Rc6Fjۮyu Ӌ|sWmͰUysG'S_.M2O]ݻur9޼ޓaSLj-ϺcDžzx\WpVj1ArskluNjeWOn~s3e\. j zxո֛Ob SysԼLBK-2 A^tR<1 pEjq{}mwmY)v*wZT $ & %0؁I4Qx!UѼFJ1Ї/e7HE܂T%`N_,z+nًW33CG9HA%K !$QR>PUS;uS S þ)16cԭf4YIBO=S#uLPF Arݱ]`qauKZK.I.= :I9aR Uv?9?mg qg<sI >;cLyt-JII_HN˧SfL>tL>T!^gK[xtH-MήJ.j.X΅̆qӤYS!˽% N%Q dqQa=1*JLBpQREEItסaxQ" 4m E1L h(AՐ`5tRN#&Nj eS0 qkIw]aTʖeVSKl0T e3S斈‰6ճ-ҏm6~[k19a5 خos7D- ,}tSWpT/tѲt`&eC&@#7n3ڏb[Lpkm|fgX{,k{Av@^yKZmtt.Ո:I$[mm]3{rb0 2-Z"Xk{y!ߗ^Eֻlz@AQ,,ϰ²–RP[X'~Ҙ! ɞ5Tg/^^;cK:IBɻcH)}'2Ui_]!^;U 9h4~\/=1,"VѳVэJO>S,AJ55Go:7aO՗9~#uWH`:Gr ix2NE@stm$h64Hn6 SX-Cp[CB  CD`:ji*8ZPH^b]Rs(\0BpY.21X*gzH_!XGe۳vgfhi% Iuw}#IJ,^R"m -Gfd|`|$Lhe4aFԂ{.K?> .3&W:7_P!Ǖ)hUʒ)O1~Eɒz|9w&"&+«:o]d6_t:p; E&6}Ӝ+Qj47򎍷Z1 f*jtfOww8Vء)Q\r7 pW?fANH~j.Y U"hPo8ч--QiuUWͦSAP^n.n?0]8=Fwӱl{Uh`S0GG pJ΅Cv?oAH;17ᐅןG2ߏs6z*!ziyF''ϸtvxxw}MQ_?ލNG 0~ݓr?#|%qWWrM'&VA$ J[) nK`jmJgw9hBjLQ4P)1gwِ/RL?ZZ2eR5K@'#g->Z] tІIC.|pDYc" 9T@4O^oj nC7l W&Љ3V 9'2u@ɮwdG0M#|`j!VbRCtV\F&,:^4ZJw>CӐJBb"NKQѽh5HWAQLWvAKc O!E? 0&td0u KFs6=7֙)F"ΜQ2#6hŒqat iJHf>0r՞jm7Oe?11&QvxQh(եۏ ۏ-ITv4'QI嗴{1'Bf,~ĄT- (c׹PJWV9o.+*wP=ƌQ}t`ʊ#Vdņk +^(+ו_0 d'F)R$) |zL$G5?Yv|#<ְqYBLRZwpXc@ N㝏qrQ1"(9m۩yi(.tSaxp(~zX|dv]AO%C%߮=W܌͊ғ9kd1\Yk`0ya>//+bw\0О Luʃ&m@kb0Yк!߸vҩZyuv֭#T;Ye(S#n]'ZlzАo\Ewt:ϫɬNM0=.7 Q@Bήw~uO(ډ^{k)ύPЧm]k|V{gf3"w aՇ>^.Z 'i%P_<uV5vOHj=%N"5dP$//&p'۔tfg]'L{:^e˚8LK72''뼢vgND{gݢxs[hC^ݹ1`:QlC0 v>@.Ҡ~JY.JhIB +5luOq^ k9ΛɈw'n&V̈́/<1@̱^ l䵐r,jS4LEİj I3MЫz[4WdA*RGCB>fK+<'M$90qІ J[9EJp) đJەmٹGPD`Vplg Lec׺v|isۚ*-Z3 EAåVdrQD%Q~GfJNZ2G Lj{Ҍ^?դ~޾9Me I;Iu$iT]^~a/h@b;k[, c}~lKcm"Ud _>?.1-L)k-˺YHBgI|DW`4KlrF$gb,%48J0n7Tu]R7^oM˚J7e2ԠޚJTcL-5g2N>Mk\6/~XS'f{#+VTھ<jFkP?.7ԿLnBz5ʒi=ܹcKf}ѩJ{+1 `GGU۝)H+5j$g(*(ёw|%ZΓu"A%re^Sk{R&dx22Lz7T6fA.7Tezjh7L&F@%sPNM!fL$;n]Jjg?R"4{RGQ*c3x$\IyΨ) Jݳlзwv]x "yij!2 !Y|r̛(AH 680.~aLV~l "*Snx2,\A6YiKB$*FkFQFq26j'~Ԩ5c_Q;)_jx8Gt ԢhcC8FJȘT rM{6! (ԝðqzIAL)lT!lڋj(fBh>H2ظ;1]`{l&Zƭ%I ]Hg ތ=ǒ`;9/ yN*$ fK͓fs44h+*֫@W![pvoEAEAL~C?]Q.Y%+x>ڊl 0S\y@rV·Ma1U5F /w|K|c p2TV(0f ԵD,+ )`!y)p} T&љO,k,q]PrvǙ3"0c GP~V>AJW{/{|VSjM&7?8|rè<"f=ZD6KA53 sej}@IK~H榼ɍ†Lphۛ WdrD)VF#pR< Z3)&^7>%Z[|2ƅ'cgIb4{y$  xNNxIbgnމR~0,1jX%,Jtn5Y]-.q]&9"Qigx#C* ~8P!?J:P (zj\4s(3D\i1CF2Di%-I$kx`:h T%# 1Hyڒ;T "pK;?Let qKАo\E;$mQ# ;٢/k'G:bD4WN:G{֍ֳsnyP;Qv,BP>u<ح Ut[Da c" 瞁e[/ ,`@(lKhFa0砺]٧(DZS ڍU @T BdOXj KxZ"9~ DŬCІERU?:a :w;b<9Bq~̹D;9הpj0+10v 5S6M/`Z^}xVEuN*?.. 駙-H! hj+4(.v,8%&rnךi׎Iq|)Q"/J W4b~^?|Fç*ć|U=?~R|KgZCW2fZ3Ja=J@ǤѠII3!QN2jLʂ ^=Tj z |~[}Hs<Ϻ.QTvlhs5\좥 z\vp; E?Kf9s3>A/vz`B4(䂥 lM&OsY:t.zOӒ6I q>M}E >cLj96ZI-+/8:8W_s.咔O_ۦu.Fru醃 sb`acf7ۙɥIM}z{Մ,) ;@Sl,UԏɷB TۖaI, uE 2qLŎ9Xh(1TDIĉ4RJfu|_;k2jj&i[_au fCQ;-ZeX)& }&sÝK8S:"DZ"5Qd9kuB%C2Q%(NluQDAWn]%q)%+Ev y8χʠCV?z7W`zv٬&jԞ&{{?u/y 'IQ6Wd>l}Dݪ3suwi;IjY1`?RY?b~G(Ugm=^Aܦ]U8K_v[Y]sɀMq]$GƑZ< qD)J]ATNTWfv8ս5¥¾DM7AIBJN5ls^R&Tz)%gKN5祊^@pBY翝KMG|{*vV q> k>IU#QrIhIG>n+J֤?q\vYUhT.^^_Ya}#@JN*)l)UR:ZaMhp'kϏ2ߞN~ًT֩&:s8*'ROZhNJXHD}9Ͻh Z6(jOպ !J;&鿺-ؼ@=Iw~Ie%!D.1RxQ Qw7 ݵLT)rTm65}IׁBQK:p1[֘~go(3p!g\[8n3dh0 &1Q~b{7vv|~=w u74<`,~H~# &K o857}OaodOWpD}wA>oLNlwPCZQZӶ.fuµ|緁9bmh^Aj9LG+)pom jVX5+o6i#Hk_?)偪5[ep(D4Vm(1R3HMAlb· &\^rknDhr]uq)v M4YsƮ<>)E$ƴ$8ijƩԆ)/}g9).R=eGmy*6$<{3MoM0YqdVVӱ"$0cR2B+O,X5P(n \1."5#WE6ޠ(粑", /SRNK\CJƝ_"rr]>H%'*cMbN4qI!8Ҝ3tpU8Vմ\'h:2L0`CPn#cs@iiH$:KfQl9yׯi!`bѶoW`9 ~7 N\$SjI]=Т_ >tO.O|rYI+|;qOȎ}~k4'—"& 9K4) >u3u~/,ӽqS{Yv׮_F]xb告?_9r6wn~ sW; ܹ;*uG'G 4{ΤP8.4YvK= [:v0'3\|[>w)8gS [ȭZ>W#O>oNTfjp olYPg߿R#ݳ zT?~l%Go+jjchqr6EݬO?Q r=/Uײ[g?AqKQYao׼+}oSc e7S՟[@߭LL}+}1gji4;=n|1dEnG?f?SPqѫx9̯8?.ˬ!G2?d )O?feɾz΂g*̴gE>~MX8Ss/f~΍ϫG[{3%GpƜy!ťsnjդDДmf0oYG< )i-%~NM=^fJXfQo'd֖rv_Nlc{Z.ۿzb:/P4B Q1 .46B:C"%pDf`p I8G*)]"p[#魟40[u7rAx0}oc9c^?1Qȇ={6~U4hiK޹NDDvYI֯<Of$/Цg@Hą/bx5|w& Moj L6iVi@RAOtAZQEX_ݎF.߮M@>GIp}d]>}"3n|)"7 >ڑde{A}q}n@]Y@)LEZv-:sYzx ݼ ߳*l0tRK,)iL\ot~ /! w7g9Xn֦hw~q&}68aS=ql33a7i짇VxN]gW[ُ3[sD)Rnv0++MgYƣt*vD/$o*0G27vi&rfM6 k `|@tZ 8-Mlͬܺ*CB8J@~e4",*J"N Ȁ@ 2Bd]W q ?w{.p%FB.WCh.:f)Rh5 Ht09^4NQ J/6H8vx^|ؿ_+,ku_gKNg,w  {@@2`H׬(ukVN5֬Y5+;H)Y&|Ӂ3#`2퇣ٍ_PmrmGk3́dpX[9&1K%2٧B& 7o5N=aԡ }aD6.Ƒ014"DҧBuF H c,%T@W,d:by:՜*R5 `؇BJN`^@gKO)IiND5RzRU<_,0)ͩ 7T&y*NJN5WH4RzRJX{O)1{֤/[J)RZq,0)ͩ洑 R&ä(YH)aRSͅ䍔^?^cmC2 u7>M'{ܬ :k85b9Fbim9hѱ@uARl͙T( :U$a153ӗvӖɲOvREy ;N?y/Y G&{A>HϤȭR͐woܣupX8ZFX5~S5s_ Dm~]G WH n+P%W-R@4*w<49Fd0ռSڋ5iGB:^FstE`DI9NŒBjU6w:Q Cn5t0CqD5ӉC;X+zhs,-kan ݶk U S.R+)Fq#'RGHPJ4N,Z(i$K/w;[\:f(D]6Ai4ѸHt,Wl9Ff~ @[ňR&ت!و-FIǰܘآpBPL&k8B&(cc_ LA$QiA0$=#IuڄiVL9aυ+eݑXƑ>ĭNPyus!$nE+6 kc2YZh)kyMN5I )rFV8j?JEDuԷٓ5w%^13ݳ$vޙgJƱ򗭚)Ҹzd2[S;=ՙ/;])cwK+ɤ_m$%G|wϊ=1{׏?'] F~L݅ ߽;^}Wd$('X6PRwXBໃ#!}=LD ]sS ρ`~QvӖ_ҟ``I` ">!ȤmbHaU'Pn\(O!0TK%ӻf)P0Tcګ!fTE cBBd`ZH eO !Jl8[yRrk xsSު_e#?2!\QVIdRԊ@N@qKڈtCjvZ1<F1I1{g9ٺ2i=Z WDIclQZcTFe2t01vUMVӝ装a=!)ˀu(K3CRUt<|:h~f5 P+H30yj("Nd=kBs!s~|шd {Asu!L4wKQI[s] ʸFƥ+ĕb`-\N Hf#?i+LOWcV bD(KWn2hCv<^z>Vb&xAބ#8Í6vCl=nu&`=["܃ۈ3vD b{ڮ02@5FqyX -ցНjb5vQKЪ LD20QY kܙjDp7;umR=S(0% Ĉ~'!D'AFs̱xxgx1G'= JI?[o9 Na#!{ D'v]Z&d bro~:D-d wP,u|0?f~dgE$w]+OQ U9)$ϜzvY6T&!S9I޸egľ#dwr:m6ywR̐|FͩȩZݍqwԉ}G6-8vxK- ΧhkNyb׷a"gye}:!#++[A&U¤( (=x䱃JTeWIڠةZ\VWxOj,TK_],ńKv1 !.v`ugC'{er:%THLI.4FK`.Ex @O8 ?cyI'(3)NcUGcOqFWo>e%eaHK ݈JYQlT!K`eI0!RɨA[AbbRRF9r%6TX1s.%LL+gąAL*(r{.M9t>rdZk0SE4t廜>TVDTؾ c+$PTBHYȑcD7M >0}zjWr~,ꄙUMz|\:Ig}T-7f\>z秿_5,GOW>n@,8kxk?^MbVO5|͕+'ZW_֐r%ÃE! $ӓ7Ę  d\MrP)"jsjN]ZS+A^1^h2 p-AH#\3K|?B 2ڛ~|5ș9]1~Qkoi5x/~?s;Lsيzy٫QmԶVg\G>Y, ao| N$rO2уd7֨MXzsk'wp e4Q^&w5ʿ-^nM>(Egv}g%՜NV嗻t<Y+WU e'kHǹ-u-~PM ~f[?jh0Gv?sCf|Z6ܲl>4z~7bChp]e4[#ϕv!mpωs(|47T}ފ0 QGa%S҈B+0dHaYa%.M She-<$, ArĀ6[` \Rk֬nF' lvS$,gbpO gy4ǐmv?{x xqYaiwC}椽 'QDbb5ea&NTim@P0Ik4ޢa0pv|BS<1;pAGr@Bb0y^ɍṐEnMi7ui5zo&M%prӠ% lo2I$Dշ]sL L8L/%RϷ.<,uW”.έCh-AI$JJmXDXs) +=/<|]%vTjmDF 2k܂LcSqJn=|yk,vuPx5{f~d-S5=><[ZmN4kHIFVwm]-C,uh xm_141j2z G_ T6B4;ހL^.v|gL:bS8Jd%i*M1dT`Yʰ1<IeOClx|^ծUQRm( (.c^#pHhyIǞFksaO2*+f~o.~`bޤ-2 g_˾zH7_6QKz3L 9 >gL6AD Q)AbJŎ4/61S$'ÞOL"$-8>s'IyŔPqj&Hʒ1eJ (h\ʐBn(=A\z2ERIFBqQrAD (8FH(HO #끽 >8 Og#(ky.$L0!:^Ii3@ۛA>žf*U+Ht͎V1I^LAf|O8|= spBHaiX eY(h4FP( v0חb" 撰i$9CPYibQ (IoCZ~ru#[<`S9=0AropA '[Ixhݷju5["#I5FF]&@LG^%6 AN'uUY3$T|nj-B/VR*`A G@y[=ʄARlk& |AC3v_Yh=hl2GsHJ9_*V1{fY,]A]wNOkm B/\5y+~1S\NF>a<ήOjzbq9{\\Y}\\_Cs`oQ[U:M6Gn{(=c7;'Džw.VY1-t@q3{꧚=Wcw?|s$r=";* 3JYQJe?.T+Uųkb6{fPMSȳGCELj;hajߵES00v!{qY8.D6 Fc 踭؞Hw&=9iJ'W籽:ylQ{Lg.`e V椚O?Odm6^]}>34# mfh [;r=_QӘ7N: 0kbePK0H weF녳ٗ~2Lڮ>FR\7U.Y:Ubq1︷Z rF[kXڢr°+t&Pţ71^+o ( 8t<ժB"͆Nv=᭣ )iL°v@T @Z鄥VݶB'[@tka-lU+7HgBGe0+e$Jr|`C WB0#6a 8 eC8tYex8X*[Y%\I#@N5 ȫbi++^2O"˦dP(+^&A31޶ߺRtQOSaDi}Vb nTceuynEX 4 bK|{u.p8d*g;5yu&UZ}]gRpuM^7IjqX#׬2$>8cs{ɔ ;l#umMѪֽ$3ڿ3Jϖg?/E(pu?gRtWWYt<~>q;{wmY~X""!ܭ{%Ѽ0Bq/s:j"h::pF3]HumWU/oFPB 0yAjmMul:2AvL!T&$Yt  `')kd]@%USZ %#'"K 'fEL "jJt՗e/bxn9_6?v(tL؞)`:a.;w48fm{_=U Tr"drU'-oQ6R2ac%d!EϦ9,~5/[wR2Y!ovʾz-},\ڄ:y>x~u/'fKoܧ4T1ԉ8S.^^ɕ]8pթ\Շ_zZyBPZ@/N},\ZPP;&ɒ@*QIe٪}*h8DK%,.y׿6׋̔v~_%"?H}\/?7$,-mLLmID#HX@!Qt}LAHtඥ[Ϸg?\\|a35EoV3"vo"oW;ϮyEP<-W7k,ԽI+Cl eB>02&0JiF`9B@T'ɼUb32z3oH 5ʧ'C 1Fj(12n|;7U8KL1rJ"g32EOCL:Ÿ)+ xBf"{-C,V42ֆ |fO3MVʛAJ+x&a =qf;zLP-ϕ 3C[I RIY]Iv[֪4@q U]2b0>pLh=1᷌t.2BJuew*KEcSL ul딒x!Q* ]tMݽƊKƲ^[vBA+`Pe Ahʪ쩀 N`c= Pcv#̖ Z01YSKZcNךRWv:(nr|rvsy%˯zM,ߎ#WůXWMsWԙ~{]zl rv- 𦋲?X+̼⼉K++Y+yP^yx%*yEWρ^ h^^q%/W<WT V[Bo>^Lg md wAl>qp'Q@Ⱥ&͜ y_~\2P$zHPSR K`2Td+5#l[óI&b5k`*yXx5IqTcT³*`A1y嬟æ;u8lǎ cG3jkqOkf6U|l}3dc23 W?-+cibU 0&ieJ8P:3yn՚Re u5hό`3#pfWv!(ox]ŧo*[ٵ,KCjkXl;jYB 5̔]/ pƛć5~IŤ'֭yoyԐC!6"6\:d} `o<8H^^37O F-FkT; 4'VŨp!OE\_g oPH`^ x^PZ'fzxVӔxՃxBMX7@XҘעp"L{i/5&%DǺmNǤICx$\Kf;5֤[=f.`0BVP2g:T$ j!֥t JB*j%ܕ<2(=E%0 `[6M3DR0/ ?8662] =Γ y6LCP6 ?UlZXFa9&A(U TF u fJ3ae4^n@z!AAsA]A}#(6_9|3xJQU;2#D2Q**߳+@cuQH,Z |8,JzA  3tM޽Zi*@,`9q!kB9j'/+Ke%Oqg.L;-IV(a3s<~Rju}A{_W6HdBO||/4Kg/{ U%4 %4^B|^I  _uX1o;"K:wMi'fgq:ﺳ{uf41p3"%Zʵ^GG%o,i8FpEi{r3՝yn(qVg_^,lU=r4WsֹW, ~< ""?6" KF/y<-iqeKKϠ%h6*NCI 6o¤hD@ޒ,Q: P;"_yt(4=6h. VZJx)( Ah %Uv\{SHX{'Dz]@`\ Ji6rQ9rw7j޹e\#zo11&S6u Rː )vr5}څ_4Af'Oߌ#NO&TIZPAx˶*~kk~ϟ?Mp8&!S曺-4`BdK>6dl,*X3~lˑBIy)SI;&\ؼll,rn}&ՙٝwɻPxqq Bǯ9)5<ذS2mI]?7yٍfk/7o e?ū:]ğ >.yaEaT6O\uǶmu%MB͍ mZ`g}صae|Flq+ D-6B`./Kv-)opc4y u_=udj\-D;IDя)nciEr |f|kuY=.aBaX71jdD8R|Şn6C%҄0=@a$?5J.UoG8Gq=V T*6^1D8avO=n$)%95&iӐѪZXYyǪG<=OXY'`x<%W(-PH+兀1MʪKKR2n=}ȸm=69P1_vuDI|-:\NDٰI8)Ĥɘ}he.TI9j"5]KZW:DRRxD]D(u:q׸ Z\Wx^{$'wkk8Ľ'՜4ۂi377 Z#PvڤmvOɮxl1܆0FyfOZ13$bncרz>Owo lDm.kȁ?٫(/()2jV3V{J6mC^Ŧ"`nG{H%:1+B!)Q?]sH>z^}anZX(4CD$7` rIc?QǨ\[`UP-[@ av@H`߲o6BPΫSU7՚KKpfH+IN]W-^DR,}ޕ6r$Beǀр3{k4hILno$ICT"[4D*ˈȌ. tr7XfC#ss&5bllgb:ͷΤQoFi/ q/_Xf uӽֳA~?_\>.w\|{l/ @dgODN 5JgODr96_#Z›רē,+d4'ٟVIS 676ؼf<`H=$TaTSa/&VD3ZUJB h1UĂQZbӬWLg/Q³ ɤJ'ۑ*b@+YLȐRv }G?#oBU:ăX̵Zhv ,&Iګz=C9d\bRhkY='FhmpZ 5\޼abtH^bE3}LRo_w:b״1*l@W\ ywT7F 1.uU#p>'y I 0-Vn>禊_?~6Jkf[U_x $ 7XJ1Irө<y}OUtcuAJ1ĊI#^vV!E|dCdIJ \SN8qm6] R8{{|6ZJ#!$ F2 a46ϭ00U!)ݺ`I7jHRhsSD{QEEN*ݏW(6+Y~ڕerCß~?}<}Ż LHt_6FKɖwYȵ ]|ph_|;Xϣ/{ lP]ՆeyW8J*"J&6sdGui9=דSzN=<1?؇k XUWv' #[܉gmu[dLw3 fʼ2a@XFJ9$BEVҘBJLKnϼ/)?~˰E3sJtj7LԿ:3[ͯ?3o*V+@u3e]k)QgJJ+Aș8{q&9.BjJyf:a TZr6W$) }N3k&RVD_,P/YPH06)*r)E >Z[OK-HP: 'C^ W!]4=$Dq7?.fsd>61dz=9v,/sÊd7חȽ[>]޻/uln\m,m&;;u#:K\J'~6[Pj/*Q u@!JLFRnWHJy''Б@/p@'GBA 䯮e$ϭj 80(RV2r=Vx|ħW[Xxnd[8͓R@Br4-@EҡȬGpWJ &H  E#W"t-ފ=WPg|\pfA^*=K(KF+i%ʵMѬ#x7tAђ,}9&WL׾Ikl093#ahI*$J`hTEG!ң*a o!M܇m9դ.܇ v*ؙ?vj0L܇}SnKm<]#mAlotg>o#]{i/>sK1ŌoV8-|\Bl+ gCp]}|T0ID#4v-$L0}R*Uۜk <* A#LYWff9\nx2߂gw{k>h"n+-*ss\]N23Zq;ucw/ gMS>:AVJT/5[+%+>7 3>$E`5G k4Vq+86&FUB L[:Ȭ5L,EoddpFʿ0xк313Q*RzԌ GRHqD&F-<6F Q+(!Z4]jrVXV. Fs.$ʀDu؛(P zcI֢Ctgۘj"nd6F[n'WйSVꁿ 4 Lѓ) 2ϼoƹ?C3>5}ȎsC;0!a hVcr ?u?$ْvs)/}\,=`dC^.m/oGESUhpt7_s-H 1ӇWUWjU*(ï7έ`q4R*H"&Gagd"e&!,R+bq9d萨sn7Qdfصplc ,J#"M#A .7f+EE8mD@#UH3GBjlR1F&R$-&%<sV;fFT5ibIK>Ђ}}+{ s7 ^ *hq$xN֥A.&\ALCq\`5)v3{L]K7|"*:tV9"Jؕwه}$NI+dŽ g^r(w ⻲{:% EYPxt"; KKO4Bv -] /)>TLcO4z۩ ڨEJKCV6JsPucbڄTH8ѷ*4w٠Xtc2`%{(Iqx J $F:XzZ"2&\'XP! ^Ε|}A0ΩP"Na%iA\ך.{lP e 2Fh :dhXGM[>nw+]qD{_'g:~M ]D;;s<}ʩGCU$(%Z(@HpE ( [=ЗK z+xN)b $&GBs# O041G^Hk`~'-/I0Ю᧌egK=w052k1[H&O)" a Au"}+ŞyrthA ddrH1LrN:.j(5Ooe:8D;+U?SzbZDA3YIiIMO:E>LK|j,CQspnkT:X/yk!wFphcFFct8-RSaF!e=DD ej}z K=]F$˅暘JQh]mC0ݑb ##Cz#b#*yQooSC>6'*Kݐ# P =US!FZn XU aX잾=t D6# ,Qٮ!Q,p'*#E9Sd(*EZ|nu9̻J聪R5 џXPRi j+Y=J-8i5R{7JxEںvRnt;JD;=d^M(2xRid)FKEˆ sG_S:F>T24[NE_\q ӕ0I@l,ظ1\v~%=g54b#''.^OQYzKpgjY\a|bH+I$xfݕpL|i_ozV: 3t?^-~dv(,&)Γ0hkpyg됼,Hh5;ޑ9-\JB,}y1u$|Nj)(0o?]&!B?ٸpB'3YN‰Hl%aqѠ'3G)5dGY6]WEŬHcTfʩhRAuV"F$Fz(и#n:g Rᅳfk":I&<3r dcKLo6v"DJEpV1s~FOȡ|o[gxk;\y.߭ҧwWl4eiA'!kcl.hZsQ46f68Vw>TWap=4 35x {$zo@OW_$荑al-8-/f7w܉~Vٯ?ݨۯ혩w^ycy ɵÏryncnAl^͊?^\,gן^a pwyZ#譭by/mkmkw^ޤNH(ikmZJX owV3Y4yMե8F};8`:μy ؂zoFAnU땺R;--iK/΃^ymJ^>%B: !5cM6ښN"622_< YQB(*RV9\FU+Q4dciȧNz{'ޡl?;3c:如lex,cǣ\|4¸×wIg;Ol& xI&&i4ptbtv<zXح"l?iq$[.wP]'}S*]$Qk5kNJj!SV4n}=!'kj8rǧ}8=N=crxRknvCK( y;Tg 6FcF0vi09t,[c"+ѻxԧo{%j]'/J4\dY /[D騥Լ9sYA:'s&198Z7|X d#WwHTwpw9d9kj$t{5qʌlgxP^֯珠)(y m(XU euܮXz75_C{JK7k0Iqaݨ'zw}֜XrlDzG)+Q2xo gOnӄ"!Ck˿DO=416muBs/m̲_hጎ,XJ} }zr܀ H5> #>4͐= IѦT4h{U <)L2lbӵWW+jnZjzTR( (iRv[IiM5u1+aw8͛֌T]v>b!rE 8bE竧#,6ߌ71Wezku:qִb&"X^/ږh˟lD`˵C#b#bƊpLE{#܊JDd*`RA霞56}ccQ Kְk )xh |jwTQCAnNnzEc fiLr?kN "K=/e'mk$ėhP7r`^2kFF-xm>0G 2bo]V 4$a2D0e;!}Z@ J(V ƒAHCjݻ+9@TuBp PRǬc"r<jډr-~ȀNL[* 2 t9 pi09|޼׾w? 7e(EOzL#'1y(y=BNh:l-U4m#g=y~5b84ձ iP8'ҦoD%6:yJDͿO'wB:*!4bX?![ֹclė7%;-3nJɒj)[7eD@̉ J b'@!eȩjwjX]dOM{*%W!7 6[#> (:iʄ2Yy*uPίVz_U~p<h=EźoɁ!cCxʨ&WJZȁ,!FyV]>}jN@ksjA]^E][t&YEpn[aQGK&FCQu();솂a~XxT\-f߮^d3e9d׮R k*%2BE+b HDj-DL3,ŘėշŗGV6c .~NA:x+!qrG Icq˛0g+N~k ,Ƕ]G* P$w^l0M{ʾ  DKHv"#1,f!:>s>Q,eB$AYcjXL#vG VDEru5/P*;Ŏ^e,lńJ *&ead(!T:4]2Uv2 솼lFiW+r#d 2 ,ׇ(hlEqYB,adW7l 9GFoNйJp={ 1ELЇO[>lCƕ"^4)8'JƸ_?^%Blߌ,\[O^tv>_Y;:=M VothNwo pʉ> /#82PDRS[DVQrVbeHf#?n &k|yhY THYeB)|ʜܧao0TFAGj` 8"V!)-E_Vӻ{> _NUlܝa&ܼ5ܳ^WF;TR$΅S=Yq|3uaj$8WV65WJv $hgSP#{:Fxʨa=ر l& +/v~S3{01 E[1"4QaTӆA˳HQˈ99/7FJ̯rt"۳ ~Qb.OP.9/EKE1fNb'PCt@{Jv=hzr{nx(뱲SuDˏ9E_^1Cufp"fEƚ5ՒM5G\8`zJp|ll -g.0ŢU,)d ptSWlBJ2MdNsZnn8]r/FFkn"ݷю9G]AT2$gtT&hk*3PH|>%x՝d#^o҇fOl{+U/cX\Mѿ4$gj%pF+"/ RWINyM|-"FeҹZ{X]{w[x!tƭ,Zvx?L^*#%'ŋ>[;oN*oKBYZMBaeO2 o ?eFHz80nK "fn@x'ڇ5RdV$5$10Q]|I&V 2 -X;U> =].RGWYDMB JmJBXX'`&ٶ|R`,~b$SFjT' ҔRD&{4$ 6V SR3LMbWr{ULvC$Bt[W 1נ*`BecFNV$Ô)yIPBi zab=P [,gDY/zR|[ݍEN3#v-'tٻ޸W<[U 0/{v>`uA]>qǙͯ_WՒLgΌ-%~$#EJ-lM)7 `pm[Y?Z(Ee I ʺ(p*jC@㒅5D(McXq'M]2 U<]<">ϭ2a]dY0\!%gITJ7Wa+nx|,z"@5\OG·AjR ȉ%7KMi;u1XI&&]`Ƣ6% MċZ}h&tu Tv%VkK<?@lQܰ[mz*Tte^\v+DSl@1lD6Qӯ7zPZ!y/68j05QowplW0  eu(hUYDV# zh*,K&?}ya+E>ؒC^3UlنjmzP9#P-KjĆ5^wXY[yj]s~0_<<˂iwϳ_޿+#Eq?ᅣhmX"7?˽<,>5?9GC"wE.a/u~OvEE,]-;E…L "V$m7X}4|L\>۝m|_sHqސfO]ḬaN Rڸ{^ Ԕ[{ҮW:k<8Qs* (QR; w*3tŚ=N8VED굿۪[HmJ.y rή)N*ך_ə9q 'K %9?Zem|ٗ/X]QHq*|Ucd[az'dI3ї;di,RS ?VVcg}վ&TةK)$[")Ctr|Z_iOO "' yXUW9j4“F0ΌQn C˯  gC,$.#ѐ?ɐc9H7i h/]YqtV;1ZLCll[ZՖ,x rZb1Vk jrt3IU!TJ8JP*&m2@+J򪻌6NڜN,o .>ZZdװo5i,6;^|Z.G$)NapCVmupת|J 2$B QԺbu]з N`' y{+,8S;c9ӧ˦5:0ΎP" .bF52E~3fYfC]6Uj連m9;Hj{+Ca]Z.dYN5#}㩁_7UǮ*ic1!XlQb,ҥBm98+&T (𤕒+k ~ej)~E2'A,[(Km_IoSnjxz7-Cg'ٝIJl ug(/WIf{R,1$bda?bFlq"OA:Gq-۪o<'=|e[۪eȬR(x{pkr/ѥUR294'֋ N\vhNXn'-x?޶KQyQšāSI9Ŋ*zOn`#+D4D O5j`iGw=[?Ժw\)lE'rJ[7؊z!pIFY#9) DMJNV b`Lanζgk7څloV[^@ڷ{& Zgɡ=|m/+n\L_nt=%_b S̥y9 .Ⱦu1Z8P*K@6l0;,y8Hn֏nd,ԀDz^ ECQñ5]8ZJ(TXU hKrl:3gZgLb'^'ߤq|⳾J2Z{6 kM^_{ۻ WlVYjke7^uv3[|*">b㨫s˟kc8胲jxlj}k1+7Nsq\'Rdy-LMgj/%:IhV& JUEUJxLaF#% zǽVIAWuՙ,В`@pfd`:4@+6ƏRdeXL6ƣN^ZXմfT$Qzv(@QL\ϯٻR)O>g #Z^7x崶1CT:țrRo9 ">ۦI8dX |ˀN9eĢ+hlK` .O`4$Siܮ L'"Ȕ3k kŏ?S{u'pܣJ@F;;!mZV%jx o ^?D^J- Wpy7ǿ4Ɗ*f!؀KjVLE5:+*/)X5:m1]CRқЩl\c{`kå)Yc{ C~\V1VYoPW3ʑUi_cGI%cT_,f_u5ԚJcj%źG("-T(nF鳎;jG䧖D$oBhlFq-^Sˁ;6d: ; 2ן-iםvĦ\º45cKShe-<{ZZ4TBY+61OZ uB@LQJ~)ۜkm/U,"ll*Q)b؄`h~ȩPCvîO 'lu׊ nQښ*PRv.~(iUwɐLv>qd1B]R 4@T'X"FNv44nbYh6-QI|1 P5aF$tGgJ[J 7I#F*It<{z8nfwzHH\/ZW* ͞Vy.j)1m,q$ځ!?2}:ZuhR^|AL-{%hfG(@ut9' Yۜ$3"7(띓R4ճ)Ttk/]]Op;J[j!fkk/% %?n_MJV:֧cC{rK!,eVbbãHA՗0w3v٦A,~dBA)td}gfHUVRev7E>h(Ʊ3MKL΄3 y'ykq%kޖ Շ1l3W>ʧcӅ8濙-m&kx1 d85-qY%TZ1Md1L A&& 9 *(bD{D{+~# ==j둅N" N3\D/^?hIT]s?:B5{,'U0qTtѫZJ*Go1 G&m/Zg1̰;-{5}a64GY{RXyV)"SR%Heeƚ\ YS2FRCa v5}plGN}ħTioݕK.?no{.#f7Q26^(&pbPSSL3V=]%qx=fsTi+d_C!2:k.w}5L]3F^]{+` >[=44S8wUS<2z {xn+2=-TF:]=nZş}nEh"s^UZmjFD$ ƱRJ`9nF:K[(d8Kff @5٪l̦݇vyS.\-E)/nҕs+6R%.Fu1X# ?*xMՄFx+[NjWǥis*1jCдyER[>m=p[uІ3-m96 u"65{H7jI^u\Rc8n1ZHSAyTe,:WO9%8$A^ !٠Iڱcqn74% BrnP"a?8ZF@ Ƕor4|"qEL.%Lz42/!V^9:`[+DMe|7t8]^W,f!rh, :BmpɃEyu6FXt,i>*=z>?7Gp "̕h?H}' \"uhT+}A=!T6ȨWW&c R01'9tR>5%g#s\sWe:X.mwƂ AW, iQHKL8%-r() '̢jX42j7/`#g@UQh饞#'B:Is&g&I\ l)٬:;-lx@"^v'iXASThB}TrݾrKYKdr^;$Ӝ5 #ZyUȨϚ%TmVV*v;An=*b;Qi3RI"f6ΉMhu Qk.-a74$lzf 0r(3HT>x_bށKB'|/XkZTdkZ0HMSOW1?s4jAw۝ .N mD%@~VV8]VsYLR+^Y)n\I4( RHh P/ (2yJ"ETJIQT3LWGMU=T{ϕZUh%+a{ad..ﻛ]iZ~6(N3#SE )a֨Tl~ @J}.MyJ썤NLCBf4f [F>L9.VȊjõ'xVv ?;ܡCD|0h` q(\>s2!=nGM'J}U }箁.kq 뫠GՏt(x2Ma iiTP/S/M<4Ī҃:zFLRMEܥ/?P,2ypp=G$1{J"ݲ]8[2<;,|b-VfSv˅=8n5CqC<#=+9xՏH |k <>X{Q$84+g>fDž,F'\ a5hP6`K T)|ȠT՟]n0|g~߬ߠrs3_<\ 2 mMleV'ݛ*Je+K2ӌA))RZ=e6'eh42)K`^K%ClVZlo-gN^C&܉{8y-+_` }/ 4M֦jNWqT8j#V'1ջOªffCaG'9EY}bGLGCk-'`\"UJc>o(M0+Y*"L5ՒujoAc8?Ml[&`f72NMY53-[`fyd2x2yq7e#X8U>2o|VȘa1w#fYn*p;fj+jч$Id W"zã΁(Q&ZB'|%5j V=Q*D|t?Hww5}دyJ1 1I+'z#K Z* } ]"{ ԧ[JpS:&ДKq"RD"a3)}]*MZo- ަGB?Qtջ-r/MT4YfVDtSDi-$/OV$_$-ASfr 8s*B-Ty$lGP[ph0qGtZyt?gy bq0 9^SK${izwSn)^{ [ Z.H.WrV]HopV{.񽻿qЏA^WYw-$2Ϥ؎#q8w:W潢]AƣѴ5hJ8 }Kopa቞ƫѰdVܗ+Mee_4XVVa֛Q&.4DHyǛW2zQX<6P]ΑHK4$#2DHdr<q7vlJWu d3Jll%ۏ(H'N#iI6 fX4A5n0'@ 5GUsVz{V@nY}7C}()R[ Srsd1'5S{܅oޡ1x@g/݄Lpw5$ g:;6s1]d42c.tz1\+N ,ȹkoT׽UE|0Oؐ1ZKXAMzD=,呡V8N}S=խ^N .sXh3j\lbiG3eRzW#@y|b(v1jz[%R[G 2MDQ3eme_F#.-_h4RMHxDN>SCbEՂkW*@s:eBuʿ4`G} ղ:݉ aa1]9ݦ7[F 44g7G'0L0e 0H^iA̒6bB(3qjUZ VHu2dd0g$<~jd8n֚Zx'r@ /E߸ (9:m)!% V @sʢ⟭V7|4E &2[EyX,Ŀ!4`}c%݋RL'Sby/#1A6M:j:A[-&T >f{j 0L䞷ήo!"tt@D ³RKgs=sFӌ֣΍jZ=ύ): Gu-57/ P7)R?BfN9SVYsY B8_# Z_,]h &W"jzo摉˕y5;oxD Cߏ䲜9?ۥ9-<ԁy>"֡g}뿐?'L C`i p`'CB);zzx rn:bsTׯ0x*v9Fn< S9#SEqV ?1!y֋Xb]Ŧl\`!bb%qY$ &,q:SuIxOd nWvUm+vVն[1{V`"XpTI"nrːFNQ25bRaFY67Hn/kQĘ;3`UJܻ 99QR?dnrQ6MJ5ZzDܦAJq;Ngj^΢/*V%6Qd2#GG& ]LL6+K8<0qqangqpeyz"LJ?1GTf^WQ'DsJ v,і* QO7&%܁I3F෮:VGO L'|wK% g"r䝎5*Bq1*׳Ӓ[1#rDᎷsB!+1Ib6ɘ{FCwSHU T֊pՕWQ~ 0.Ա 6#ߛSDZOѬ yU(B} b2/L7hV):/ O' WF̞BWU(.t ]e -bј A:Α8<EB! [GTLf0O) oП 2,MuC /fͮ[&n#) [,Nm ÒehZhB6k-36%`6FR+IZ*eTcdR:=ؔug¢y[&˔+N5ۍHNZ.SSB [gmcJ3E! M*0"\G7o_ns̊DHV(Y1FK1|#-$ )b&.1*&ȃH8ZGȣ][g iɺt1=$،(!]\x۵?7 E:pUK\Ȭuw?I: Ƹ/1YZ M\2iylhՉ ߃JA1mJ1y{}Is [sF2ıX\u*Z Sƥi`/UW#VLyPdqt=ݹШl8kqړcq(~mFglL{,byYpI@w B TSDD+w!c@Blui6K :[{ y{m'dDxs5:"woBv:wؔLg8D^'d ED E*J7|y52O8_JR@Fn&RͩXDn4 JÃ(d/YJ療ٍάiCF))$B`ls0wZ }*C$"\kʄHL8'*qP-C B>g!Ԕ)BdvͮL/S2LZz43nufq!q@arezȾME,~½$JMեWZx5]լKot:UУљT]{7_ 1d+{_0D19 1ۆ C$* >٨ (4/?o2dꝅuo%.T'?mZzp8NLxi ULS^tvMQj;r5qz(7` 3y$pݹu}B(,Nt޲Bsgʪ^xSaLe!EnCԉќ! @KG_KB)hVIyxróZZ)vRf_9^(D#%| f) WS ƨ7ɮ~0kJ;dg` Oim "me~yȏ> oǓݑAHO+Kv$i,xM2.͞Iwu󇯘vvL𤤬cyru C-=OZ:y±¸rG 6>'pIIlװ. .JFᗢ/qCFNS&ml8M5\D4ss8rnJdu #AZQ6|rYr9)C"7mn?-22Kbca8fO0g[+<f 3J"pbZ&!XKQ9R4"e/#M &v1"Ж21JdS)Nha& LdTA.R BI;|5{N-sFC6]&LOsm9C~|[ \^,?\0M6n>=#(N|藟ߝlOOi3GBoҿ{&f:Ǣe{twshK \I{t;*&"APK>=Tb9KK]|ۖER-W޶/-+0FRf>;n7[0Buih>K>դD^V*-6: =Xv.e,/K;+-)xhhBW1"ضaSmQ'UHW VTqC|q JִڹvH?1C/V܂5Q̥o $P/ZjLy-N9BQ8ýSPz'^]2:ubw:zwsE& &dr(J/kЊFPBƔFt>/fStˎ0KfR΄/34,rk n/Y1}6qop,ٜr_N pNaj-֖X$&L*`rkEJ ɑĠUbǏFp&Ȳ&8֯6V32͋\ȋ/gls}T M,Y/7֢f`6<+ K)Cc#q쥍Qb >,WĞ4f.`&`]qr&EkEvgFsw?In\CD@ݬ Yh6G.\وc o!#?q,(f0d2>z:9H5$+WsE:ܪ;]hЫf,vǒ|D9i S:A+{aYLjYKr[ (G%$0ˈnŴc Aw #]dMO!WC} nCf^cC8,L%*(AqC) P+tHa"GdQ@E -}fh1p{RT=)-% J5-#Z 㰎f!gwdUВ2! )/ B;2>[ta <Vkix/$pB)&4;/ ep!R=7N\'@פ⑆' O3X;EnB h-i9IDF"D[][,@Gzh2d22HkXB7F->CX)#ȑC@u;3Op&u.9aH04ԷHtZkt#rµ,uSLHncmP%܌~> '^!*u4..>>N/qx|8[rz~ʿ}eibZc*E S§J򘮼OO%T+>uxU*]HrįF%Fy)#/DSFX ϥsNx @fqC[n],kvP5R@R 2j^&Y3Df¬B{hDD TϾs]qt%JeJ%/4CRlI_f.$lO𦹗uJE Bbgs3&qK\Y Ф5v|סb Ivc[?&$݄s2Iwvp}u9ez_/ יҳmU1nwNhy{nw5Plruފn-u\5 q;X7xvW&Xn)2ǜK|RFw']`֫G+T`󌓆\ +S#l[w?a{KK>vŚz./XSe]zч^=F2 PՇg٫Wb+*jyG 6A=4:8J4@sI>LJ~b{g:+lD T"%JTྨdV;)Ec(KPK0)G r~2#%aE%՝#>^T;}ÿAy[ ۋr_g>Zʖ\wwo˜J^Ouq*.eI@8M}(V6s6PseRiG<|ejnc=K~zzD,O=W]iG$$~|.n0#dg*`èdgg^aJEs6PT9"OhReBȀߠ95 b/َ aP~Kj mTj3~8.soҤ/1gB*{*\t?Zc9.K܃ޙ/Sk._MD L2s2^MhA +ODfN3#$2;X\VɱYޓc[9V^cEiXP(:D!Vx(*mT5 Fpxiy,7PZ9Whe=FY%u L^_FڙdֈD׀OEkwܔQ fMm'Q%}H꛺ +R/UW%m(˰t]\m-+x9乿,RL^CiWMޢKXTBR%[)hZ6 EN-u?\U~9=n݄ \6:e 2v!,4᱃F)w8Z%? %.>=njɫ, V<%@rWO?D:ő/:j# inZ0 O2/g6n5+i"0!6_bIb4__u=͔!/nWC,p5%.e⺙):9 Y+ f#3ٝZZ!YgڙN0R]73Z' \ VԌCpVeݖO$ L jKxwT`ViFf$њ~h cCKV'ۍa&83iT4vkyWϴrnG"~{ȹJ1ۃx{u!ok\7ͺ8J^o7(g{N:Xm4\2=8/@L~34 f0cHS7azrOoS-ꬨ/G5^(/[|SLuR[ l26aZx*(Xڈ~Jn >-PpϜ;iyf>%I;?,M]jvElF>\q)Gat#+Ufji h-G&iɅ1D[B@;됛(CyMj<fBQ2g k$2lHH*4zuHDr*v0bZtNU8+"+ ՒKvr:bIzl:MQL2QEQ*P2cAԎ Q^p^D qX@ _ZzT^hUNaYM8 \`1SZK !XƊaؗBlaiu9pmVK,3 `m " EAB}!oRK^bsX) ^aZYF86ÿ4~"}2k,EK|z[y?LuM&̀q---ڮ5-d'[/}ܒcD(lB-2by*JA J:XJD 4HQ:x§ 9kqj""wve)-fײuE[W$Fwɚ/򷃲S\H6wz,dq~'e+LnF% CmTNiD-xWOdWVWϯ-W.=+4@mjqѕC:[w `zؾWi[- FW႟'q !*jvZ*<*$%cIS=^=կ]auFm;kl3,8c}8(&Jtdh(.Ah{??juPp%Oq'åG|wR0t` #G(!*8Fo064up ((8^C_Ii,і!ƒ<8-b F:"L7@;;h/'hY,KN33ˋ6uC:9wS#2KI޼RzKm9ˉϳ> 6NJI}SZKq9|Vj3;Ye;)5RԥuXY[i^c.Y$c.FZ\t^Z< _zH꛺%.=o+[=:J#cZii-5J;֋qi/.}Y=r$zy݌#GPkh6-TIwK ¼{iU*9{YdtwE&Sl45´PeKBǯ[Qퟓ2ȴ~{zt5/ꐕ-c#eeʹ#(;TOU4czMV4p U%C>b/P {P޵($W!@u(ި޵ƑEcփ/pX$8KMʖO+g$f,czFd%v׋*WUJH%ZIq3z2znV"U_~X7*c Uder豪uUo=xjΝ4;%0g "K؉;DꬣA:bMfY7tY>8iٔ R2'~O&& IP*i#&PumGQYdeClisGZ\Z"mT6hɘWK'ф1 4!XuTth 7A3BY9iŀs 3?it &4%B䈈3DV Mާ h:zIXY;ޗ H2әﮗRFlvSQ9&W:͎]wB|l ?B `(3[#De!qz-׎O!6uceq]7VpiwSt~6 u(#'69L":H܍bN4t&~~i A5+cfδ`UHZSOg ޻VRkә ް+3 oq`N`:T{DVL >¢<W+p)39m:h$ 2zwN 6iIȔ4m[)PcVI-,Tֳ^[TxOm mS-V 9u1 F_T)h#m-RF0 =,$lRc4v mdy :Ho\x=?k4p[} )ֹY 5FBVFO5IhB- 㤱p0d=hGU?Ȑb}I*ݢD=N5&=#मkkKWJCCԍ*4#(K*ڠ .}S`u߷O,HE *i'H1Frusg#ߞP/Fr}vs?b Q\̚Aqe-ߩS,)lbŌrҩɚ~g'kiv.BK>7J8gLPܳH +PܳnS.iwJq2 )EGl Dٮ __~-m;B!B+%$`/,gpn:n^.uV̇.b 551|EbVdLQ+Nua!JGtp};K/!2n s oF aT(Ap֠ ÷`UHpnVc床JZP9!É+yN ^( j( nA֟_tRCĵaNPh6CC܂жg8u^ݽ䔻OnkƷO~q>~:\m"m-==۷&\\]~|znۛׯ!Ly!凿zL]],' 4L]~D) +Me*ՄLgj7F^Ks]!ȣ]I430:#mq ݰJgX;o[sVډ[şKҗ:3.+oEk bfcq{b:.CL[*|g?f1> Rl@lUh4Qr!EZxXL"ZT`[篭ވĪ-b\=`iwX~ǷZ2X߼X~eZ7hZBDLqoB,i7H_L$eizc"E& ˆݭ Z&;9.lH[Qr-fhliNHb%PC2HVAQ*k7`e0B$,c6~,$0ZkHB,my-`cZ ~+0V2|}U鍶jIXSd: 4 U!j!*cF} X򆝘@.~]g:`BQO!CwIFSy{C|^kK|&QfM/?.:mv{֯*2*Ӻ&A$]o.1O=kvnj݁j/WW9DXjgHhwUclᶯ+K/g.v/$hX1CooJ`|sژCRx9;'4̯>WKlEmZ=Ҋ nd ZLs'jtPk7R;raM~.eE4uʗƘJE+!vIG% &a7,- Qʾ32 :4y3|/ƴeβBDj ÐǠ:G` C7&F0&@#1 QX#I=r8;($LX4s©!n5Vr_^[6P׺!\!@Uʽgw]=!έ_Wv?GÇ^H{5MO7-\Է6P%:JG<{لj3U7Ԧ4ȉ Ϋj*@4Ĭz4&SE.A1cZ#Ib2 @1(Ll^ߔxM]+62` $C2>B!:kR@c6odaN~ncc.Zu*IpMq$=AJrMYIL;^E< .ྋ"}Xo P 9 s:ڃ|_}NsGbev}|-j)&#~'b% ooASB]]TB zjJ{⒮ӳkr7tӨS~ ޛqշom:S _7ߧ ph̺ORhaJ0Ff6B?@h0>r0]gHT?}׏F{0&0!<"z Ftj#+nVV otjϒ7Hh);0^HtLjͮ`T 0;|кi6(E墄hhʨFHٲ_r Mh>_U]|w۴ɸ'<^McG)rG?6; P/ptWSmrS82ry>}YV赪`kb`E!xL*VޝYV眲;olC6w;hg6Y||e(KΨ,ҵ`ѣlD.A9l;g4` zv^4O>{`15R$CW (YсF5<[qr*PS;D4-{Jlj(4GeSGkHS }cpFꚥjf`ISVG&3emި9NdbYGQ{m(IU]ì4uuMV4Hy[&/BɱX4,hLB6^IaZGޑACk )$hf6&B eq;Z.@־LNAэK &'2=gJYVю">9)S&]eh!V2% ]:Q=9Qfj nO!i[ FU8QܣG A$tD"I A6ȴo1R:Zzb ؊ĩ0Lek+O;łCX8&3N)F#5. jY,~Nra˰)`ɷp p_.o~kgn?nCn;zShڋ[B/ S斤nf 1rF`4G ǫs_O$e _PkE8qL?1X%1pYգ=S,X8\q._%3*3[InNQ$W?#HD>tP^}xb]Z%+WWoUstj/mml||7)._==w1Qxڌ7sznv-n0)Nr8Ѱ6ZCXDM`t5lP}u.PpsG\zjKSpUXL}8hIs[njq/~M~ysn__/o'jev#nUqC5MX~ɵ"7y=niz3ٖòQD3ѻAFv#S#lm<֣лMa!߹)'ȻI'ʠt~w;m֩/v-nSXwnmJ'?6u:X ;y׷Y*™(`BdG( 5%w@Ab^%h*;kV7i$ dT͎|H[Q }߭OͶEmb44eAصkc/OwGߨsXYxɨlRLg2GjBղ+8RdA Fy~2Q)Bdۢ$ϠS\tb'u Sec =6~d^u(%_x2&RbP|&aG<+RԞouќ+j@Z-Z׫%}H5׌۳ uFzߧ7{Fn_aK*N՞J]v+/_ra0WgYԒcoj58|I`vCn41soDq OI#dwCM\mkƐP6#K.aͩ n"lUyBAy"u/94f8%T(Eu~m O(WV2p[uH-,3&-ZC^xeDxqX8AsVrBs ɶ㬔TsZ)%bR0lS#`Jk<'BSLeBc .J)l>P~#jġ,eG8JbFH|e F-j s%x%*8u`@쮆af w6  DJD˜0,ܢS&Y`TIA$NJs6ף` 1 kIa@ر^P|~a hkQqq%/mr0rq'慹 ;7YW׵DhYy0P.26!Acj|47UHbjzO5lx a ikik Q]0[BW8i= y>lI1HZQ-8b=ӨVD=ϴFwi!8r!M`~?6V۶k}['o7WV߄CFo6%H?.fSKC??s\ Vs  n0j+O+kAw w@NS'eBj9Gۆ ћPԔ#w}hQX/ E gYtQg3CS):HH*q{"sEJuP̥euD⑔vΓUnOa7/.9] \sӛټfwfv}3{x,bhU \br[s|vEW#qWůZIO pu`6{=_/ਇeaʢ`Y ,g/0%LU N-,h Z&Ti!c!:/w*#:}~KASJi#hzk*Mm*nȫ0?VX`W2=K0GjHd'@B)_z)ZMف`Zxg#!%_4O.*bDh6J+'}fl+7\t%~&(TО;7Sϗ3e/ ;`2ة=tzc,*5jԈs>'r0Rt"aoEKMuzx_wjQѶ]XKN`@Q/̤D$h{ApmƜ)~7KfG:+l> עϷ;EP ZOL # =vO!%aB#PqتKrg36ޚuPc) Q9H{eWdsʀZ! & I[کʐ7PVhq*Eډ㋶MEL6 u Ͼo2k7\DvAȝg\ wRM@N^RQh+hL/o& 7ws>sY?YssI<{7GPJ0ieeo7 RohF?{, RDߥe[j{6Ynrd| _n6O.xZua-er]0ιsru֪čY=$'q}cH$fW"(y' #T\(hpJё%B.j'A~HďIUBM0Ty>c_ 94JDS&| ݥpuZs3ߧP z){mܕ!da{CE&Z2T"dpfohAsZ8H⚪J; 8-;Cڟjݻ^ɺQ/Y7e P xH]!&JgX) Rք*"QWY(vKo( vn $Ą2VӪ3$*(pՁ٘hB$!\Sr_kY*%J_LkY(]h;t}ljb^3P3 fJhyQ7158LNaJ NR`N~7o6>n?d{2JGWH$6{D;zc$gnu M@٫c]ݓ#WT$j[O.Uq9˹ݱg?SLZE#[U E(6ud'_i0bH ?ICWI>(/@+ Fm6*m'F#Nd"g'$|җkO Nvr^aWX֣7wIS_aY1+,")(Q?ڌǤ{ S'@,| '`'&ZR/Tͷ/H-ww_\ݾCW7dq#wvXdx@/,'!Z=ZuD]jЋyG(zK"enr^ F= !6@Lh7W< &%VQtL)8u0AYHOL} }gpq='JȳU ¡MRʝ{Ѹ2hξdk l1ǸsδzAavy35eIѽ)6ul)֣,!=CSdgc'$skI<' >x͏+~9EZt* X\ >A~t|s`LrvdBb. !8>1 &e;ȏwBGShΟlaV[v xPR")^04#e0Lic)sФy(E5K K҆j/ZzZkq-5x Sfj㌺hYk4-MZsj4-mh9jiLW[%OCK@Z[KҴCihsT٥U+s{sRcҴ41>h< -5&MK ˉZ *O3.Ӥ֓Dtj. _.8ƽϼeޕ'^ϱw켱R/K='}2{azeg{J|%g.V 2L)ϔ`L{+TN[:>] ek!ݙXQ[Hw#(86dž2zXYR㸬, oԘa)јu "lj_dOF'Gnl|M?xM!ׄ~N$XgJ+0[n~U:O57^-~/H:yPfmq6?O&z#_ SNJG<X|2 ۲(&9T\J笓+uAL5|USX9ܱY+,ʒ|RPhLW}` Ur ):G)Nv>bHwfO$y\Uڮ]L"M&s5׊їαp/+uigӕgbrizd%4nJV4DixҴI7K"zќn{7OvxIy ?]K5VUEc)ݘhd,ZNK_ v˴o~yrс`#n!]"u (c汲-N/ɕqK`HP0þl_CbEΏZQ+@5 .cѼ?OWic>!5@]6O<+W։X0XiU)tX:K"U^:$^FYA7^l~1s}*E 4ZYVZ]dԒNΔȕV{G/ !H0dB%dJ/R) &k&"zT%(誜+u֗*!ˢ4@H;edjdW rJR l <9yRa .͵QE5}6Ooo.,?=ֱ_Ykޛk c>e> W1sL۪߯?2>Tl;PE[mD}l[cf[]=?{쑐7ϛg*S<34+qG^ؾjNOg>Xk*5} {N5wՏezo[ ͸}9x<-6R|t3fh3 :<0%.!ʎ:/P ]m}6:͓yn,LL\y:Ϥuk2F$Q=^9fRى9F&*&wc"Q+nHg,*&cu )5';@NLhU;tuhs#@y(Zai Qq,Ḣ~.etXAWPr=єB1X [T*kl@@ l*$9y JBc*onl+s} Mv[ࠋȃk]dK6F=W+ELKhGz:^5ٕbN&r.J h\ yQBKhG1TJI|kG`>B^5  e\z+\ΤH`yty&g9grIN ִGA~$4Tpm=Ni#yֶjX%|24i࠵(Acg>o%UpУ~47uc ڵߘ~~{8(i3n wAYpl; 63L4j'fpiҗB&]==,gr8?aEi..d$Qc};bvщ9 x_É3z҅^1v1tuO8ĥ' u>-G䆠-dž]@q~ HzYE {ae| Bh;CYB]2KlUaT(U YPMä%ё/[}iV,GV?#Zݭ6@*_VxRwET̫_őWi$^Pj~~?IWϯF͔_97f^ob{DsQpcW͡ŧL&jkw% WPM;﫻պC$s5hڲ0zBs |QDdO?S,㣡 ߍɾEUx ֵG,ͣ46xuqFЀ%W6"R+h&^r!)tҤ^rQ=\єOW\ nuLm Fo ^ԗajH^=8o[ _Mlxw+t! P~ +4"`r]׬Ԏ,t,GI8 :r|؟~oZq r!sn@yTUn"+2+rgl P DL,1vA냹kZx/E.'ִlEִtd-pjӯeFlx;mxf>Ec+ZH,!] ׆Hӌ{Qmi1ҁծbrtKX1+S/b7l!R>?D5-xG5-0hOщb뷶g?l!;#xGş;ɗk[c7acWj(z㨗eQݭ([Rok|[Z?߶s3Cy{iJŪD{biB.όdDͽD˚Fvm[#W$jan^- ]tX (V7u ]vx5*V5jOc?7^-~/jo5T1Y}xkvȄD2!?6,am_դ֔޾Ăz\ Ue,\(T2)wU+ >X(,h08ci#gfɮVY(اfqҬ C~Jɣ CZiVc jK|3ΧfEE|6rlJaIE[BGX n=2$7wPC݁HWXoӒmCF7c&&ZvI`xKs4k 0-g9n]d4?K֣7YiqS484kM(Y$5K&,]J8'΀R} 'H'&jRG-AC>Rчh_=J'w8UiA]".SȥYiϩ>W2,z9(F_|oVojm[\4A/ZY4kEǘYD[ЋH9r#I/W%<ޫ;:d|^5M@y6,B)%$FʮY4Of?yq͝:E$t,$e?t%3UA2%qX yTcsbD)̘-l͑:l_,RDĻ5''^jJy:dqI$bRw((K1?{WɍJ_T3BXz{ö 3`U,/YUN ]Uc%(Ⱦ0iL[CH 9.!:ps~0+:ry} rҡy_$ %恠m}P):23H48 I@?@Ճe (z$@BKrf${04 S ZqEƷAD%w DH-K64_˹_vM]&: [؇/wL4wT˝J%_>H%dcDt¾JQUqR$M~!zϝui|ĻI1FeۑW=:6GIο/V ډAf`g{!/ CY*OBz7oT)au< 0vM` *:?I9 W`2Ot#x-|h~QN'x(<}w㈟Nі!-aD6k{zӣk=ahZc GuZ`mƯyQaЇyn4f[ ^n/~4Ǹ ꟎jCAX (g mB !orII ߗ, Y-^e&*v(΋:V y =A( |[̴VM 1:`pgMd޲ʨ!2>6shqQ0n6Q\Yj[g?v$N 7wV7e ݷnuO鷳YnUzU cS1>\즩_>9íOTM\QgSNyz!S!Ƞ+mpgm=O5?5-U;z7nE6=Ykһbd:n^AJܻz,䍛hMisw YxR RL9m]Phޭ4'ѻa!oDwl*UFBƣ55uJ"0>LHpp euc㶞lɴ(z%x;=谀v N~f\$[i-F+Q%G BA{$dK"3h [q(u-Q XGO%Kh"#ޘ 9-9?rm(e.fX; HVU! PWѝ\:BJNzx$FLmq&p7 LJ7D+Vb|X^h%4eS*OJd/h*ys0wQl`]v\$ k]4|CV>|C H n_-cy3GŪ.1G"cTt.VZIHW,0jɛrqMwW\1EJ:Fj[[ ӱK1Hu-FN::ݳ vt!1 y&ZbSRXnN7rۈ:ݳİ>nkPͽf.֡ZKF7'hAni7"2t}ܒsEE2K}FkP]œO¢ѻԍRKS²!oZv4UyKY:!EgI1B clMmEba` f"2H4֑bljĆrNkdYT8ȥR:dp<;=TQ0IYR4ViX h<)k,Dę `B{kPZZ)U%o`k2jD nX$z0u,5sDDc#q_`t཈)eͽcbkw!z*Vboai ƼVǃEo>}6sWpW'.gw ?w7ߖ BωLRqMؿ̥j|J0Aܛ"DR5;Ht28W&zXa% K`jRuo eh(  y&ZbSXsVa߭ғ6]'['OkW̨sp2oJ'5h]{MȦ>p{߻V”̎$~#q-r+ٱ y&ڶ)LlR[$+D&7wf^-w99rGĹ%}qZڏ'g[/)lX6u:JNf+J)ľZD~j"U3R"Į hM C V6jM%48.^/X_cM"u|bZ|W|ziM,9O_[~[$x[N}N? ]TwEB# +*g\Xv1Fю~oz *zNp.kd5*`~ϡJMA]n>,iٻ(.z`wtҍh2AĤ|IX3rlV pԌ_HE_H̆=Ht$m}Ԃ)Lsi{0 HzOT;P~ULLg|}fIQ X{p;'>#ɸ}#Q<,BALOӃo7AancQRW4CH*.7 zN #+%Wbӛ±4׸)Kz8Xs-F/ŚYZjɒS(N { +N,N|0Y;s2zvy-BNxRly1OVٙa Sǹ {<0I>I9Nq|nO˛y|j0M Zk뮚UG]4\$wuc̄P> ^%>bA/D1f>3dR-aB깧jӉ)o8x mwv4vL)zg/`LJUDzG&USk_ 4b%Űw$1l/*(ӶNRۗV>B:s&(e Q\#ݟ跛a`cg-7l0[ry{~[Ziݻ1T D|:~Zw}wg?GW/?x۰3-(c󷇣≘ߧLvߖ+#w 4V-Y+*)Jg߽QRf躞w/ H`%(F^A-F,?)>!;ALCH Kme]Kȅp[{Q9i XJ$֒Imx^)m#k_+tCFtC,;]{Ryߕ]y %IOOmaVῧ]hqG D %R/]*9.XjE|!>s !>s;htFI-!lj1xbīN` q4߯` sw5c>]W.g8 %@vUWC8ߗ;ݧ; qȗ3ʖWjqp=ݜ.53؉cE\ޏn.^in~cNݻoq&2!f9ͺ#gSLʤf\eA8J@)FRpsA{^=GASgĝԺԻ:quҠ@w-.Pot Z@Ƅ;_xDžJa,\KB?ণ͔SRzht jHGz 0}v M@"kMz]::%h 1V"E)*Xe-Iާ4:=` ޻6q~W|nTXx.SԿb~C"h aT^{_WfM .FÅֳ*V*X5];W[l1X,K!HHG@B> &QjY\%EK\jE<|(ADZ@+-)JC+مV2V': /eB^EKJIU iG v;o`)4PO+DyNT,rO)\n[!(uɵf1cW _|C'Z(rDcwQIǠz ɦKн`jw$@9\ (!ephW$\4彛˝>}j50ɧ,շON`2]Ee42x,ʹH08YdȑEqfs6`=2C+|v} pjeNaKWdhbbeALӳa0Ƙ)֕QrONϟ"QiFbsHԜLELH}ܬd ^@ XG2LPQ:4*8W3 k\0Ρ%ph)x)= f;8ѝH~>qԧ6Pm[xj|!koa+u{ȑTqs( q$QЇB Pb %PI[X(-y>EHI_LWRY[nvT ls'rFuA0RaTәD$Z^aR#0(&<:Pir)qc 1s`s M4:N؂1J@'p ;f,]:(6eM ރrD&|! >cu~msd\tФlV`-yk4eMF|fc<_ hHyJɎϭf{2&S d-ºukB%2K<{PĠ%/Աc'TOU t;hQ7t}Ń٢Gݘ/IAxI%5/{`h$]G0|yki.vVK%lq0Z`?o~t$ʈ}QYd֩ ّodaH\$`$n*itWQP$N]iӏ6 %sv$#vX%M4Ve\z.bFjFxCz/y.]ϻ&P~̌PJ{k)8ec C:Wҕ -$<\`[.t ^%燚 7cU\On1Z *_Wq~z޾|}U-067WUu MdѦ0~4Ef$qbq?5` /<[O]$%HJK.]r /r G^aI'ac?5GQ`e~jbt>RHTk$7jN9tQZHͤgK/`ަ٬M(QG;y2?ۧw4ѸK'= ߂1J2EPZ=>e ~gt [N> 9su\p+2~/W.|lO$l= ۨK  zZO~Z} )/oެ>3_M_>W// ft`pc5>V|w?6j# fS }yY}GGV!wGS}m3wFw/X={WH/ HL^n Avq{{I0 Yz<!%n}R]UXd}eEjYJ(>uy ?*LeQV3BEK=.p'#y`1֝(#L{/X5߹N&Y%{?~J{z=ʐ*wTT%:jk˄}.J^Gvp1*ގָ̚_SEkX01h;NYYv&/=@#ŁBJ7X_0QYv 8U5Q FF7t>t{ӳF)<9bb./e=ځ\2ƒ%4D6A@,B ֒!Ѷ䐵ѠkIupCUCs:A1eplMZ݋9Cmhkp+L.:9U -E!5!!m`!:k8":'zI>mg3m"=59P j 6('Y+B9% xkMЎI\B[ jZNX Q6 !uZ)BK:i FP6ٺiE#,DQ.꼴߯;A A~=<Bբ֘%/d M\bഓ('MdCh4:(PqO4#Y[ %Kmҩ ?ܤvge*a/-DL\&|[> ޝbԘd_S+\[‹Z\IC2|Q@"qr2ac. -Y)f-O|q]u/=n"}\ɿC.7Om+V&7:rlOqus^JLFFY,Qĭ|J"0TOP+sݡ ʁ|@|@6"[@:d`7Y?/^-\Vw!ٻ")a=nZϯO g<[L-n^x%`Mӟv02%Mv_I#d`|B)BHS,ʧ'Ve*}O~.q/PvVYvqVɹb;pԀhgjP_{y\$`fxɷnio(cyOjm+ؽ $v@BV}n0=B±CX:B p%Пj-X Fcppr218p{Ȣ|RTj0HvqUAE)U)dW?qQU&:]b3 U?1Ȟ /|A`;ɃOvQn-Id-%2Mz9a_/?Kyw-V.݊#IW.dNdjXQ ŠDtv[ht{h6<6P!!_)T'oqnk7jrD6.혧Z !݆g`j1$+2sȅn+X2BFYBDJ=OOvٴ'Ǎ >׏(ΠSb $#pQ }?5KGui鑝L Pʽ|hT/AHk]}ɇ1߶ ;VSzMKŮc#ker:xhc-f>nlRngsg쨞8-B%p4Q" ݿ) AKTɲ&6M4 DoTԨ7A4Q 'UCAe""Di5r\#ӲnlUS[cҁC$R{rMsaO=^T֚aM=u"^Q+51/J?Vn]s+<$BΖͼ!L!`J)`I+R+!uKZvShG| `~5Æ t6E:6 Cabllz Q})vՏ3~4JB̸ß#JƪI+0I ʒ`tFW{bS2J^ק`$rg6Vn=oTBWnX+%U"d i$utNy1x鶪Ըɩ}Al:(GpwnE r6݀+oS]rMu'nAN2+ ӓsq7#euᨕƚc4ºeuizZ(G 4Fӭ42?kt6cF}m B-|VΡʪThu㱶(b&#%ҞY+.HM5כnVf' gmE ZQjjhLLr4(1(F͞!.8šBL.tӖS=|څ^P=[5MPC \Ԙ=XW鯐P0@ʵ/chgr^%MGN-K-ag-l)bmZkϭ [9uPlv˝K9zb2@+5, eAfqcF"i]Uf6eѢv \> Fd)lZ,3S==SLb[ZŖgVtgJ˙?!SƜ{6VqOJIlH>f׈y(]^ҍh"LT۽mIDmxeV~k?)HEkj*%YdWB ԺPWx#Xuo0 RV#84)!}M/(f+!ְݓ0f !-:g d$i;o#A!%iȚ( ѹɩkrfS8ʴiC1b%: kiz4uXQFL BF Qe{-dww`Xס#O`[H/VD+:`gB$snQ"]glG3%UA%ҭXB[YhE [Nçc{b\w^w-:C@įSƨDP(yo- .yw2:{f{H=LhKgp6 0:Q9,*@Tl%/](w9Eّjb5Elٛ:>5}uX9Wөfu),Eʰ9~ԯut*Sν _Z毒ҡۦwBn$$_At=iʧbYГՐ8G+8r"t8{*׳wYpiN1r3L!Gs#E)-Qu(+Z/Y&!;Q'rkr̰Lz똹Gk87mdF9t#[{V슭<>uO8#wy#rX)+g@k8%}.Z$=>Xi0>@wտ8&́ƃ`r%8Ǖ_+3ɢy͈E=t՜/Ɂ}(q饿4_M.&6͖.-'rI>-NjZsTqQW`~,c( (1D#wm tʻ_3|,+]~Muhv|+xeI=LE$4BOY[j$# +j#)}*b$MVx‚sR򳟨%ə%wb? vW-AM5n{NT\ GA4E 8ف2㊣by/b˙pЉIl9ͨn'|D|VN)@~ڧq`CDzD i*mdJf[ 6Agy&08IZdrQ̊@T\IIbexAT;Ex  $7Jt06.t #"FYYn3*- AIFHd#Ld,~%lFyE=f6XB{FD8ѷ(Pچ,F9-r%V/r;$襞Z6jX\ֵD_.(3Y؍yc<*SeY=C+Eru&goPy6@4C.޹@q9CZ JT RNiDK)hw3e(V<+2arh)>(Nv"V[ n](6<4޼.nsr(D U)#6\#MFTr158RvUf0tkbDӋzQ>1w;?r}&6eϦcV!?5բ _|WҠJukzBaQ ~$ >m??/"|FϾn-fe.!c>vg5?Ww l`$c}ldZ0e!beo:/{f}$gbߐ>/שFz9s9E&z`IX4x`ُ5<0$:  9@ [ ֢ւ~Z7`nj(>/שsm٩;kLL_OPUl٠ 3FuuRv{ZӞz6ĠD+B֦!+V5'f$-FwsaW@-t]uR^.GCc_/ΐөL;dF~JJ03 VVֆ͔="-6od \q7TWwy~b1W$-[Izx\7Ki/#-LAa|tӽKq;M!Q wםOմ՛S9`DkRK6^ vzte\P Z{k*t:CQf h2PZ$BK AiFob*:zJ;% ;f`'O-*=_'sɗK5׈ \~rj/?y \Tt{e?c:(/|͚EIk VU7EIkճs9i[hQxD⛱(kX!d%DܛoH W-gr% K@^\r8V-@kMoY]oJ9HPLDDb̈́Ǻ~EYXn%`ut6qWVkR"EvWJ|rrM6G8J$o8FJ-z7$6#ffFwWOW+%-*%IUJ4WJӽKq;TJԒ({⼏:wu ]ᘠM`};d& c; vC6;*Й=2-%PU@Y8#q9W"@fѶdnشP-* =t۶ͩ!UÓ`"gF ű.s.k Y6.Oֱ9w؂j$nÑ-*RWO~V5csJ,,QE^}_f8}|Z^ƕ"{nc!ȾßKVg~vwuu[ҠH->^[?ףKm.Wi؀/_W(xƂRk2L(ϑ)<(#ڏ"EGUNYyuuk˃iuۀѝѬѺu!o\E3 릍8 ֖);X@)ެ[녦Z.4䍫hc(dFe;GeI5J#V&W$0T(mS}7lEXѢd(( rj4֜+bO"0MK~8 \Z uFbbp=t4-T#'"ġw˽}/fWa}=>܍l:O"*~0 HΩU ?~ne"3:3v;ėY&|] L[Fnug5T1WhE&$>o/?˦+X ?8'.O(Jl ~٭YbRtr Fk1% Fr(j q#2<B{2aZ4[UG&1ښWGT[)ϕ *78׹d9:>d rG Vo>ň)fmbM^@ ӣqns>V2f/儘m~).^\jVԗMrt 5fr\5d_tɷem,<ztXpVPx}UV #nyܠ\P>wm#NU3Ԃhu1%{p?}%{4V꧈f+M@6m}\iZUnfL'2QOli{6 iLR̄Y&+Gn3@׏hkҜVO/`iqMz8ʜNԓEݏu)GTȄ~ :P֫k8_;U˖{UVUVuWߌ`xY٘NrjSыs)NRJ)b| Y9ƾE(%{")o؊D⩣x#WQoHy s;.\#G䛈Ynx6%`nXQ}T_S;aZXrj=kIk)&ޗ+X8LҒj ϶T&HFcC5Q1g-=5-M+RBKQ}N5VkZzZ6$>GV48 -&MKK>kIk)4-(dΩO[Kb*H!!MKK5tϿT62)QLI\=d`>5~%^<9kǫdW.]We"Ұ)&  )}'oO UduU1ʎ)ఫUbfgq\XQu=LL&RACOTLvi}0itFCh3A0 [kt|~pUT+P( V8GƨxVB;#g b( 7`H9ܡv0Q 0 &y 0\N"Uh#㠝o=` DK6[8ݶRؐ_Аx-W ԋѤjzr{l%JvR(( {, ֍i R;ՅGgGK~+ N H9ZGWO9:sV^VtByZ2-(UӥkCc7Af(n87^z֙Ѫvj2ӲX=OK&Ut m#V76%RIiDR[By î̴RϯRV\# juؔeaK˷ Qˆݯ3_f>b))Xƅ} ;K; WsW84ӌ쑺_&wѢ`Hݯw3 N]k7 &k2,3On- uFvuf8Oy@"-`Y@ÄFmSuƅwL2giZ2!g7daMl@T 91HN>9GĦV ހ]A56/ }l<2 wN'C  uxi [D X+FWFeH7:OQ e`b# u !H%)yDPm2c鹩JX>Pav{._M>eZ?ӪEg$»M ֢3_TP=VC_fiogpWC5ru2KxQ:3"О`' I%m˭0B[7](v8r>5pXJ²# ө˧6#˶OL ybxr<%y5 /-?XD]y0֯īߑBپr92(qww4BKT 5[X3Jt29AÑ*\5eNDMNNW4%i#J}Tl'J[͙2݀cQ)ᑷs)] 9eHl5ݾhh5VA_%pƆKNpA$tBX4[Y .Sܣ#nh2Q4)  {((ebLA6[ 67U: m&6 RdzhՌh(cymp-L1A(۰hMC8\kCUX!HBG2z6h+0c4Aՠhv.(S;{)U8?T_tKbg5+9,wo9??/83~݅4Gs`I 9YR ?\__|wj"itϗѐfU$~z}vjՖqmЂ}Yb燛zWbl# S12f863>»Ԝ5O2B۩/ϗZw>P>_?|!E,HТ )G3Оy9]w)K%\L~L]?غ~vu^:OHd9?\)z y~G (Gjڬ'Xyղ+]Xϟ~}5XEBλ쪺ӣKq;TՕIdi5󵻱sKqVΔN>7]0QI%d) ezg@ \f7?f>p0SFka=tjfl<Ч"*B|O/q⾴_/k@V) 3_4|oޓ{Htea3$8#$ޚ̑ZKju& cWfT.2{WLʆk#E$Z]$ U"4>Q$Y dVsWP8caZ>G_'GT0A[<\!Z2vnj1&phÀzFT~Z`D@nj K^ \Ɏo읜D>Mkr^e2%# zܻlzFVr?jrَLTj GHj2aNtV(K=A&4֜xƬk]hAFqj>JJQGEN\2J9΂Z3P{ulUQEiu 6( Ivnc@({|Ka1~>Pi`#nb^z)컍)XiKfeأƦ!=>cﶗ +6[f$ Zu#?9~*K^>S )*֭G?ī.dEOƢ{s=sN&nR f.^Iedù 2)u-(0 zQKIQ\8 ȷ ~)dU 殊9ob;%ߢ;$ar{77@>]LS;ʼ-j۰ O |bMKւ7.46U+yOyyJLx{{*Y)t鹩1=d=\;>~b^7M/qX2MjȄ[3 !ue1֝!Z)G``Oحy Ìdы,e8֒q xoVjRLk W#{g@H,G Y"7".e+% ONDh 4ڑRWޗOlSMtƱHt y᷼~} gju`ő鐙|-H6LGZ G)t2 kYc"(upQȢщF}H~жL,)nO>F WF&:##N;xLBk9sĐR 0B&0Uc>8ё&xle[$1 tq= I ]^o1:z å!,xBto)[cZEo55sqLzN;Mq5Mq dbhF*1ŖC`CKC3\A1LөL>%ADnItu*;<+`%ҋ!nϢs9g7~%Hya#ܨ/XZ,zS+T >#͢ObԘn$[>5-yEIK՘h%]NEBc|AǠ"00Ƥl3Y\ :*CͻzV 6+sg~5I&ڑrY W5mX80e=ڣ^WW20ϯL21O*ԟ% um'3-yS8_O9_ȓ'uvwkwQ*9:p51 pzkP-Q 0 :?SL-__d c 0Dŵʱ2X3[9Ln!jA8o*򮤇ޜ wV@ ='WC1Dx g-Ƞ@gZÉlA0)uy|jt`a.b`#0|J*:W6qN> #- U(o$hGvrLzB33$ajFq}SnyxHm_Uh&kV>7sSC`4f0wDMj#YU ۞.ױw>Xv)l:lm-A4PLbS,Vr, /&8P=V}{6Tz7Ḟ $>k@n5֤;X ȣhM@1z8 x#rl +Z4fSc >MeEnlBY2 /ZRZ4njjg%Ł[eN6e kvL~ e!3@!/Jߺ=wUrU_=r 'od28v&ON~T@^S0)cyB=[\=iZ, &y/S2z{e#di \pIL$Ry.UD QDγ!HOb;U]Evse+` jgA"X,g %$R!6MsS:B&IcΝIM!#Hja43 $j=ȵ6 3i ReV@g HRIui,MP7쪦"Etc<>1gAz=\-wW9%,;_iP{$V˧Py6k{,ԪjˡjxiS#L8d~R+:T>Vk^&}׉a>~Efh:L˄$-xC]lڄVW=jQj&HDۺX4pbi h ',".1ul ٟ<8 wee@9o)2ߤ^?^U1="Zwq(u|遗H3$%!u~qwһ~!IrYC(L3e T@@E/3w똊$ o# |`D7pD3/] E1\|hB1 F |A 8DsjuZRLM dڦ8MQ 1Oȶv2G1k# Qe Ds/ цc g;N[Pp40Y*Rk~9,^ ym#(l{gz7 XJ|&:@[~_ysm%G#U#7?ɪYwQ2al$əBF<QĪ$%e?e?B;[&°DWVIC@2/CD:]#pe$Osysan]vk%qu`cUn\i,!5F%L(MJhI.<ӚaNReS("M1x͌E%y"!i$GK`+Xn PNSu@6+LqCfϛG^r+k鬧;X ӧe U^S?i3%*Qʭ(L4ḁDjPG$yjJmb Q;DFTt~tDUv: OGkò%Js"_!ds<WJ Km'#%dA}~!ź1=V`\"E$/T8pg0JL 8r{A}SVlE,-ar&Z0ㄼmO!+9d8ɊcڋU[޴75T_f7 @CÀ`@sif맆ثUFfW'#L>Ф\OD3}9<_qTUXUcUeP*AjM31*iG5ǣ;' V| ѱC8>2sZme0bY=]5WWaƏyuQzLк֗n`pwsוV LP~X> dDP'*/ Vn⍭ ¤_bv34k4__<o +4C'?tX]c-s^r]e~q.ǼrxG3}ŧ[E3yW2Z\n]O^͓Goy3>`mYx>^<˗=pvEGU уw159G~"WIy Yweű@c҂8hn_ Gel0BS@X$yWRcFVjੱ$JL ֙$+éܠZ<6؁ u$x 8́F(5lB]wiQQ}٤d#?lQN͢P Hh&A[IT"|)5P}=N!* 韛qƘlH[C#XVxy48lwauˈwa9isLݶ= wkDplVp8\K~ö%%xOz#,Cõ7hp pݜ`8q!>J1mpLp+V)8,tGp@r92DuX\q ]sй>,|11:u($% ܻ8ғi_-ߵKBrcb`@ Vn޹aS{jqq&I$c,uA;b Km ^-,u{!K}h3WN:ն\ր{uS FWNwDY.Y6u܆+Ժ!\E;T=M(itAt}Gu3PvqhֽmuCCu C ;ka+:jP}w_?vqΓ?$_nb>ua>NR1>s~ j5[97:+/_TYUռ(׫Ͼۗ11߁뫋/kgj@\ۓ/'Ќi2NR]LK,@e92qkp7D$Ö́deWe!x%s4ޢi1PЍ D/`~-yIJJֳ!ZRmcSR L4Lif_k# VJ+=2ud Vjҍ08nwj!ǃk8>M43#( -Ly` 9s駈,&S) .4Egj_a]@Eu/|eǕ/Im fWݣ({jH03Wbj4n4TTAKƺof .?CPk}lN6"axmJ19 #k3Z}I 4Z>M&H[oUXsO 7 3.|MvХ }GȄI8']S@N`V+ӱ D(IRQ_jpX+%,RuyT5UJʱQyUlKR# iI}-5+qܢD3 /44g/==/,n.L!r$l-JR٬JIx8/mf6Xz^zë&j_PvL/kyZ9Rl5McG+ mOKYy)U&,㼴噸ऽTEVOlF~ܐWUZZՠr&"v$*~ԳRkѺw1? :~7K>MFM> !o Vco֛k<MrJIUn^)d SdQo:f .CZGj9ۨ=p6l•0'X'X^]=\ j_O~hj~r~i|~|f_e*߯%SfA1^[ǃ^71S'CF:/(|4* 5["\/ܖ E^HTBaeڪ.&"h[}:y㿑Zi.mrm*s]J-np%-,Ő:U@tM$lounF*9cyf£H/LZѾwሐ=EP|Q!t6.1Mcu|!plJA:,EH!4dIJRFx$%B:' P[q%XBR̗~'uzXB{VL0Kw4lP<'Ws[;ue|Pvv)^ ^tK}-5K}="xZy N:л%x\@)-ykek祕 T(*{S~[RGP Q$sR_mKDgӞ[aM 4NK)ro `x6 (+P5XKDZAKxi,$ ͚DWX $3~[-66 $[Kkdih YANDSlvCjjq?@12d% 40Yv}t4^<"C(Ɋ h-~_p#zksgbȭ6z/al | w)H@HvM'۰07<"+1-/6֎w| f1բsR9>YSPץsК$p.YZpIȑzJiFDV\ 沨rI,H*Jt֘kȫ 䥭 #Jk69}](o<䳫0V!W\@UY3ZPd yjlw+p͏|sءۋpe\$Æ?\rmon|w gВo~~;}?5h{1߮QϾZcA'ә0I\}oovkX}6 ,j}15_~zwsa؎6+< 2j2n w{*Rf$ ƤSILP;I k/5 a;BgUn_ lݕ a[=Ņ5-XJA lI ] au@#k V 4%9=>k#@& jG_P; KaL ",\%DC\]- ;!l& !뎄g@Eœ+2f @QBp5S 6 j%58QgbpO7y,\!yx_~vdږ'V{|MԺ(I1֭'GVз%Gy!1[*+RM|~J SsS4R%k9t2=x(fiU|fa}˾x *}~}@-}d{LOsęۿ#^. l{O(t`sM% {W,+7{ a)tIU%2H&2T#24}qs:Y߻(Pa`_\4%cN$d]QQU!tUc]ʪ!( y4Ed]H"|Hp^wA!˴:5x0tpzI6z% -$X谟i>~|.s $%x khԻL" ńa{9e)Ŭۋ_SΛ=Q߽H;'t}=c%σ !"g`D.y(^Sc튎# Lú:K^wյ~pw7gדo3^5[r?El];d! <#׷3yo>ˑ:Zkq2@ 2&C8.Y  68#qo },VLgf3םaF*Z./߼B}3̷Vlɴ͝~EvIƯt!vR7^Vt<`{rЇSv& ut {;oXw-Tnu<~/#~&ZH/^r(<&M VǺƱӣn1iK(HFM,m~nt+n 0~XAFMrH=a9BXzG> _v> IV261H WaSUAtkm>R-T{v Mt-$q_5Z\I,;v*uZJ?b ڒO?6+>VSs~ q5 @JM"m1lzq} of_GꣻkgCw, ^&h2U#M-A17~n2 ݲe@]O=uitc u~#%/6UB%#]Oz+q>n)D:qHnG"4TEmy.SS MMVvdK9HIj#f4t].]Um}s_=fSyٟq >p4׋7o$LּY!poVlY~m>q^ʅBowo(ٰM>(mv!Me i )0/減z*.iUкgA=uىom!<5 nQ:Iog@I>IaI M$^Fioa3h  +;"EЂ0 ,[zY >_*nC.LnNʉyz.7?2|7{B|E7FsQڋ>maZ)2FgGC)"K#JmL#GWdX]F=©f;( s|&W^(n> yHeEjE3 iAg@$]~pU{-} '^^-́˻كb} CH5LK-Z>PQ{ԕڒf]Z֥&°(JYzUTude*$v.Ak?T Izf7OomnDD45똍H1UDƈR5^KmYH;Si TJKQ.W#ՈVl=1MZW0w#!eZ+yPt5T^GP]yO #VVg2%WR0Mh瀱.Q.K̤$^.M#RC,eS%]WP=!$qd}dd}&#P>r5B2H)d_Qﮗw$58!T 02L8agi!L0]>?~|^7]o>`.WWd.\`.,Ab xGÃӔ[CÃ)2>QWhD #v4%a240a>9r f<ѦN ޅs Zgd<? Gqǻ=TnCLQR)wLO$xiP[֔Qơd-胻hӣ[q'Xi&6zg؏DKP+nHʣfUN9悉  [`ȑQ }(4+@) }CT_%ED}Vir[6 #Aí!'@*tk{c5E^-PG jn'yEoggf;cն>S 뫃H!;mJḿ@d|Cx+w^>R?nUHXj}#HXWV#)šG)W=|+S8D7!Bt뫃v*tֿ-2 !E\ Cν~.E6 rsD1oxJ:xFt^zb+bYi? >^ꉴaHNg&;<Yw?_̗ FZ⛋lpİ|ÂM$}AQ5b?Xw4&bU+b9 \(JFJ{?Jm#Z[3$(eϼFPc$n(퉈 e$o{n |^&q݀l[_Ctx'"۝֬6~Fg]81m&Nͽ1|ĥj'nebཅd7Γ ıK=iUv?-C2Tw."Q$F5r^uPnX(@_iKyZVN&Y ьl !FՂсQ` ʴ*9Y8d9@`j S{Svyp޵=߮}ctӽEiIS:W=|+hyfeLNdFW"h_\g;/1ȹrPaQlFё9)>U=VλB1ې˷J_pI<2$gI>0>@{ghEP5Vb@FrSob3X Gv؏]w^L0Fj5xWsݛ _$&V"3pdžC(k5cQ?5#BxiXns [)A`\S4W(>Wq*Xr珮uu!5#,@<c ?ưc ?~Yb:;8As&@gkW`tt~t~t G9œ M0Dd> ;j-=C`BrJL@հ<O &P/ѮHmUUk-xkmW3܅@xжMcWsՅ6tC9K4n9S𷐕n/|`dž9ٕa#&\(v((9͌hR2e_K0Z DbT+ MکYjA4g ;]{Zy"cDpa "t$!\:v--=<ӅBlϝ_"cϐ@S]8S+hʹ]Qhͭ_|E`ƒk< u[==$O;ms*m_b~jtmت'{"诲yWɨm8yw1A!xz, ##&eE'80VTA4#l@%nKؤE;)u_Yބ<eSK, [8  aQ3՜b,CaVZU,z e{&&g(W)4ӸJq%=W2]J:n_iQKôIdNIћVoanQ#&Matq_vڀ)<_{ۿ`*:5& %AY9F=0Z9ׅͤde8/2,4Ą7 `%ڱ,עaE 3 .uۀRqخfu/S4~9u p47O=XKO~].g Jt2`F`?7K7~p{MfiCrgE?0~v N򧒼5x۸m]3@3,{I4J堭ʔY^_r*7&sXV.2t++,b{i#x!Vb+,5Ag娈$;,H ą$^C:X! L-,[V|SUwԅ7ѽ3Gr"5mLfB}ELd5 R12W8ǹ9qڎL] ng'U(@0D(pN1"8f5qt,/+"G!UB&htY$ yf].ҁ Q bB%ƠIqAWЀ9I&!~Vޏ}.qqqM"z_~8!05G0a NY vaQiǦPjp=^EУ&Z,N 瘵1^^ Z^39C= 4i~ZPz)L&8Ofq:l/-aّ1G/qvT2.TЀR%E3g} $z뗏0/{L^"(YEbk_D@Wf?c_7tH< eeZ;  A37Juw>Xگ羚l;P4 g9ΟĩÄ("9)>MAL>MѵC_bc\Ε@2^Ym)bZpuj'd<윤7%`^7XAuDpO%wA`JIK $ $#~ty87ΊXt8@jq.\0Jd%/F501:SMT>K㡗$sk_uZ%Ƥsc9r31 ̫W6ܐXċh:xQcS`G:X)YEۓ{,=C4p,gBy%r24(-"$o?jNVUIj wrizB<<^JӛIٶ X[ P{.s{QJ.K[)3m32F bh C#.` J1j\0$sf=@Žvr`2L`\Tch "}s tPu H#Iriz+rv)\/*s%ZOKwL+.brByP\0^lI6?Q:҂B ܰ8m}tHEkrޝ/̌Xp/%.S~H2[~t~zQMƅȠR^f{oNc]=ߗ0CXt9 S{KW&}ooA\{%*.H$T6J~syH}SW .arVс/_F2aTZ4*tZ3)Ll)I'c+H^0=PX# T79S(U+r a:+vL%QM4tpt":PHa*&uX5/bf,\JMځCjddg`ڿy@lw+歞zwPاxCyw;44JjA$qdU!E!' Q"0j9:vi9\#΄5X ~Fw5by_ZIVʓqOOhQ%Y$wF'z㼌J*twЧ[~b򫕓2 GECV܃5ƫ#y4],_R'Mk:oE}5?F٫^7qȩ 7!qyF , oϊDq(. %HMP%~,Dl|DK)lƷS# g8HE"j""VM&n>8rӚ=E7h~LXG;)_pߍYlwwOV9;qeIAtF?>.O6(2~5~YR^=ڭY/lԃ+>es]QfwSۧB $yƃYn 9 ͎NRZV?eYMGo[2j'#9]eN6Ja(z~$dj0t oRt2% ]LPۼ+04Zus%r罦`E`v;X"Y \8)zY䧣[\Yf"Hf./d^y`:M$ub!ӓ=^(JNLV`R4j;hk HIvtFLxl3d& p4-"K)yJRj(ػS_4i|֎^jD kE! }Ry9juڱލLdfkN{I{vO73)C_MQ$GJH!Tc!~I J0Y.j31zz=2NE RM#'CFX4c{1"3u4@#׸R<ge0fF2AfaŒ@}!lOx~j:qz?VFQW2Z>?n_G=_=x5P-BP 2.=3ѰaQ(!/0:~{Oww)o[}ۚxJn72[qyZfjo<[iLߵtiͻПJ:\W $Qi@%@1 ua-~@QUxilhZ9:nkԛ'֘I;s2cqfbgb&Ҋq{[[lHR<),kno!pS5Kݩ hݺ Euĺ2imݺ$Z>4]tJsϺKn]y:]hbE(0|4{/Ѻ!p)H,hυ|,-ҭ7UրxZ_,ӇeE)G~8nhho.ߨwuq_p=f}H#lp;$O6{wH]=3{F3czf=,.VUdv9$joۛ}NjwWqk"TVGʺ$cE7v7n߽} $sU٦V7!^mxX7ʻ5K->5EwcJu w'^լ4m8P%*)`IԅF~sWR4@Γri2D|2E}T e" 5 kMURHNdE5gj/5_imڋm:6L|<92xWWC:~_zRAÓ "z[i`yC$VI-W%ʘEwS'zOq;hOsdQRXt fz`;XX|y9{qC/A awU>>HPpib/)| Jߖd]ؤݼ`6r&~ڣѝAo,לpWg3yY")Yu3UH%_*N䚭~)3۬~kܹ:roQzߎ\KrgJmMxrvfBma0p@;XI: .G ]YFӧډ4TUK -m[2emkK׋‡(G^Щ,ƽo,+>VϪeX)&v7iR4sPB+JʆWZ?a&Wzs<4A%AD U `U *QCpH̒YUt-eV^/ߞcB fYwHDv|ִWV +!Bv|%T.zR%{wMJi5o6u//։i\d2B'DM[~PޏOb_@#D:H>B{!V~<9>58OqbM SHE\2)-4m%$;&!acK䬴`)ڿԠm蒏W|^p>Cמ=+hبGkɗ,l^6>%nU!f^8dMf~I޿Yn~եp<|?6M{X?󧹒޵?v+r/^Uуz7x9DK%_ޞ,i ioq9|-vrC%R ;CE%5TA@1GC7&E6\ i8lժx1m3o@B{z5m/h~h/}y7v!N8a7Q9RyO3CgZ=wW_:>}s#27?_}XT6av 'aQjtۘcE֢v|(âP(ɪ}0ߊhYQYq+ϦWcZtTK7lXsiQ!}+(âJݛr"/Q}MrE26{xV݃|+oq̐M:_J+ww%vu/l_m߼d}gEs) |ab- wB^܇Ԗŏ-' If-7 jSR4Xpc2E䒭Rlycvk󹝅mv!EFdDnq08OE| Z2I&l"G's e#Y_aU3iQ@u(9 HؤikV* 6?GZ{Ckk-\R k N1`cii TAj mZ9d\.:lE,:Z@bܚK#ٳ$Bd* c%ORٴlo!Ws|-JtӳtAN-uyJ~ʔ wLbbٛ{L7KͲ?qIƮn8NYQu2ϟ %L2`IKqw'Qex v~i5dJe] Q"(Iڄ/VR2xurR {'E_zEzbM>XQLQ eU=GcH UPZaIդyTZ J'^zjKr9{oI滓L^OBK_zjh9kd>c6ޞL6ն{\Mb[3'htI5w/ZzZJlVNBKIV-hYk)<-5 %`/ZzZj3s%l̈́2nL-VII6g^ aZc SW;d:8`?b.9$CD֏Xe6֓5 0['OBAۇe~ZA*$e|ᢶylӯ{*hNb F6^L=t wAw0xO('@"`a|ra$x e>9/A!UЖ&01VWq]B]XUT5[ j- Z#+ce"gR(ml68iLNL: 4G'A/Bm!%"@ݻ &#C-I< [p΋<ZFTta,^<@=ǒ;'uݶL!G 6V}AחlyZM筥yGt]NBKꏬr#笥.3-fs7|6r5  zfrdNw m81Aw8##"8#zԚO֗n١*g5 0[k)E,δ  SpTaAH,|UКŜ A6AdUB]d4HFhE6Cd1X~^|z;MOsݤx2FhOף].!UroǧZMÿ-ՁIjuu9z G3k8jl8ꒌI਽$՛p97K8j/%'ar?&z*N6氽) jR̢Qm=F vpTP35UBjE2˂k24X$QK/Qsf Gu޼CEc=/$anYafƅ_ؠɦϕ+цl2-::ilY5u )o!/4:8KvHPbEbI62-8d+MR56-ZŴ H&rC|m-LVpcp1]]}?*|ϿmZلuM+YSf04ؙJӓdՍ`::L/ v0K/A!On8^"7* 0_J+қEr6z|2u09/RD:Eg}YL :Bs t(C{?$ 6huY~%=8a%w+ΔuL/x{ \N< ?޸+`}N4_)Xp{9{H* >s ~ 9xzsYf4F4ԲGÓp4Ts 4v q^KK(:bc?}:k)GiJ'ʇ/p"zю=h_/g&w$S(]hm+z!>OuĚȾ3.?عWWwݯM& ?jy#way=^gEzӒlv q3c&Ls XZ0JM8@z|-2Fn u(Uth>fjMB;vtFl3M7pwYݜ<ڂZO4i=dta3ۥf hK:IBkx{Y:x{*ao뼫jwU𾼫Wk4rzqKZϗ=7\gϭɘ${GOlu^^v]",v{6~33ϦA~**9\^j傱#1O[ikxAq79˓1C(5[t\e.-rs\_ؚ-nҲD(o}z[VX JDg;h^Sr66>RPM!!߸)mO@nFHu m,%3})ٴ̎B6|"ڑ)Q(Sik.j^*V2®ihj_=$Dy urʊS/9x~يA?[jI`KƱҍ)Ŕ3SOa%^zx>%c)0ic% ӆ2bFy&}rM$:f$\z`Cl.`yZe (y`K[ynSQ*r c5{D%[Mƫ\̲}hH~z5@+{vE)=f ɓ=@ϊ'7Jd$O_F'y&Z02 ꜧ %[s@}T2me|d!rТw#Zho/0.]αZ;7 !$h jcp4;>#"#Cj Š+үZh*)aaIA UV 'RjmPjn$- EYI^QA2+j*2K ?ϾٟMnf/:?[IuYY']~M8僥䓸u/͛׃%ڪ(n$BYx5 aa,V1MEhtR1 },> ZQYRǩPJ!=7*b9%m oR: Ǣ&XF!LJr K{͹7LqpTk 湶0RJP gccFO|F؅UJ2>Vt +FHge% |Y1Ʉ&`f-32D%I=`f`wE_uCªsff %n3#Lsfli9!O&2n9r6ָ-:o8P|x6!zEţwʚVu5(e+6v-{ltG1O.|ECO7o?v?uvM.mĶxn<(gyIK={\ = ֭;]Ku= `mU\W&^e=|Hts}Ҫn_9[D=cE%<D*s. Ȓ|MEhj]VΑ_iwHtxQA>cQ͌KX!2;f1BRI{Ky5+L\~YU.E2*vyإ s ef^e 4Zl@g3JH~G0c,?Ңm:4#탠Dǣ/P]|f=]fL4)Qv:z~zMNDž_"\]q#ʖm䢇+ۛ[ mUy^ X1cu_bzzvQM$)=9jϪ{.G/Ga͉i葝Ct?͵8F<@ wNC=hg{<_;Y_|&[iԱ({+8ku%/eFPmp_m5R aM!M }a0FRэv眕M7kuQdM8P/KC|/k^ `܇_=q$b?L+&C}++_aWN5:Ttd8sڷ1$SFkN pjIܿJVS8 /gGֆ`K^ٛ%(d)VѥD<,?`ɒdkK}Tܨ*! ʃ7U.!V[Wyc>[Z%9 A7?V˫*嫁EMc2l f9N;8GSkwuEk>plvy8cᛟ3#Y^k9 ӮDPd<]Qz5ppy-:7q)_1v ?zjJea࿽T.\duZU=)[֏ QXBie@tsKC"2Z9ܒ&-$ny,uY.]ݻdhDI$fhwI=uyr%cKڙ-c_4wؗa~r/fo)VoaKf N1ͭE58BL+BG4p:&-i(CKW7d Դ)&Ї#@n^1G&mji.MNsV.[pDժ1mۆWթۆ$V 'J{54k觼j1vifg?W>Ss`M _AR MܣoW@v&7&>kp&as wd6>9U*Z귺4QKmU8LSYȪ!Yy@ɞC'n]y=3_qݞw7?`fO3YV-rJbo(@U&PPo C;a˶Bw֕G8?^캀+Q/LvW"V *nw$LMZvv= 5S|ȀiŇ7xd4! 'BϱvH^rȶ" :d{ϋ,Y>/|^}y] tQHL5`  #81Z1C/u6/0 z8Dg'Eq8A(J+;1O VbIZZv3aZ.tPQ%+<:-4KuH#0,h (Q#9:)iVUZ(AAY;MO,FBrA8,}~';-twiàCќ':Ia)3v)Tm*N- ,YSRi?aO#8WeSk8 xK pL^W/SE6X1I^5R) 7u Ŏ3 DjJ'"f@Fd\,WuMy2zvc&,©cp}uԃ`N-@M!N\N^&GSWP3[oD?||;?ebAPw˃Xz-4G#&Y* 5[/׺+TH쮂)Lo`]M EQhH[mQꠕW(Y,$S4&%Ax䁾>bu,lƕlʂqx=8xrjs< VzN&NR;.eA>\6|\殹ܢF8k/?n cif ]6|^wbWjѫ+J-hpڒb$*0%֌֠c2[λSBmE!g%ŷC#~U (Px2sއ|4?1@*JRvO=C{hP)@^7 -9Ǘnxε9&&0K縬_x0O;u'2%"C1H.r0 ̕z Z\pe\q*B`3ώLa .wΣ&$ⱬewGNNtiO 3$F( c@`s\:qcGRc !O8@ @TzSP 9%ăWNSW6> N  Ťw|Go|uV. PPz/G sy@_)G}u^Ix#g8s[ zK/jCjlHՓ/ߺC``XgdcC+f\QY5YQMw^?pˋt3_̾U]몬O׏v~?Yrz_pެ;Ցٚh(otube<{},FA2,)>uqMg:񚆸0 A\a҉Xw`f1%n.2 ;T'>Q*DŞoōM{pcuuQ0]MZ{LB ~v>r%Y0LfFxߥm`fĪ;Sj޼XÌOUbJKu2])bk޵\Uy?I+paFacA4 .&Û%/>>?"T {c’QyK}KXԡ%`WT!Z$AG%'r̵+Q[hKbۗVW ǐZ{@jE $ɨU81v>Ē*5< 'HDE`"H~9˳8sL/ q98w()89?3@rA{{ơPǼVRZ96ځK |E8e=x v;6DC@r}YI2/rAV@T^VSs*JI0 e(d'/S@fJp'Aؾ$"Kf/b-L DJ"dvLᄖr8{?'>>5NNfeqJɗb gF?rfoYߟ/9u=q\rѓSW߶$(L2-08%cP,'ٜ\qarL0VZiȁ\+ft)˦I <\"yE Bћu(ma XZP8?RXcF1frM,`WMtw2Yq <^cU+}d Q2̀d!y>n6:_9un-Uu675v;uV_nBw%xjM&s^jvWW$Z %_-|zTq;5(NuG3[ϽwxZiuqkY{YvL4BK-1K6K3L KQҝO>קh|pˠ|!-&%.X'WEH1tx5 D(+dhFV,XV?.W|.g_Lf>\:űCz9V EI(+'N󄻠?qjV<-?K-?g7IzyeA-2d۷[1D1ΠOs3i@Y[nA;$|*X I`]Oΰo[:< k&/Gۃ=dp(N$JwW>b+bKQL4<ÌJyJ:h曀dJT7!qaO>;}1'τ1㜠Ѩ,1saLsǒ0vz \j9FhTQw%K*`l*`$K*`O4Џsb"{.b<̮G슁 X0N}vCĈ V%G.\Sԅ G"-cizl02O\`  #">GQEVdYY@/Z%# U XPP%J(??yFH$1uKzV*9:}lۧ'Lfw 43F->Ang Bkyn B @)ٶesj߅?y~32ӓ-Gg'Tb=AfJ!+r+?ofz^vU/JlŐ8{P5G(\oe᯲kySr̖erhݝtDG eU/6wYCO9bh;Oe}_FǓ=ړD @D*} Yıcey ioqI䄴VےkX-d\kq'(DL)*y^ ϭc[EEvT'3v3'( v`噿:2^w%VF*!BqP%fP"qj#3XeWS}CG qVhh$LZoդuvé51ϸLpw.톻Аg):kg,DY7A%XTN;bۀ2Do̺CZ.4䙫hN?TرnʃI}Gu;Ai o/%кu!\E딐mÕ? .m_|J#$PEK9ό+ac;_?co".qpp7 Λ p.ה +M9 YӒ"LP]~<,0)}v?ly}FPd)Qkgtf>ycoxOOF벴~o!r]mm~=իޮ@Rgw|1s67[|5nimgOpoܺ ۲&/!"QvjA_9sv0 2,?d[v'6ܔ|3;^н2,NZ5r=>Hl (F)pzvjzC^"n>r3o:Gg6%˃b懟H7>V!+z \" ]gjk?{/ŬU|s5ĮXVz.43Ss)zR!F(դj!'OP[ t+BQEs2~8`3#ıs3Wݒ,l2 lzXŮkdcƔBC4}VŘ[2oՏN{pK2_ˬIjK_ߦD&5VBdw28.A׺ZxQC:%msu:/8/s~MZ'AA=)د\sZSݶ;/զHHm5:G!)n y^BF7VtAîԅtױ.BXpCC2¢Pަ(xIHW D)sB^ @@>M6rL.BWR)[j*dk &FXLEVZG"u ɢ eXPz,H"[Wjzk - Ht^.r&iww? )O[1\Ysc<"m%KO vy~O/M\/wTS{Wϩm:k*s|K(͞޻ד1̭F-_y >WJmkSah׈pP88эȯ(`;bi୕1sƒ26 Dԯ;``PPLB&X .2-)L=jFz(\ƒ寵_^$ uo8;i.|9%NOwoFn2Cﰯ N77k?ezvN,)gWE7 粐ȕͅX\:?{z^:_EtF4:xO&vkaWc>ug`P8 RNW2E5=oNzL7#X$ItW9pkG_WPkW5kk 5%#if gֆGP<T#8Wat*16BfV 6+e&[;@Bj"gkV: I dZ+^V4X[VK5Ju;;;: +5JOJpYM3R2+Q9N9e+H#9.c穾XXumҕYlVJWf-ը8RʬTJ/}Uј3>䴭Jm_kYiK R4S^wVq)Σi{R*H'A9c4 k#*ͩ ݐX`k[zm[CPmOw>l7gHc@ZT/фx8]_b 'V@ EEx8ݧ_wR^#T[*G^}-TO"h.Nbt vAZfO6QAwf!;Ef¿$In֧Yٛ٧M{}pn2R̮o T6֙jcoQ {VNTA+(j :]?4β;=?ERfMz |dG6|˱b@5%G8J!#Jdъ@i4P:om? 9frY4j+#W"GknTdQd#8j2ķ>6rt=|:kT\Ӕg-pGzH@g@eG)X%uN)iYfxyIֹ2LrJNR>ό5"y`t *YҡXuߺ4Ӭl h-{yWdi:{bP(Dxc< bf7T汅vbw<\sak#.EyKYߝe sk#kq? v۪dka)si9+4joCU}e耠H)M!M7NUp` %쓬GS}cu-!=<,a- Ь/ cPa7Z]] !VʉcBtnC+ekLwH-g=j36.(0br6ij]Hu{G*ըVzV\ҔQXreVRmsӶ2[ Ba8, xF ZDK%lSɓ&48!Gˉ&+ƬyLZ6:}Ҫ閳|{e|scz/)USXVk)<8ճe8uQh8ho)S N'Vp'4_UydTCŻe|2arr4sXN5St(Xƒd^!X7'٨6ӑ|n6 Z e" cN:Sk=Cߗ ttVy/<\̍ ]45Y;1j -ziO9Fzd2ko'>/lkm.p餯gH1K@ḏ1Ix):Aې-R~~jN4'_ĥhDF?(azTv66sSg*n=myN%{c~WqQ~L.6jB?nE6+՘ fVTdq=쭃ú[>(q%*ѡǨ#JDBѫ$"( (9J$!jlQn#G `l$Ʀ0@XR0^ƶ0ZJ+.iql "[3q>ώR"\, XJtǶx Z! ӬJ]%>BMô!,Jٶ6Qh,48Sy9+͛{ñ ,'\iwc\8s[٪؊.űOZCXz Izt֦;xKqxWVN6fΒ U?->:AcplO, Czdžl_'W!WS ҿc+aR±mMdBY5Hޑ2Ou!F ^Dydvţ|w˵oN:ywP3n,$_ bXΑ#= = @jˊCPm:+F )3yp/j|kQW`o݇}}y?+'4X?B֏C|I1v̳m͋ե퇦2s%E_[ hW"?5njz&+3kPNtv[Fsj}u#tsva`G82aJβ/Cs9,WIS`1&鮐7SN]I1^C>)X5uz򃦔rai^6y٘e4:\ggB`"8k=dI࢑GK! A@Ρ*/vJ)|: h')q:+1 ]7A 5t&RamM6B䤱"p<aSV`}=IoX3$PdA@SKdҭmd؅ UhnQQYsb$/Q|gԙbP11 @CMRfud0"b1q˲B:Wc2wDs/aiy,:oj5V$ > PNGqvl,fDYLr }n5{?XG#ucrLRH„S  pۀ罺޿5lֈA2^+4 _O?1 ٲtcܽiO 6^`!ܥW]a*ϔWLLM ћw(UKW7#91/? i(ʓx=ט=/Oލ~?o+9Tiy﻽&O#5elErٓ Hl0A7ٜwvHځ{stsiٗ#i9hu*dWOGpmw:<pk. q],~s//?2}x" wOx8MRU`M ۆJRHjBZ9=]ف3ھT \ј ~Nj2&~I"&"h,b"cqڀ`5q%$QL48ib)4Mn jTQ$*SQ[hۺxoWmA~s9.^3WN){{U]_~X|lC=SZrC)c# F| ̽H;οj.6!+]J!3+e#̄|Z.a㴕8ᡚOTg4Suƙ"u2t"<8@a5|~|\I4j!83'uZ`yAyTA,ȵR#CY0c¦3D$0 Seb[U?Zj͌#g%*f+@>*h$Gqxq-/8?IM|bM "CIF 'OKWnvV7vm:}wpΆ럾-QmR_>\_/_m~(3 e|6?|*E Su˄L3[6;{\6aRG>w=Un/&wn%6կ.8וлbb:n+F_A zD6s-)w;׈A6dEVA~#ǻ:\̻+&z),䝛MD5ӆ114%Д7e m7o;&:(+h3{iVĹ wƿZ9bMJhY*r#XL_hP\Npz=jX.U]}4UL^8b/03'yP`Κu4IAG_RlIL} )-b&NxI Py25R0Q!pb?co=HqN qZ75')2{*VsvEHFVkM5:)u, sЉyX4 d[IFfċP6@ʬъ>Ѳ Ђh[Q";Dv&89hX/\clKؘVPUR$I 1Jrܐ6E~(Iq=H*kfI9DHge:!|B}rN&"L*!)#q}X*!:cR I7j`P i}T#M*,Bۈ}'KzwwLId S5U]Pm 4u{"Kxju_ݫ˔ǖ$b @J#ֻM?j"Cl|*]r*uuΈ)0{S);z<54j1tgv; SkdS n6d[ϼSgomRu;R5ќݍP" #xS?+Dhb^7XZ73NGQ(fnSkп\kAKK`&QݣL8jj'&‚g0)jS5(o\abEc7^|d"F}ϼ {rRKm U5@J\ijnj%Ұ8w ȐXlsBGZiG~%HF2yN-%jUVQv~0:I\/a鍦>WmO]vwW#kcm=$>>ulB.f?~Ih65չ\>qUbt7^]vdv-t/~ T+ēsjƙJև a ƣW7v6H 1rHI?Ӎfy &o1CdRftߘ$pړ0}Yܺ!e'+Ѧ"JƦ疺ƫX \ҍwlC3ܞoO$ ڼn8UU6'/鹸I ˌu{k'G+뷹Sv",dҡ])zr>103!O'h9nǠQ!79ja`r)f6 ZDmkrd\7`M)u W΢Uz4) L U[ɒR4Q!2^%=@B{[F\HֆÑn::@v2'AUQ"ww@%5Zr1 gFA;WH㹜g2qjqrCҘ Hg*^EYi~\8cǯr! oZil |I/ʳ1#9\U_ uXr[84%r# yfK tԚk5,(^bq+fihu}ɧ{+AbBS>ևS#*wn%6e>WLIn3QI1C7ΙФGiNP zPHV;耮kzw{T @b}Ly< 5B ӱ Ftkm0(.(NV%&<ؘ_.YձqW`6Տt]ٻ*|6 && (Uֈŗ'- vTˌyQ)8TjuDvC8y>3{J_sL1S(=G8N;jJ\P !a/=:n~SXTMfϼ&Cs=ncn/_+Ϛqm'$Audfs)B~9ɋe4^F{K%%zHsx_]Qmo+í]7KN+GKQƻ]R`RTt4NGG4U7݌ GvNqOaؤ L%N=4JɔC#qR gmkfxmY!8rTd`mנsGɄ?)S@'5(UYd}:qhVwGI2g`KEuX#ճ:N10P:bl$3)ƶ&S]-šfaKJ F1^2NRbF20dXHnb~4Ϡ{2 ^xpOC؉OrI4w%M6^#۽܋D|1qHuظ&}(f0}V"z<|Yac< ۧ\7Cé [1&aҨ ,9YqZq&do0Lzqc_=$w{IP5c_WR?8 XH&,\)0% L1ZMi S V  5r 礷/gZh3E&R$$S1 eI&nچwmH B/)[|qIUPSQ}i{ee߸^Ӂ$GM =#RK)&s_a! vD<I/tT^vULH+$`l <[o,)K9q~ݤr*z񇿦c3׭gʚ丑_aGlLX?G; E H W\ά]:{|8d*/$2/Wti㉿FV4bΌo.FyZɫkn^c2oBw檲j %r6 x]L!Sc*&$ִwdYz)IlӪ4ڠCQѣkP1(,BeU0EcjP?]1O gyB_wW"tw5Zeu}4۔SR=1bx=kDֿ2:`rgN-h`2yUǥu_[2BM+K :5:y)M 6їgI9 x5ǰd,*ϭ0 L~fπm{3و>OςPWɡ]W+GWp]D7F!c_!@- ue#SWʝ㜰sB]iUz,|5Jѕ*$geZMKg%sg<$˚nc~BbcpFܝ{o; ;pZDBʑN5\E~up_/\ L{W/v>~Q+mh~t8pDh LGk;%z ځCF\QBqАP(^զ/e#rDPE%bDˬs I,/OlGdqD)/~s0GzjA % t8Vaucg:ݷDrHny7<),+7Q6e$ }sN_w+F6N-#׶ J?T8n;c!nce 1<(:faBz),+76;G4g7~(7<=q2&7Q@otc>}c5,'gZ]ީվ|O&xB`ijR͠ V;du$$ ==UVp(v?:a1z󮪌p8Z';Ip[t0v,đY\% [6*¡J|viW)3SsLuFC /7 gbw" FMN>Z!ɽ Tg젺.† Ud\TmŞ,bMu/!.NsCƐvAt4LGhÙo2'"ge8BNߞ&75>++[o՛ݬOJtz;ӏUcx<7U/޾!|zs?&mL_u~l2.N cBw$U4Eyu6eh РW"Gͱv&ap-߻Z& ,P xrօ`Qi4 b!cH:OO~q4Ղq8 ݆R΋QӨL? k^p2|9fGѲmT}&1Ƀd 2+(tOﱗZUJ!$4?4gM *ZEG݉^о"@ vA'wc,"%<Y>gk\\ ev"7zrBkmޖq\㶌C\kC$Z4]"\kҧ VN9v^B cN!o5*}XQ$ v+er"'fRəXBERUGBHmHѿ0ZL < r&+dno ˃1!ﯠ,Y0ދ(\߶e/{iɴv{N 9K`ȥokp9N6o kbRɻeVж*Yԡm?xe|Z"(Zbb&hCRsGUdۤgi۟ӓ`գ-j氯a4$u<稾ۧ-^ Ь ҨU)GN* S \pl56V&RK=QjmڢN!ݒk9.tmm]bajo8Y^ڷJ]+/"Ny}:7+F+-RЛS]>G>զܬTm·.Je9gťmn|V*˞TnMJJdVPsV 5ՌloVzV}n -Q~y\n՘BlEL2g0FHv,pTΖ;fz? k@-җh RIO`Tkz=Rr807we45kttEɷ3ZWٶ .T'"V%XvAv*9ϋs0'gƁfnJ_dB&80`yFN0:`>IlRlZDmtiBN#7ۺ] pޣv[ܭs»\ٴիcKmmUhrΡk8-'%%VFnSc\O3=F`҂O=Kσv46 d9/p46F)9{XaEhW٬ }T9hlUYƼ)3+]xx+o-4?AS-Z{c݉Z"v˯'wc$"w'B9+^Y ;ѳ,$Ay!_@c2׸=,ؐ1 hVyTh̺Xh |}}uʭT\q\4yGϊ2Q9O ;7 VxӧX}N`^Nlb'WyJN 9HcUV֨zly(̋{³^,y㣰bC[u(\hVe' a5zSckڼ.z9 tl2ZOadeVJN<>PO[9Gq7+j+Y);&EX)JT^Y)nv|MaheVdNJJYvX+}A^:䱤:TogWlZɲsh3Kʥw_nڜ);zVk+%vaC;` Ljj'΄&M4zX#ZobI]F_Os4T!|`dmhVR  oC&<|_~Q;R?:K~楬}n:KoGC)P?6aF,Y5Gy% A9ݘ*5VUd]|6ѡYWT'Rg)|n@]%ӹE]v maI/o~;)%RUc#T*Gf1ދ? ؙ+Gl W^i>m3Y6?4ׯnv;{O|kj3A:"ESzЁn3:n:u:[4:ⳡ")dOoatj[mUM^isۘ 9:yes.Ws1B`\2 FœB}Mx*g؝db'65KC lUGڦ&`(eƪimBKo3 5):;}=hn:90]JDƆZ7-VQm<ТY^&.*M1Eۮ`6iNuK!6d5iJ`lFk:EPDElqE@bNw%n)h:w[y测0Bn;^Hή6~ʋTVJy9x%8=:Dc"J!8u"j hY;bs-.Eǁ;֨9ܹaԵsܦ:=İW@Y;7;W53+ OܹU%gWY+oL/ﱗV_./YwˎI^Gp9<gi#'4c&/QĖ̹A-$S[KY=| rρW>qxe_y v \,y׶Bg;=A{ Yuk;ܢpg$ɘF t3itprROͮ?͓ !ggsQY8hM\zMXs^.x~ewVd g.ktpFeh闃LxAN Ule̅vGϞ%uxﺫɏPfl&omTM|9ǖڇRB yf7ckp@j筐g;XMC{ѿUcȁ-|p!|P/lAt~CT\3C6. \8cܫq +LMt*j0sFi4A>Ѓ``!vB ԣ1D#Ϝ*[_&ǎT9N!~>D Bm{dIXYIٚjt|%e(҂sC\v6EjКԦjCZg1!]4-Ebm|j9DN5R]jڼ TseRClsݦ0Ctemyo/#y5֘Gn@sZ#+a2+P EX)JT}uV Œ LJDXTSvRp2+kc">>rv.EZtmM)'W&4:nι 4.g^V&0DŽsR}v.-L+`O]F#Pgr?{`K{GQZ(=ɏ±wfB??Goِ̔aab,TT~:!n=GсmFDgD_q?1艕sע",7Fr2З ,_dQ F$0${wy3߯3#IdwOHʽŪbXO7^i NfղeOfº㹾PUll`7;V`8}*ت: z]* H ]gAv'ҘV4Jp?&z7"=M BY&'Я٤Q隸/bvw3'7V Mu)q?$!I5h~1Z, E5{zP+ȼu6BPhoi55"f^OV7iSJN~z2bwHaTW"f|s[mnw#TYnC^hNzzwVEеWZ"ޜE@߫Y"-N %;L`+{/H`) f/R^dYlI5֋HJ+|V[?u=_i9w75,،\Mvv̼ 8 !mq܀8uf5״k4I?yll42<%Ji&cK71JLi4aQnC i?3 (ml՚$`ujfiѴ]̃(]Qb}z^Hԕ@_ZjI^NL)3bb͎ZP 0Uդ]5w?}fA51 ֐Z4=b3zI2h;^fз5ɔ_Z&ww8Wa~\o6o_OKKO/'OcJH&3~2u mGb~D Veh4=cZ )[L Q?6YSMո>3: zOo߽Hmb-4-qpN^K6s[K>,oB/4/t=}ٴoNhU?=>.:6E9.zS`E+~HZj eFlVO.d ?U&Q-Pv;ύ.]ΰ#-Pa:W%aQ7sFqNʄqc;,'caB`FUwU*cWt6+ѯ~l苝B x8taB!mWNĎTN`Nkgлx;Z1á;ȥw7Tk+ObGDKe:#prkN29-:v4;O7aj-]߯w/돓@}z^n]#rcU3q4g몑v0v{Mlονs-Q.Ґ\E{2f.޳n-3Z5qU|ssjyv tL$4 ljFNw8 ?w+TXf'O\o~C^Kec ?dʆ0 mw/S- Lvf\)8Z0ܶYV pA)LXpX߲sPzޕ&gX{J3vۭ+ ؚ>’2 (+F`wZw/._tU _#j NHn)2M|57۩umN\\&N2Z'*.yMrE!ˆw:5sYa<8c>W;4:V1' R.dE(r +ᣔVQJ˜\PPk1qY ୮lŪJ }Zt/.yA'1ÊC%iRp%Сdek I_)ްdEg1|+L2gk% DiL%|/-yP_F L-jKrNr;R[R2 MVLg2QSRK ;/ rڒ<܁r+Ԉ%_[<-2ڒkK2-)0Z-sԖ1G`k8q?J(x%Wcq*Sң%8um N ˽A =Ե%LSm |jڒ)SPNaAj7-rϛJJɺ*4DfRxµJ˻tYhfVZ%Ch% [*}k4@5Z=hN|7ݗ J6{u\Ť:%<]6bE)SV]B76yc\l>r>sEҎ'Spt` O(d+ 2v7iKCʇNBɈS%y \fʥ X oya5- %%yب|rCݗe TX1m\XDJjKwUЂPҜ2NxTbפY{73BӞX_b-SS15":FQ, Ȱ>'ca Ks|Hr0Jh | S3Rmcl7$H4}䬂 k6$ u#~Ƽ\]+r:|qi >g3fӹ1⩑?.[qjg2SiӦE_t~L/{5RRd!|C[9J,u179!cIJv-S9mDsM+51#m 8[Q1M-E pp C+Ud!y@hlPA+'NI_QrkB;lff曛tݜ*kL->ѵod$?N8ja;vݬ;63烠ݯ3W*KZZ{!dWV }]=)/Uc|Vf1s? 3G?L{8d(/2͞sUe1m>ƁZ?S 8gV_}G Ѝ[<+2˳KIq7Ic4wq݊{]gT,qE4I^H7ZNt/ v2h'znpE{>tF [_NwHzDFҭxO6YSyurϹNpNײ;}Vؙݩ=7/>x0',;'0J#[yuho[к|FٸMrVG$ Viݱo68qP})lRkOoߑH駇/U$D[xI44Ii҇$Yщ$#oaI$)~qƗ^Dx H֘||j<' dn' Jv1`TdP8>$V I^[' 0h\&e;(!L󱻤l޻9ɢuQ}Mĩ}6F2\y_d ]"Ԃ {|rW$?]W9䱾w.1j}՝T%q6 \*7s>-VvwE6'Az-X$Q>sqS9K@pIV\iԍFᲑ说Dͭ, wvx 1c ?LWi)OʴtU6K斧U!t 5P $ˉ&<)v+GzQ b1BAV.nX.SBd`aIyjGi=B_B]RwN!7`;SkY%(G%j\37Bt\ " +:*0fZ #&AyBE sP*ݖƓLz!5>#^5S:e2Xk$" =6B5 9cTf>'p3ʑi`8^e{w$zWOjD&aA*(j>x1^=a M[OؘY0 ނC6ڐ57 Q$Y:]5C: (%ye?Jv `,J=/Czl<2"QIQnTd)/ žl<=7ck(oGMO? W#7hQUPgYd毜jZ >Ytt[bomb]cp7iNe0L c6''> #,L$fӂ딉h;8lxQJ0}6jX4ON.sI|ȲAe+9?R7R(q TA%Dfq%6LU*ƺC]>GNκOj7m֦h{ƈ8(0S=u6T$Z .U8K5Jd2,-\P2KJSL!a(0h:aM|9F? i] _ɬ4,82/ sfYYd– y%r&A #9aRg[oti 0S(I TX'<)s]J܀ʍDdyE uȲʲl4~lpT`b@W! |(º(:dy @?% ᨔJT֦BHOL:f*VwBL) ߝFy=\-y4ƾHUU@R%F*#|Ymq#}3iiȒm/G8eҮXBEXnUy9 {zH?w@eYBRH:6$\+˳( &e!*bҵcYo\dd1m=MWXh`[838g'ъ;VbZZIt(bA7 P]LŜTK 4d`.駈L/ϋ, vSIޭ'I{3)[$(krp '>g}+ <&˛oVFM8Q `ɶч\B@c$D6J"HlesKyy+l6|^e^xo֎FJn4"'aP6hcpvٓ+ FƗ[~k( Ij쑜irӼ(7Q+oW thc99:>=8UfBK$% @+2ȭ,dNߨDf2QYbΘ5=ٻb7+Sca3H0MSZG$P%y,-JL[e^'F$&c}tlL}1R(v3UyBv,lgZKmW{{R1Oh_?SZOf~9x_?Ӛ^ciQ^t+XT1!jFy"ô:AnOS( )ɶSc:FÁse2[11 xEQ'0,pC;%SjH^ Uþ5Q 7|҇QKHpr$ 3Dgj5 mq;>K N6[fZLU$Yۆ]5R.Y*#Ⱦ.˧勋Qu=Ƙ)ɬ>סmC9״L+ RmK7g_&8/2%#0II9H%Gu >+2LCYgH̿ ?yDcD ?K5e$4E0U$DAV$Rc.,YU$,/-STgq;[5Ԗٺ%J4D~\ԷmɶjKR:w9z+qn]A%^~R.ȕiYpkP߶K/Ki<ЏK+ h~ҋR2R\u3CnXAP+^4UU Q,ԯrv5٥#̥yOs1 8rVe3l@{ۆրCfg 4nEZ7J|{W`!S$د&gwڹP5pڹ@k~P`Z(o FG,1 H0 bIe$Md-&Q& ʢh;*=cJEPsJn4iSwM'KxKۮ9F̺.UhփZbaRXČȍ.ޫrEQ/Xk66лkQ4@@PyrKц)9}BPG}mepifx娇%2 9 oA !xL֌&T2RF")`Z=v%l Kز̰܈(UF2,iVDyF2)ف6vMYͻR'K9mjC5Sp3$PZRRLũn _"SB޵X̦%`mYB_;뽺AݙdBWwXWH*g}_uROR 6fջKF+͕/RfE \"Yǜ6np5AJ-G8[%RSA["2/D,x P+R-22^G!Q+>!ʚυ߈#[ 7-(k8ãRqRյzԴ8)塱)Ii 5vҗ(TQH8)-ۓ`)oHxOCrv/wFY.I} v65źYjt}# M`n8oLz߭わÍ//:AZv耄1!V(Δa-FONSǺ|C>0Rq*Az]1U1+C: X.sz Y8aj{]\˔MT>ט<5g(RЩ#&q0aV:+k:Qhc=cT VKAyy m$2:i(tdIۅT g"g z .4𿥡ɝsSiYrAaRzFsrb=7b2R!ŷR =GIҖJN1\_ 7>M%xUR*kDe ӯk@C_ӯ E u1l#F ^.G*s$Y}*T R3*YJG%mC8YR Z*D"s@P{ׁ 3$.˗O+)%n÷|`7 e06ц)&ߞR-5dh7PӴbWI yw.j镊 ӑ§d&c'Q ili%ښړ\0xi=zz98μYN+z8vZqF/ӊqT iQQ7KiōMHftogHH}K>NhӊIs>o% a_#ӊl"ӊQrbR'sL%֐4:rIDOO+Yab\ %۝t)?=[;ge'^->}>E" m5?b&!*kSKf>hp"ā̟Z ^x9_̾M~~3Ҕς01&N ܐU2Y%OPž,yƳ6]#C6.CQB ŋP`FPqIY՗Nx/jK!mO: {[Q<Z. MďA+/ˈfe[:8#'~©er{M@sϷA*DT.Si% 0Jp=Xo^^7]H+x-NHr0wd68t亭J~|b;ꨇE_TT aڎghY܊)ЄRslm|nt.np 8T)Zw#z_ QC"2/؂ iFCBiwcCƩ6qvf}{oMn7K7 a56%(D\<<$q̾BB DR˃Ëesaz{Rwl 0Psj#h(mw,YˈCt.sj?v>鋷.+ur?sv{W5<5Gg|3_&ټ]X}toy"U9ZX_>G`L2Abimv77*֓ELџ$Msv ڭ+ bD;hYݺEj>$.dJ飷8n]i#:]DЋ ݺn}H\D7dJDʔ:iz/fA>eq#*gv4{g^ Zi{]Qc&95>2f_G6j;fV(-hϨ=͈ .,{3ۛQ{S}x?ߗшg);^_M&^>>y@=4D ko A:׵^\~aD3l0,Z1(,2< !DH> T!9Szlx흰VYZed%4Se8ܐ ^fn=8: +iB)xpK.MmmnTģewcg{j7C3z?V_Ԏ_>v7VsqL]|`xc3)T߫Ph'ːR+8U*fN; _wNy4&R*$E".Uӌ15]&wŮ5r6t2YWE~?Zxd x6iC1=4 QaȰ `_% I2neHQhB1lWy .Tm o޽{bҰ &Lh< uIhldw IMy `m?.ҫׂ82 ?YuqB.\< l͋O,tkx}DբOӏ|u9T@(`gN,1,i6Klm*ԡw.LYkAb[jz`K[+nn|]ЎCڎ߰ŪBaw4c7Դp;!Vx$|ng_R#J:t@s3>41ʛa=sUܾ-L(i.^qeȇ+ ȻpKv7X-[R#gor_W$ WjgnT/茦7O7_!j\Qȝkn+,* 쁼 7& Õ,l%}̥eR1XVwgfN_FP5?c5`v2/MMN\Y1O--eZƏ7O^w 2VrᄍYyo#kF)ϱQl$7} <)ɱUýr܀}$DSҪ~oky/f~[BnOOfq^=|uD+}oڃ >N+\X1[.fV9Jf 7A#;yk|* 3 H3n\ Y"jH)ˍ}TJ)ЙLkeL38 HYmi$.Tը\l[?}I7׋>R!<+hӓ:z3=R~[j `nwߟU{K|;vy^߽=C*a<fg߾=bɠ˻T)*>|{rVJ9h*/[w{bp0^OM gZvIonhخ%)fZOOrPt}>~%yw.*yIFM&}a6TC I>g/ō7؏7: dh1qB ^WNw;B˻>{Srvl肀&  W.mPox:R&i]"QJ4ys FC+Ggkռ7w5G2y߿|%Оw\&R=mPzDiwDqD =N:LDb8a^baݢȝ)cJԊU-@X/ފez ,fCpw6s>.c)h!BWL8J}TVT~MYyOa˛bI>HTQ0DrɬyĢ2w/A 84Z' ̀EkMrh?*"]{Js8mgoF REK"!f2cE+391YU1iU `LHqnȉ7x#CH$~6tO -75ֺϓn7ѫ2|K_??gR H _+y"E]>+dۄ\PJm+ T(:!(T;ff*UPH>?\8gw>n ]zM|9VZW4-o$hV˼[ 3^Ch22w x.l Zs܃Y^xfpNֹb!lDDP51",s6 ],MyZ#oFix&,zμt)r-pRYy"Jf9GHպNǘ3]s:P4PNx*)Xʑ]9EEN3Ğ0@q8E1:Ә U/q9%&du؅>%f*Fa'tlP %yw-JUl>'h ^&NSFʦ< \#_PnAJO:1;H b ^lX|jI}\!@6vuK hoh0l@mS BTۘ(A:,mlFady(rECzc#yd{4,RɎkf΋g 2Ȼ g{!Q`DVcnݴID۱6iޔ&[Y\D"I0w‰zRj?s!Ȕ94Nr '/$8 W搦ҤS,( *A(O"r|)fZ[>QMT(va_g/o Ҩg )'胯} 4 Jt0Wd*ܺ:F0y}Յ5O@ kD]W#ƢQx|zzu{cV4Jٗ#' 1XgP @? HOsiPtIu֑gV='э5Rr1@+/05(5*j+nU( nMfWɨ%W[>Oʴ\%V $?xUH/m$~xd`·KkܻTmYr,su3ɉR,{ U0۵P0)ꂺJiJP-'ѷDK*&N.Z58)h΀gEp׻+XD+V͒y p0+g>FHAX'kem5ob DIpq+^6} C w sx@pMWurƁJ )(+,ޏDH$h@΃K۽\rxV|% n) Jn%?HJ8!6̲V=y Ѭf41FxkTY6Jh4v&JeE:ATTN~p_oXM~ V7~E&!ǂUCӜSw]V{h*QBujLÜU.r,Ŕ{o#5; [lx>\?m*̜1Uٟ{\I.L}Z-bJCypVaX*7cvz*~Cpya"g䄜Ė.&ZJNLWN,Mw=x")cu.Hβ` Co7RKD,hQ!<D @$ hFHV%bP06z&2 Ry"1 JK $MIZ;MͮD R>N#66u4NMX9wŝVR 6g 'j4lV&Aȧ7;Oc 4D{D q;9Fյ൒V!u0*"'--6ApAղnJ$}ȉ`0hr"a%Wa JqJO7;xrFo pIL3rvS0 CoO?^rQ8xxշϿ*pByB^(|K$!<]}V`Gm'|ڿ? }eq|N8OX!=½Ћ@;_YceZHlY?^MAv5vA@x$^ ea֍ww{U/˕D;e} 쩇*67۝?Ä}z`v8b LݍN9E.n -y!yXFż:J((XHsSéj2 i//^,u}Jh-Z29:U O>Zǹ &rnyq:7~lkxG/ʹ嗟3 R8"LrEXVGKw>Z& lcBB9{TtT7=r 4FLp&"F'4К b#!8`j$: 2LBS{`6QFR*WijU :\Oz5Pi,PKe@6L (ZZr#i,v8P׏5bc &3/!ev-.TϤleHؤx㶛ƽcWDbVJ¤^[_]$A8hHX>AdD(?˙DiE*=Vy P y,9> 614w';A`HA7YHqRH@"1ͱÉioiZ(zywwӠ?/cno,Jvzpuջr ^l}xmY4C7YcdxW~7ol`/' 25!ƫ((M}{1_/֯4bYT |e? 2cm3uķ5ּfT,U] 2>|i\^yQ{f`ɚ/O;ʼn%k*r`\%l* d@"MwG|֊yA3?! Py`:],7R aw-Lr8WAi T3וUi?n2-)ɴPH?=QhKͻ%Xh-  +.4/V5*??kh[\jy턏.ߨA" bhe˧FѤShŻ[O},A2Z2fx)T%y)ԑ}Or1v0 epT+ W'ĝ%z,8C36&c&R#6dj݃^PszK>pNPZwwןk~Tb+㏫F)ym2`{<:V`r?޹X2\|߹ ݢo0Cp.2#o2ȽEoH])5=s{M/fƓQdh5j%/=k}"(;&SAb [ʪ"z1Ec0$ JXSmp[”' b9Q9poL6ŎT 0t5Ծ2JpkM]' ݡL+rA%?gx2ю$~xdJjhi0th{ΒAIeRg_1JxaYpmoΆ\Tu.')K%=G4Fi+mC{-q5$:Vs,ZǺ9$8>PV:I2Mjh-iN')FP1je`TQ" Ш ,^*PdKEnf4%a",~qztrS ,ҮKܭXoX&_o2RJ|szhIJO\eBy$z]&&!DJ0ӌ|Ab&,0Mj{.0=l@y?ORcfڨ ԩ$/t*!?ɥ˫GW(T Q#^pT=13kl/ʀsE8}aUK]ۛjzvt8~@Eoz O]@ke d:39T:.Y \:ay0Z1 cu%lC+ԕj5;S ʴg',;1RZ:A}98SoA$aTΒ_Q޿cp"pꄳRJvj6\-&'ڜ߳1ͦ~1ZӑW\ViF ٝf D)$(Lj $B;Ticj\yB$4H HKQO۾foJQR7u`eQR\3.$RQeنyK8!BN8! XQk6Rt;*V5AXDr_PcN1x'ͬɷAHv 4Χ!fIi&G)pbzl{\g3Pgq 2}ew;42:6t)EiCcuCaZӴba1=\7uVQM]^Q̿\o :1aEʎ ϱYQ/DžJ6-3P4l( 1_8pJ?`E&kE0˪Z2/DpzV˯Kt2Z2z~tc&A;.c͓شe[gT(F>74qvIΑRu0WIgXLYC2\8k?2pY#epDH*[AqyXnN&:C. ء)75gR0vePgD4ܑĊ 鴗u(2+G9Ndd)'wCRxe h{BԒ*V޲cShGkwI@8'E+}HR71ы(" 7 z̓=QSɽܣC"UQT k4z2~s|8(D.hP!'xJAccbho%k4W_{Lz ޠb '^|JKF|F!߾F+.Em\L0*Y48An}wmmy)Z2l4}9=Exu#K$;9Pmٖ.%[;yH[p.0iC$3$ ҩ zt҇Zv&kdb$&썕tMۛ3~Z.$(2MFמ!OHFّV)nA4|D2MCcf[XwZϏ$ec$k$<'+x}'o}n+t/~w;} PR9<BM\D|bFPmٌyMu5՝αk/S`ew^zml}oSIsgEʟe)mI~Q_yŻ_#q"AgmTʍPHq=NDv`nz?۰δ{eF 08`i{ظFΒ4 XΪ?6S׭Z{g;e qIn\$ F+6';d5wDuD { m#=n^HR2c wr-ghrWK:qp'[otf~^ fϐV:]*[J"-ѽ^\s+ Gk:3L:O pZzuz2E#|Uؑ8`PXf CHO>HB#| ґ;;e|YvIVBGY Ϛ~T;L8_먆K.%]L;~ւj^nAkD6u}xqQVé9[^,޿[5_78WGzӠD>;hz@Aw<^mKßfXV&/Ip/7]|8 ~ XKHfka$ȣaKHmf6ip˄kw1%CTZ" S ۈkX|ٔZ={a::S!H>MJ)R,pBya6#"(ZDž?8%9Z<_|ib.Y rXM.'/.$.1Jp ]v?hGBpe@rEe9\ok?Jon5+w巛 9^h,x  p)NI; ܗ2J5j:&q`^+d t4٤('"])a?ƼttD#W몐`׬ [C\|E KbwV+R_@VnآF;Luy`md^EΜ"s q>~|7+0l(NiYsnkKZN#SF%d+ p1MM.{#bY_jⲒV8DmfӏӢ~{}6[ b.gփO}£yϦ֯ aKvߊ%K-ک߂)}@dܹrۇ)ޔ)C@@ukŅ'9hPz"im^&/iZ7&-$e3yyߍ=ҳ4ObdӍ3q#>fY=aя0RNzDdjF&LVu^*RV ߮tuEgi0RBkG :@mP%/y씻noWSWO|7u?rFRSBwOC館"3"|qn]|jbSĎu;[q KQR!l488?8X.ҽ CZe*/?O\"FFbȆ++vu >ȕ!͖~ѿ3i덞p!di\PhdJ@>?7z#CKjͽ SB!Ya84tEY5-C*L  ,H" SRI}idEύ;#(3{S?#{r_Vd /d)nа(B)H{IZ7X>`)oyZOuŠ a^JVoIJA_x +b%#Ok ¹NxڵnvЙXdVzjPbQR l+AFr%700?dO9\7(uV%_Nh1*r oǘ}V "xgT#>Up̲H҉ YYlx!|&I!DȚ$m<SKeց9eOVJH``GNᶵ;:؝|빓_/KIZ7v'_"XOQ"mɷ5a_gG\Xgxb&A In,4Ck|4%lbKs|@~v&Mp짦0\s&dAq`zUNY̋U~8(N]0AfI{DdƝk(ʞK"~mf.ȔYq"H 8|?4l#0F ﵞ[,. l )T"%_'.27} UN#ݍrEtH#5-`|uFKťg>")g5gF}./3F2F՞$WtdɿF>t63)I+S[,3@DFzQ靔>}ۚ\u}:=m;6xpԁ8>phnƮp`"t̒"hƱwgujê')nAMbf>0ɉ{jºfJFϪdr(%jpOH3ȅ8UUR{A96׉9HPUφP2`π<ֺ ɥOlitV1ݰIrTa0D{INٽcv(N mi ݳ'9yo=[ommF7}f&NFAa_ut܉8\Kw, u"I*ZP (a(&6N/L<:rGp?6FĝcӃMs,Fzrq51Z>zq~B;)87|ߙ/eYL0a#i5=r®qnV*wA]NH#{óg1OPmi!,m(B#jb0'#zهO}KΊ{1C.ڿ%](j+79FN{TM[r]9g s/ȦojaoN\ɞ`JĜ|,9?|>p(Qrq}{3CgHXwJ^%)q)"8l(bтJ\jCeQt̊j@k"r"s'Q͑w"yL."I P 5dWxJF!E!HxRݣ߄~JA*ͣ׮n{$ 2_40!D_4e?*,S !ǢIh6DP3e.g)@+j20i/N>[cA W#p(}F-A[*r%xCTvqB_n+j?)ݒE2$k$DV͒-!`ML'I9ɒ-S6d[gt\yAt4<7.U?7pqpv:c`_thӖNrAO egUZȤ׭`f@ uFC-<׎Y 뢒Ɋ9'@>յՎ@T\Y.X ɗDTO3䩾(\~xC\@#hq:-+ېEoڴf83 (΢َQC_u$'OesCZ]l O*+] <"cKt/=縴CL}kS@0ϗrIPd(-HcƏV3܌p (S%h7L6rJ# ,4?"Ufu\.[w__9aX4\Zb7:ƝKQޕ,׿"Kdek#.l\_OJ("O5II#"gġL6`zTu)ݵdkv[ʹx]I2 ZJdAfB&~7Ky~ʁAu:O4y*#7f0y!Y~oT1IYb-ɍ&gG0ݰ3ON{w-)W㰏9-)hXi~f٨̿kjk]B w`)H`?kv;Otq~i:4@'.V$}2I.ʮ]q9U,u<.Y| fը4@!G*Rx>唹6Ҭc} mZ\!e,Ǎ3NQaYb9nTmiVYjda&~F7RJ(4j45i08\ԛHjHi)Ў5?QU[)8D4v5sU! x1Jiv,+Rӄi2 aveńJWNeR> B7, vL`y)-0Y (!b` zJ)"\2PDŽd3Ap-(釮\җԅ0J3A;!m=شɨ1#%dlE04HRy.P[P1"Pca nޢ j]aQ62)HJ%< J'. T{Idt2ӯkšEl g]\dKuT:% 9*V<,1")"@D\e +{(.2mCpe-aiÌI"Yr^J?YBWrKB0:edNpgY 9F!4cO2 h!*J[H)q"ƣp x_7_j 0a;zK} U9\D([J)Зj a9Aeu&yWLI X5hH-6]eTǗ` 3`e(`M%0H"2%i@) Z^ C-Lu[V<#z-2I:IpIHFx]EgMI΅}~8Nn,UB3rmce 2!"5=K3WI4 \+qGEfE*D wCr3 gRvP`MƬt~T!b! r>B hĜw_SP`vjeWh~6 ](ai_68\gc_rX`/\[k@F0\RT;\3y%N7+eA KV{'8/Ydy\{ϻig2!den>U},$LuI1X>m=[C)3$_ytQ(&cv\ct9bqz8dzc\_f{E` ,TTH?Ku;hChH #Jd룂wsӬV %R-i,&(|MޭxqW>'Iʇi JBKz[秌.{OzU ߦiٟRAr>mo⟓vs`:~,oVbs`'C7L-AOHDՃ,E !ѕעc1a[ @1þ.5ot-ލ,uM>wǤn6X,4kxڢTGOHSjCgY-P`SխSy8x=>(m"/E(m$jq沪IQ3F3IW-)B~Ѽ5*42JFU:'Nc`C$.0ta*M;pO43;WZz]{=\~}f=͡}|BV;G-ק)hO\07Jp:q(Iy WRNجc麲1z=\}c0*: L "]O>W;olIJac NbB.Q/lW%tq#Pcl)0?GG\1Y9O[QJbcX@cWP42ٱc6 WF ЁKH%9K.1~,l6-ƿ iThE&JxZi)~]ѕhK,Ray"XpvOfjхzL&Wa.Вr)4 F+Ų3򐢩NJ涘edQp2a(8рHȥT ZI5  %S;wH䛚 Sf=@>-j+1ѕz]6jxK4+Tws#&./0GZ ѓs9J u%5bxko{mVpl%f`sȗ3sff)L|WzH!8ᓽ:.>WO)ָ5ŤBK۱?vǿ]gߏ:8&PB( 43FF2y78B 6^>?/t.WML+K7OҍJc ڀ4))#۫łbg<%X;_v= >fģݮcܻ92~}puk@伡yd/c-he;"hEV"ɔ= ;,zE"^ ( d-,H '/>mŐCJR#CJeq8SOUuuUOJ.9{6=](l[ TJ+Dnb{TdX+K[T/;XS#LZI#Z֢^o0j5R̐SZT,Di4R_ŵ)Å;Xǒ7UX6v )&X-vyr*MPP.gwY\3Yܟ.H,ٛϳc*Dg/:5[ẐzOzgiw$Ѯϧ!,7{*,`͙X|^zGeYtqyтmKx-ϋtt,˙!Wl]y'&BsūziH!ouo/şW ;l,p3fɌb~R&$&ӑ)"G')vU0%wx ^*jQ9Ր欮 n,)8O$ 7$yww4'ݯ^^ysb9\R%V7eUћP p?ywu5}qzʸD֜.9yqUVxqSzMJzCzDe}&3ը^).ѽa@ˁg+͝jF=3N%'@?`qI'y/ٵ_(ݥhdQ2fۭ\dݬF˿3Lk$ydĂ*E$ rGqI߫<zmWF=ZXw͙R uZdWt4׾R-)< Fd J\sRdM4 U Fpv]I_,{לŪ(CsHxmP9Ȁ*@zt6}!s#ܱJTC!omYڐh@`/2x UIӵ{C5އ]o,o KCYIQR8ޙj.dBkBF,Z^0Gw;Չ;3w @ MIacOWŨ~OW:(?F\N$EsM,(A2| MLaaG%~Dy݅0lFޛ9sd%p`S*n":ǧaJ#(L9 ׈KJz̻뀊S` Y5t!+z; fM5g1+TUGe"y~a 12SE_GntS\=5᫠} #͸茠C `Ic΄/oV?V=*ڑIUHF:ժ樥dTtFǨ>[$?]Baa$a xo~=};srݬVV)Q q<2r4dTZa|֋UhC#BU?)Gm𕵲yu5Ӫ54a򪭯'8m+s~xP MIOte(=[6 %Y.4*RkuZMds#E⮕«,\/WN.X8.ࠗPx es-"[II <qfR=-cgs-@hHsS*dTRhӯk!t4݌t1)@޺"T~-s ew2~#Y뱻e 籮dŘ%7 &7YC؄_0CCWbHD~[>Ny@>RKϯ~;tz`ÚUc%/ocpxeX/_>Xh rĪq Q(OxͰ/q1(q6AfpؘiO]e,zfzdpC&h®~?f[QR2~/ixS/P}NjnS׃;u,BE>oX!J,gK_o[1ONs V j@Xځ4\+Xd%I1e+-2P ^1-e<d4,h0TNJJ#+Y&V -Xb\F'}cmLQ>,y5QCwՃxp@6,~f̯?rn>?gn>?7rVv~-N"TVO ~%j)ޡmA[\Ȼo.E=jBͯnY^2O (׎9@s_'Wv j LݻʹC-nPm.zSc̮oU~ cqE5>W)n} waiYĊUg7)dRsYguˡ0ycgkF߿O. . ܺy\kT?Li*7{uqPȕWl_UX]K.p?:1<~{>+' w_㕛Kcʿte\+Z/b# *5>Uv_9yӃ9JfCC6J+?l=D27;(f@d8BĿ8[f A}:z!~<{B$O̽B>@J/g?cGNFPBˆ.T^/Ǖ`r6}W-AF:y~QKZKEF";M rӱQsŕu3g!l)dy/Sr#(ܯE] eϹ-[[5s||mAM}ϩPZS?$ȽK-}ubǂR&ʴ,o*^hcTݣڰcjakiY% --Tpy<sICj4rh:Ie1.]qg0Q- ҜkK)rc_[ʴi~S}N l|ݐi5 $~^UKj*܎)WeΕ[TFgթ >oUȍQ;ٕl6kkЎ'+,>_uNj8dT T ɄJXHx%I%7Y%ۤ 9?QuQFl6MuQ6F[%K?`CAl,>fSNpou %:YR悐AJ:ER[g&1he0IIL P!H 6ٴ 'uCSDWG!MB&U`55%QAIH^ob!p[^YQUumCVI4To@j]iߠoikraZ(CGֲ?^ōdjZ`U-ɄHqzhbM:ݪHֻ\_P%RXcfz{=P֕\记ZuS gYpk:@a}ڿg^,Ynڿ͙X\zYWfE-J_'CφmIKڿṈɿgAH'&Zۂzl7r A/5@vfDF¶_6Xb`d2S-գc*I,V:y!M5>(2̸#:ŎOLTLɆQE mflK@3s_@U,@Pև o]r9lbx/ {Bftd ROPSwTƩ5W#*gfAtaPmE)B-ꦖlZٓsb?92%k5EYu$%lw@ 4/._zOeLo((wM{4܋?9~Es :dFڕmL4}:(k#쩯u L)J:ƯݦM?}%`?M,Rk6ޚC@mW;NxX93볟NP*.϶k1 _]vo@+!Y(օ}ec W1nt`y6XE@tBީm|㍰;bz7Yxْ] MюO7FοtFA1nOgcO+I)E8/H;㶖pJ=6!&&eW~ss K܄;7͢5Ki?w~ןϞڿ&ܵX<am۷65\_:oɈo V-o rG7vڪcݦ Gt~cv@Shbo6m xt;HNn Gnxt 0|"ƬHXd@If4R3 Kbt@#*sV_m~~!zlag>m%oY/?H$Ʌ^}9w>gU&K÷.v9nkeQW=dN_L34Vf8T\bk D91d{Leh-I~EۇPN+'g_̮r,J73s) ӑ}7XMmuH`A1Oo It>d& 稓miDYiO!swۮ$w=N9BubO "|i,1 4g~ 7VM+ObGP#;=e>?VDZ轨aaD 'a. ^GW[pq0scozBk#a8h$s S3NE/d 봝%K;XՒU-e"O)5:P{}Աj/IlhG!Ij$ZJf[] R̓jũ@w#(pYBT )tiQ)VH8 4NhU H` P+`Q5 |wZchУըEѨ\@h\R5HT&PҶ@eT#1BPԃZZky^=F6!;G!iP=-q{R2!,}c*>|^ >Nwg6,w2tﴚmHl`K'qhh<ą4 N =uHtRGQ~LGRe}4kG#LA-/k^Pw ׇ4oO[|ۡ.7uJx}ߝ}WMeCN`eg_O3f~@77Ӂ%|;;э(}-:8$bdBu0 +(L)jbMլߢf5sg RH=;hvwěQԱnHRV\ټ?zk]g1bFC+v hjR ak!U:Go-.o4lv.ΖPrI}9g~U3Z)Wft<πm`YMJ|.8 UuM[za~`J%yV3s֜SڡQtT%qȏ ڃD 4RRR:鬉6wQO#{Zr {.e7a2a"1vRTs[?pA¯aI56sJ5J 扄 /ԥB -! vľtiS/%؄̀AKIYRx[I;4ҘEBH ֝s|&}awMgQZv/|TL5^sIcx"f.^v*]NŬsM`o:;zXt1&8sڽ3/2J$hb&c7 92GNQAr1l/ëm/ꧻI+i^ ؖC W*Tuae-X>sRƗsް}Q]HC^B;]yȲKuFԿ"،JZ8 ȬLVIkY7Q ؠ2h4R 3|(Ke] QbEjC\-Hz2X5O-|dž 5ZҬhׂ.dVb#=WJٙW:Udu锶?Wis} ,[nz; S5 6+*:|t̎wy?M|Ok޴jغ"Iں0B!ZZQkc辭0r1cb)cǢa+g _bіov)HWL.w[v:CycqWG3LkCjfڟp>h]䳛bۓVLd)SdVIχ9vMr uD"4VC̀(FVQ5T! YKU $N3l5EoC'S)4mi#Z!\kUHX  `UWURQz^z QhZ##fK!5zdlLd (EMdj$ih 75 CT@qtjFF$#P|MF|6ƽKmv1244`q[*Ɗ4 U-jjy˴m7>fqfivar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006110022115155004663017673 0ustar rootrootMar 13 10:03:44 crc systemd[1]: Starting Kubernetes Kubelet... Mar 13 10:03:44 crc restorecon[4586]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:44 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 10:03:45 crc restorecon[4586]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 10:03:45 crc restorecon[4586]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 13 10:03:47 crc kubenswrapper[4632]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 10:03:47 crc kubenswrapper[4632]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 13 10:03:47 crc kubenswrapper[4632]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 10:03:47 crc kubenswrapper[4632]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 10:03:47 crc kubenswrapper[4632]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 13 10:03:47 crc kubenswrapper[4632]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.397897 4632 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401812 4632 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401848 4632 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401862 4632 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401870 4632 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401876 4632 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401881 4632 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401886 4632 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401891 4632 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401896 4632 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401900 4632 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401904 4632 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401909 4632 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401913 4632 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401925 4632 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401930 4632 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401957 4632 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401963 4632 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401969 4632 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401975 4632 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401981 4632 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401987 4632 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401992 4632 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.401996 4632 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402001 4632 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402006 4632 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402011 4632 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402016 4632 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402021 4632 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402025 4632 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402031 4632 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402036 4632 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402041 4632 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402046 4632 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402050 4632 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402055 4632 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402060 4632 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402064 4632 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402068 4632 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402073 4632 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402077 4632 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402082 4632 feature_gate.go:330] unrecognized feature gate: Example Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402086 4632 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402090 4632 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402095 4632 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402099 4632 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402104 4632 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402108 4632 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402117 4632 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402123 4632 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402132 4632 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402137 4632 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402143 4632 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402151 4632 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402156 4632 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402160 4632 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402166 4632 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402170 4632 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402175 4632 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402179 4632 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402184 4632 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402188 4632 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402193 4632 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402197 4632 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402201 4632 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402206 4632 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402212 4632 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402217 4632 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402222 4632 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402229 4632 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402234 4632 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.402239 4632 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403275 4632 flags.go:64] FLAG: --address="0.0.0.0" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403294 4632 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403304 4632 flags.go:64] FLAG: --anonymous-auth="true" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403313 4632 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403321 4632 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403326 4632 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403334 4632 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403342 4632 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403347 4632 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403353 4632 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403360 4632 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403366 4632 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403371 4632 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403390 4632 flags.go:64] FLAG: --cgroup-root="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403402 4632 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403412 4632 flags.go:64] FLAG: --client-ca-file="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403498 4632 flags.go:64] FLAG: --cloud-config="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403507 4632 flags.go:64] FLAG: --cloud-provider="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403512 4632 flags.go:64] FLAG: --cluster-dns="[]" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403521 4632 flags.go:64] FLAG: --cluster-domain="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403526 4632 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403532 4632 flags.go:64] FLAG: --config-dir="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403537 4632 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403545 4632 flags.go:64] FLAG: --container-log-max-files="5" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403554 4632 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403559 4632 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403565 4632 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403571 4632 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403576 4632 flags.go:64] FLAG: --contention-profiling="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403581 4632 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403589 4632 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403595 4632 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403600 4632 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403607 4632 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403612 4632 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403618 4632 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403623 4632 flags.go:64] FLAG: --enable-load-reader="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403629 4632 flags.go:64] FLAG: --enable-server="true" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403636 4632 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403644 4632 flags.go:64] FLAG: --event-burst="100" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403650 4632 flags.go:64] FLAG: --event-qps="50" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403655 4632 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403660 4632 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403666 4632 flags.go:64] FLAG: --eviction-hard="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403674 4632 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403680 4632 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403684 4632 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403691 4632 flags.go:64] FLAG: --eviction-soft="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403696 4632 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403730 4632 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403737 4632 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403743 4632 flags.go:64] FLAG: --experimental-mounter-path="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403748 4632 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403753 4632 flags.go:64] FLAG: --fail-swap-on="true" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403758 4632 flags.go:64] FLAG: --feature-gates="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403765 4632 flags.go:64] FLAG: --file-check-frequency="20s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403771 4632 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403776 4632 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403782 4632 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403787 4632 flags.go:64] FLAG: --healthz-port="10248" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403793 4632 flags.go:64] FLAG: --help="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403798 4632 flags.go:64] FLAG: --hostname-override="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403803 4632 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403808 4632 flags.go:64] FLAG: --http-check-frequency="20s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403814 4632 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403819 4632 flags.go:64] FLAG: --image-credential-provider-config="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403824 4632 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403831 4632 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403837 4632 flags.go:64] FLAG: --image-service-endpoint="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403842 4632 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403847 4632 flags.go:64] FLAG: --kube-api-burst="100" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403852 4632 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403858 4632 flags.go:64] FLAG: --kube-api-qps="50" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403863 4632 flags.go:64] FLAG: --kube-reserved="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403868 4632 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403873 4632 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403879 4632 flags.go:64] FLAG: --kubelet-cgroups="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403885 4632 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403891 4632 flags.go:64] FLAG: --lock-file="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403896 4632 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403902 4632 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403907 4632 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403915 4632 flags.go:64] FLAG: --log-json-split-stream="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403920 4632 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403925 4632 flags.go:64] FLAG: --log-text-split-stream="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403931 4632 flags.go:64] FLAG: --logging-format="text" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403955 4632 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403961 4632 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403966 4632 flags.go:64] FLAG: --manifest-url="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403972 4632 flags.go:64] FLAG: --manifest-url-header="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403980 4632 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403986 4632 flags.go:64] FLAG: --max-open-files="1000000" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403993 4632 flags.go:64] FLAG: --max-pods="110" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.403998 4632 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404004 4632 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404009 4632 flags.go:64] FLAG: --memory-manager-policy="None" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404015 4632 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404021 4632 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404028 4632 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404036 4632 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404050 4632 flags.go:64] FLAG: --node-status-max-images="50" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404055 4632 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404061 4632 flags.go:64] FLAG: --oom-score-adj="-999" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404066 4632 flags.go:64] FLAG: --pod-cidr="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404071 4632 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404082 4632 flags.go:64] FLAG: --pod-manifest-path="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404087 4632 flags.go:64] FLAG: --pod-max-pids="-1" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404093 4632 flags.go:64] FLAG: --pods-per-core="0" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404099 4632 flags.go:64] FLAG: --port="10250" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404105 4632 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404111 4632 flags.go:64] FLAG: --provider-id="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404117 4632 flags.go:64] FLAG: --qos-reserved="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404122 4632 flags.go:64] FLAG: --read-only-port="10255" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404128 4632 flags.go:64] FLAG: --register-node="true" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404134 4632 flags.go:64] FLAG: --register-schedulable="true" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404139 4632 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404156 4632 flags.go:64] FLAG: --registry-burst="10" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404161 4632 flags.go:64] FLAG: --registry-qps="5" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404166 4632 flags.go:64] FLAG: --reserved-cpus="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404171 4632 flags.go:64] FLAG: --reserved-memory="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404179 4632 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404184 4632 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404190 4632 flags.go:64] FLAG: --rotate-certificates="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404195 4632 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404201 4632 flags.go:64] FLAG: --runonce="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404206 4632 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404212 4632 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404261 4632 flags.go:64] FLAG: --seccomp-default="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404267 4632 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404272 4632 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404279 4632 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404286 4632 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404292 4632 flags.go:64] FLAG: --storage-driver-password="root" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404297 4632 flags.go:64] FLAG: --storage-driver-secure="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404303 4632 flags.go:64] FLAG: --storage-driver-table="stats" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404309 4632 flags.go:64] FLAG: --storage-driver-user="root" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404315 4632 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404320 4632 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404325 4632 flags.go:64] FLAG: --system-cgroups="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404330 4632 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404339 4632 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404344 4632 flags.go:64] FLAG: --tls-cert-file="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404350 4632 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404357 4632 flags.go:64] FLAG: --tls-min-version="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404363 4632 flags.go:64] FLAG: --tls-private-key-file="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404368 4632 flags.go:64] FLAG: --topology-manager-policy="none" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404373 4632 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404378 4632 flags.go:64] FLAG: --topology-manager-scope="container" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404383 4632 flags.go:64] FLAG: --v="2" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404392 4632 flags.go:64] FLAG: --version="false" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404400 4632 flags.go:64] FLAG: --vmodule="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404407 4632 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404413 4632 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404563 4632 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404570 4632 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404575 4632 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404582 4632 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404588 4632 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404596 4632 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404611 4632 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404617 4632 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404627 4632 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404632 4632 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404638 4632 feature_gate.go:330] unrecognized feature gate: Example Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404644 4632 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404648 4632 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404653 4632 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404658 4632 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404662 4632 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404667 4632 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404671 4632 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404676 4632 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404680 4632 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404685 4632 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404691 4632 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404697 4632 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404703 4632 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404708 4632 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404713 4632 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404719 4632 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404725 4632 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404730 4632 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404736 4632 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404741 4632 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404746 4632 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404751 4632 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404757 4632 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404762 4632 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404766 4632 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404771 4632 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404778 4632 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404783 4632 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404788 4632 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404794 4632 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404799 4632 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404804 4632 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404809 4632 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404814 4632 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404819 4632 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404824 4632 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404828 4632 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404833 4632 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404837 4632 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404842 4632 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404847 4632 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404852 4632 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404856 4632 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404861 4632 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404865 4632 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404870 4632 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404875 4632 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404879 4632 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404884 4632 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404889 4632 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404893 4632 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404899 4632 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404904 4632 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404908 4632 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404913 4632 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404918 4632 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404922 4632 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404927 4632 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404931 4632 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.404966 4632 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.404983 4632 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.627477 4632 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.627543 4632 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627725 4632 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627741 4632 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627750 4632 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627757 4632 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627785 4632 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627794 4632 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627802 4632 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627809 4632 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627815 4632 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627821 4632 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627827 4632 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627833 4632 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627843 4632 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627849 4632 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627854 4632 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627859 4632 feature_gate.go:330] unrecognized feature gate: Example Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627865 4632 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627870 4632 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627875 4632 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627880 4632 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627886 4632 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627892 4632 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627897 4632 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627904 4632 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627912 4632 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627921 4632 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627930 4632 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627936 4632 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627965 4632 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627971 4632 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627976 4632 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627983 4632 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627990 4632 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.627997 4632 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628005 4632 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628012 4632 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628017 4632 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628029 4632 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628037 4632 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628045 4632 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628052 4632 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628059 4632 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628065 4632 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628070 4632 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628076 4632 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628082 4632 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628088 4632 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628094 4632 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628099 4632 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628104 4632 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628115 4632 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628123 4632 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628129 4632 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628134 4632 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628139 4632 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628144 4632 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628149 4632 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628195 4632 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628429 4632 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628501 4632 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628506 4632 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628513 4632 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628517 4632 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628524 4632 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628529 4632 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628533 4632 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628537 4632 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628542 4632 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628546 4632 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628554 4632 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628564 4632 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.628573 4632 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628769 4632 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628781 4632 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628786 4632 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628790 4632 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628795 4632 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628799 4632 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628803 4632 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628807 4632 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628810 4632 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628814 4632 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628818 4632 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628822 4632 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628826 4632 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628829 4632 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628835 4632 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628841 4632 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628846 4632 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628850 4632 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628854 4632 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628859 4632 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628863 4632 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628867 4632 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628871 4632 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628875 4632 feature_gate.go:330] unrecognized feature gate: Example Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628878 4632 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628882 4632 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628885 4632 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628890 4632 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628894 4632 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628897 4632 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628901 4632 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628906 4632 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628911 4632 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628915 4632 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628918 4632 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628922 4632 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628925 4632 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628930 4632 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628950 4632 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628954 4632 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628958 4632 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628962 4632 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628966 4632 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628970 4632 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628975 4632 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628979 4632 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628983 4632 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628990 4632 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628995 4632 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.628999 4632 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629003 4632 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629006 4632 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629010 4632 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629015 4632 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629019 4632 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629024 4632 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629028 4632 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629032 4632 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629037 4632 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629041 4632 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629045 4632 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629049 4632 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629053 4632 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629057 4632 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629062 4632 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629066 4632 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629069 4632 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629073 4632 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629077 4632 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629081 4632 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.629085 4632 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.629091 4632 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.630121 4632 server.go:940] "Client rotation is on, will bootstrap in background" Mar 13 10:03:47 crc kubenswrapper[4632]: E0313 10:03:47.635016 4632 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.639682 4632 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.639838 4632 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.642954 4632 server.go:997] "Starting client certificate rotation" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.642997 4632 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.643221 4632 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.780356 4632 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.782392 4632 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 13 10:03:47 crc kubenswrapper[4632]: E0313 10:03:47.784671 4632 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.861819 4632 log.go:25] "Validated CRI v1 runtime API" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.899782 4632 log.go:25] "Validated CRI v1 image API" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.904835 4632 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.912343 4632 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-13-09-58-50-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.912404 4632 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.930422 4632 manager.go:217] Machine: {Timestamp:2026-03-13 10:03:47.926538474 +0000 UTC m=+1.949068637 CPUVendorID:AuthenticAMD NumCores:8 NumPhysicalCores:1 NumSockets:8 CpuFrequency:2799998 MemoryCapacity:25199480832 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:e8be0c8f-16ef-4a1d-b190-772a9f649bc5 BootID:b5d63e17-4c81-494f-81b9-40163ac26c6b Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:12599738368 Type:vfs Inodes:3076108 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:5039898624 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:12599742464 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:2519945216 Type:vfs Inodes:615221 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:3076108 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:429496729600 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:35:93:68 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:35:93:68 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:06:1a:b1 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:1c:34:be Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:ec:c5:1e Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:2f:7b:ae Speed:-1 Mtu:1496} {Name:eth10 MacAddress:c2:a2:2a:ad:63:e9 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:52:1d:6e:98:22:b2 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:25199480832 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.930730 4632 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.931025 4632 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.932003 4632 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.932227 4632 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.932266 4632 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.932553 4632 topology_manager.go:138] "Creating topology manager with none policy" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.932564 4632 container_manager_linux.go:303] "Creating device plugin manager" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.933079 4632 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.933110 4632 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.934190 4632 state_mem.go:36] "Initialized new in-memory state store" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.934329 4632 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.939929 4632 kubelet.go:418] "Attempting to sync node with API server" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.939988 4632 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.940026 4632 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.940045 4632 kubelet.go:324] "Adding apiserver pod source" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.940063 4632 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.967405 4632 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.967585 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.967581 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:47 crc kubenswrapper[4632]: E0313 10:03:47.967721 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 13 10:03:47 crc kubenswrapper[4632]: E0313 10:03:47.967741 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.968556 4632 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.970127 4632 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.971658 4632 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.971685 4632 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.971693 4632 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.971702 4632 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.971720 4632 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.971729 4632 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.971738 4632 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.971751 4632 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.971762 4632 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.971771 4632 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.971784 4632 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.971792 4632 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.972873 4632 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.973489 4632 server.go:1280] "Started kubelet" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.973851 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.974574 4632 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.974561 4632 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.975580 4632 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 13 10:03:47 crc systemd[1]: Started Kubernetes Kubelet. Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.976357 4632 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.976419 4632 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.977001 4632 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.977030 4632 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.977200 4632 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 13 10:03:47 crc kubenswrapper[4632]: E0313 10:03:47.978025 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:03:47 crc kubenswrapper[4632]: E0313 10:03:47.978132 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="200ms" Mar 13 10:03:47 crc kubenswrapper[4632]: W0313 10:03:47.978446 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:47 crc kubenswrapper[4632]: E0313 10:03:47.978492 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.978840 4632 factory.go:55] Registering systemd factory Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.978856 4632 factory.go:221] Registration of the systemd container factory successfully Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.979213 4632 factory.go:153] Registering CRI-O factory Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.979229 4632 factory.go:221] Registration of the crio container factory successfully Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.979295 4632 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.979328 4632 factory.go:103] Registering Raw factory Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.979345 4632 manager.go:1196] Started watching for new ooms in manager Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.980082 4632 manager.go:319] Starting recovery of all containers Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.981045 4632 server.go:460] "Adding debug handlers to kubelet server" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.987763 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.987837 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.987863 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.987886 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.987900 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.987918 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.987954 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.987972 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988005 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988026 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988045 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988060 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988079 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988221 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988242 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988263 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988285 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988299 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988313 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988331 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988344 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988362 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988379 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988392 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988410 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988422 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988448 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988471 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988487 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988502 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988521 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988536 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988554 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988566 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988578 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988594 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988607 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988623 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988636 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988649 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988665 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988753 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988767 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988785 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988798 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988818 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988833 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988846 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988862 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988874 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988891 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.988903 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989068 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989086 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989107 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989123 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989135 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989151 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989163 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989179 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989269 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989282 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989302 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989316 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989331 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989343 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: E0313 10:03:47.986967 4632 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.182:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189c5e79366017e8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:47.9734538 +0000 UTC m=+1.995983933,LastTimestamp:2026-03-13 10:03:47.9734538 +0000 UTC m=+1.995983933,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.989353 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.990884 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.990900 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.990915 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.990932 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.990964 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.990982 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.990997 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.991008 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.991024 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.991037 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.991054 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.991067 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.991080 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.991220 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.991236 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992437 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992450 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992463 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992479 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992492 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992510 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992520 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992531 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992547 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992558 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992574 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992586 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992597 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992615 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992627 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992645 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992658 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992702 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992720 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992740 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992760 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.992780 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.994709 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.994761 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.994785 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 13 10:03:47 crc kubenswrapper[4632]: I0313 10:03:47.994813 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.001455 4632 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.001549 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.001588 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.001607 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.001630 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.001649 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.001668 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.001685 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.001696 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.001713 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.001730 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.001747 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.001866 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.002073 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.002102 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.002482 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.002531 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.002554 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.002572 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.002596 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003047 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003081 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003153 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003177 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003197 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003219 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003238 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003261 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003283 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003299 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003319 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003336 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003358 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003375 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003391 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003413 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003430 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003446 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003472 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003487 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003507 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003525 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003542 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003565 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003585 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003606 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003623 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003641 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003662 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003679 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003702 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003721 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003738 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003761 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003807 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003835 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003853 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003873 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003896 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003912 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003932 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003968 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.003984 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.004003 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.004018 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.004121 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.004147 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.004164 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.004874 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.004925 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.004982 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005005 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005021 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005039 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005094 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005113 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005137 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005156 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005180 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005197 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005215 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005241 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005259 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005282 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005334 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005352 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005426 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005445 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005467 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005485 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005501 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005524 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005544 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005628 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005652 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005668 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005688 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005705 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005720 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005741 4632 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005756 4632 reconstruct.go:97] "Volume reconstruction finished" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.005767 4632 reconciler.go:26] "Reconciler: start to sync state" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.011514 4632 manager.go:324] Recovery completed Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.024157 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.025769 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.025816 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.025829 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.031511 4632 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.031566 4632 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.031622 4632 state_mem.go:36] "Initialized new in-memory state store" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.040300 4632 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.042850 4632 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.042904 4632 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.042932 4632 kubelet.go:2335] "Starting kubelet main sync loop" Mar 13 10:03:48 crc kubenswrapper[4632]: E0313 10:03:48.043172 4632 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 13 10:03:48 crc kubenswrapper[4632]: W0313 10:03:48.047446 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:48 crc kubenswrapper[4632]: E0313 10:03:48.047737 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.055398 4632 policy_none.go:49] "None policy: Start" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.056422 4632 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.056607 4632 state_mem.go:35] "Initializing new in-memory state store" Mar 13 10:03:48 crc kubenswrapper[4632]: E0313 10:03:48.078278 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.121691 4632 manager.go:334] "Starting Device Plugin manager" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.122825 4632 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.123137 4632 server.go:79] "Starting device plugin registration server" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.123606 4632 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.123619 4632 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.123857 4632 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.124056 4632 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.124065 4632 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 13 10:03:48 crc kubenswrapper[4632]: E0313 10:03:48.131150 4632 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.143959 4632 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.144055 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.145086 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.145116 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.145125 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.145229 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.145674 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.145697 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.146373 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.146386 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.146394 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.146473 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.146797 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.146818 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.146853 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.146919 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.146929 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.147298 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.147320 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.147329 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.147386 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.147400 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.147402 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.147458 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.147527 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.147556 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.148002 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.148030 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.148039 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.148124 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.148232 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.148269 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.148511 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.148573 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.148586 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.148719 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.148736 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.148746 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.148856 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.148877 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.149701 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.149734 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.149747 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.150546 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.150606 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.150621 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:48 crc kubenswrapper[4632]: E0313 10:03:48.178867 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="400ms" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212227 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212266 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212289 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212309 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212341 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212360 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212418 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212469 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212492 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212507 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212577 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212607 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212628 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212658 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.212677 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.224089 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.225798 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.225858 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.225869 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.225902 4632 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 10:03:48 crc kubenswrapper[4632]: E0313 10:03:48.226538 4632 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.182:6443: connect: connection refused" node="crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313596 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313672 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313697 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313720 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313738 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313755 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313773 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313784 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313825 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313794 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313848 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313895 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313910 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313883 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313862 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313961 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313953 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313983 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.314006 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.314019 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.314026 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.313867 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.314006 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.314044 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.314088 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.314088 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.314108 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.314122 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.314112 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.314140 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.427349 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.429967 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.430023 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.430041 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.430078 4632 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 10:03:48 crc kubenswrapper[4632]: E0313 10:03:48.430569 4632 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.182:6443: connect: connection refused" node="crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.485748 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.495682 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.517496 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.537154 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.547275 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 10:03:48 crc kubenswrapper[4632]: W0313 10:03:48.551174 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-8a837521daf688e33fb4e6fbf5cbabe8729fa8258893d103fa454605c4a0eba7 WatchSource:0}: Error finding container 8a837521daf688e33fb4e6fbf5cbabe8729fa8258893d103fa454605c4a0eba7: Status 404 returned error can't find the container with id 8a837521daf688e33fb4e6fbf5cbabe8729fa8258893d103fa454605c4a0eba7 Mar 13 10:03:48 crc kubenswrapper[4632]: W0313 10:03:48.552743 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-f2ca7e88f9e3349411e4a1364ab9c541d5ee7284178c249a401a6700ea5f9269 WatchSource:0}: Error finding container f2ca7e88f9e3349411e4a1364ab9c541d5ee7284178c249a401a6700ea5f9269: Status 404 returned error can't find the container with id f2ca7e88f9e3349411e4a1364ab9c541d5ee7284178c249a401a6700ea5f9269 Mar 13 10:03:48 crc kubenswrapper[4632]: W0313 10:03:48.568846 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-127b7383b05ead4e61b986500f7a6fbcc9b94e8af451c60a18cbbf65b5633c3f WatchSource:0}: Error finding container 127b7383b05ead4e61b986500f7a6fbcc9b94e8af451c60a18cbbf65b5633c3f: Status 404 returned error can't find the container with id 127b7383b05ead4e61b986500f7a6fbcc9b94e8af451c60a18cbbf65b5633c3f Mar 13 10:03:48 crc kubenswrapper[4632]: W0313 10:03:48.570031 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-1cfdd9bf16760719f287994cb12d52eb0c1a9dcd9ecfd2e1c8dff79662b1e333 WatchSource:0}: Error finding container 1cfdd9bf16760719f287994cb12d52eb0c1a9dcd9ecfd2e1c8dff79662b1e333: Status 404 returned error can't find the container with id 1cfdd9bf16760719f287994cb12d52eb0c1a9dcd9ecfd2e1c8dff79662b1e333 Mar 13 10:03:48 crc kubenswrapper[4632]: W0313 10:03:48.571603 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-28009d100a8e1b9b5c2aa69cc316573e1ee346b21b26aeb590e8663fdbd671a1 WatchSource:0}: Error finding container 28009d100a8e1b9b5c2aa69cc316573e1ee346b21b26aeb590e8663fdbd671a1: Status 404 returned error can't find the container with id 28009d100a8e1b9b5c2aa69cc316573e1ee346b21b26aeb590e8663fdbd671a1 Mar 13 10:03:48 crc kubenswrapper[4632]: E0313 10:03:48.580186 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="800ms" Mar 13 10:03:48 crc kubenswrapper[4632]: W0313 10:03:48.811264 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:48 crc kubenswrapper[4632]: E0313 10:03:48.811368 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.831494 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.832847 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.832892 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.832906 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.832929 4632 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 10:03:48 crc kubenswrapper[4632]: E0313 10:03:48.833399 4632 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.182:6443: connect: connection refused" node="crc" Mar 13 10:03:48 crc kubenswrapper[4632]: I0313 10:03:48.974686 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:49 crc kubenswrapper[4632]: I0313 10:03:49.047730 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f2ca7e88f9e3349411e4a1364ab9c541d5ee7284178c249a401a6700ea5f9269"} Mar 13 10:03:49 crc kubenswrapper[4632]: I0313 10:03:49.054303 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"28009d100a8e1b9b5c2aa69cc316573e1ee346b21b26aeb590e8663fdbd671a1"} Mar 13 10:03:49 crc kubenswrapper[4632]: I0313 10:03:49.058849 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1cfdd9bf16760719f287994cb12d52eb0c1a9dcd9ecfd2e1c8dff79662b1e333"} Mar 13 10:03:49 crc kubenswrapper[4632]: I0313 10:03:49.060484 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"127b7383b05ead4e61b986500f7a6fbcc9b94e8af451c60a18cbbf65b5633c3f"} Mar 13 10:03:49 crc kubenswrapper[4632]: I0313 10:03:49.061645 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8a837521daf688e33fb4e6fbf5cbabe8729fa8258893d103fa454605c4a0eba7"} Mar 13 10:03:49 crc kubenswrapper[4632]: W0313 10:03:49.085046 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:49 crc kubenswrapper[4632]: E0313 10:03:49.085148 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 13 10:03:49 crc kubenswrapper[4632]: E0313 10:03:49.381061 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="1.6s" Mar 13 10:03:49 crc kubenswrapper[4632]: W0313 10:03:49.384694 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:49 crc kubenswrapper[4632]: E0313 10:03:49.384757 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 13 10:03:49 crc kubenswrapper[4632]: W0313 10:03:49.443649 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:49 crc kubenswrapper[4632]: E0313 10:03:49.443759 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 13 10:03:49 crc kubenswrapper[4632]: I0313 10:03:49.633605 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:49 crc kubenswrapper[4632]: I0313 10:03:49.635861 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:49 crc kubenswrapper[4632]: I0313 10:03:49.635902 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:49 crc kubenswrapper[4632]: I0313 10:03:49.635914 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:49 crc kubenswrapper[4632]: I0313 10:03:49.635956 4632 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 10:03:49 crc kubenswrapper[4632]: E0313 10:03:49.636827 4632 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.182:6443: connect: connection refused" node="crc" Mar 13 10:03:49 crc kubenswrapper[4632]: I0313 10:03:49.834419 4632 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 10:03:49 crc kubenswrapper[4632]: E0313 10:03:49.835833 4632 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 13 10:03:49 crc kubenswrapper[4632]: I0313 10:03:49.975236 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.068238 4632 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611" exitCode=0 Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.068323 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611"} Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.068369 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.069630 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.069658 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.069667 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.071288 4632 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52" exitCode=0 Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.071358 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52"} Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.071424 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.077100 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.077134 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.077144 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.079732 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"2666e43569f4f3fb7d9deec4b70dc873d86a3a7c4ac2fac7eea45198e35ecf3a"} Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.079766 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"cc37f87081e3682692ee20f12c80aa65fbcb8604b381f313b8073f8019b96dbc"} Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.079775 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac"} Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.079785 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d"} Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.079812 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.080799 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.080836 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.080847 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.082717 4632 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990" exitCode=0 Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.082796 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990"} Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.082824 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.083505 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.083521 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.083530 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.087988 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.088358 4632 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="7b16440f9dc548d378b6415b89f879e2684e2ea5d0284feb13fcf67f4fa9fa81" exitCode=0 Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.088419 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"7b16440f9dc548d378b6415b89f879e2684e2ea5d0284feb13fcf67f4fa9fa81"} Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.088615 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.089145 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.089168 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.089176 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.089506 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.089546 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.089560 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:50 crc kubenswrapper[4632]: I0313 10:03:50.658356 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:03:51 crc kubenswrapper[4632]: W0313 10:03:51.001267 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:51 crc kubenswrapper[4632]: E0313 10:03:51.001372 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 13 10:03:51 crc kubenswrapper[4632]: E0313 10:03:51.001469 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="3.2s" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.001628 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:51 crc kubenswrapper[4632]: W0313 10:03:51.102099 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:51 crc kubenswrapper[4632]: E0313 10:03:51.102192 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.146550 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ecc4f274e9e301fdd8acbd720c998a2aa1c5e00df4fc254bcacab4acb539b8ce"} Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.146601 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"9bd8f71a5e4bfa40758e7d51545f1b2eff43f11071060201770a574f89d391bc"} Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.146612 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1"} Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.148533 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf"} Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.148572 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe"} Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.148581 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54"} Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.149861 4632 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3322cab0328fe5a4dfa2407db67435002bef607e711821802d5e9a81ef8c8476" exitCode=0 Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.149895 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3322cab0328fe5a4dfa2407db67435002bef607e711821802d5e9a81ef8c8476"} Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.150012 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.150953 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.150972 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.150999 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.167157 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.167688 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.167952 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811"} Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.168414 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.168433 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.168441 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.168975 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.169006 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.169016 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.177143 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.178261 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.178298 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.178310 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.263169 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.265537 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.265579 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.265594 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.265637 4632 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 10:03:51 crc kubenswrapper[4632]: E0313 10:03:51.266091 4632 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.182:6443: connect: connection refused" node="crc" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.353272 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 10:03:51 crc kubenswrapper[4632]: W0313 10:03:51.397215 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:51 crc kubenswrapper[4632]: E0313 10:03:51.397341 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.182:6443: connect: connection refused" logger="UnhandledError" Mar 13 10:03:51 crc kubenswrapper[4632]: E0313 10:03:51.406297 4632 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.182:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189c5e79366017e8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:47.9734538 +0000 UTC m=+1.995983933,LastTimestamp:2026-03-13 10:03:47.9734538 +0000 UTC m=+1.995983933,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:03:51 crc kubenswrapper[4632]: I0313 10:03:51.974626 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.182:6443: connect: connection refused Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.179652 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7b911e0926eac089864b65a40a27d5996412dc7dd93176dc1472b5e6fee82ee0"} Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.179730 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc"} Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.179733 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.181204 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.181259 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.181273 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.184586 4632 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="462f33d50e7ba9db6ec487a1cdd2e211e7969591a4699212fc2fa82f4ce990c8" exitCode=0 Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.184754 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.184788 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.184871 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"462f33d50e7ba9db6ec487a1cdd2e211e7969591a4699212fc2fa82f4ce990c8"} Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.184986 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.185681 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.186024 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.186075 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.186101 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.186212 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.186243 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.186221 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.186287 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.186302 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.186254 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.186678 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.186704 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.186714 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:52 crc kubenswrapper[4632]: I0313 10:03:52.434363 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:03:53 crc kubenswrapper[4632]: I0313 10:03:53.191501 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0f16c27db31aa07efc2dc28bf5f70651f3f4fc43df633bf31c9cb10e7d7d3305"} Mar 13 10:03:53 crc kubenswrapper[4632]: I0313 10:03:53.191556 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"18912268581dca1cde5072a095c27c7cba60d75ef23e45af7fd8b89dbf56ecfe"} Mar 13 10:03:53 crc kubenswrapper[4632]: I0313 10:03:53.191571 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"9415e213590ed2127f1b3ccc09bb2a37fde1a6a96d71e06b5a599cb4264ae0b9"} Mar 13 10:03:53 crc kubenswrapper[4632]: I0313 10:03:53.191592 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"13da8bb63b5666f05bbda4e4ade0cdc7995d2d3562fcbb0f4e5b68330aad8232"} Mar 13 10:03:53 crc kubenswrapper[4632]: I0313 10:03:53.191607 4632 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 10:03:53 crc kubenswrapper[4632]: I0313 10:03:53.191684 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:53 crc kubenswrapper[4632]: I0313 10:03:53.191698 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:53 crc kubenswrapper[4632]: I0313 10:03:53.192849 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:53 crc kubenswrapper[4632]: I0313 10:03:53.192877 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:53 crc kubenswrapper[4632]: I0313 10:03:53.192887 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:53 crc kubenswrapper[4632]: I0313 10:03:53.193818 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:53 crc kubenswrapper[4632]: I0313 10:03:53.193842 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:53 crc kubenswrapper[4632]: I0313 10:03:53.193850 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.062780 4632 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.198811 4632 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.198870 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.199565 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.199835 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5b9bb5108316be08fc018d2bbe5bb5c3f1e0728b8aa5598243f55c27332ef9dd"} Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.200348 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.200377 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.200388 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.200984 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.201009 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.201022 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.466499 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.468196 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.468249 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.468279 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.468313 4632 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.523723 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.523908 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.525005 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.525038 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.525047 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.903083 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:03:54 crc kubenswrapper[4632]: I0313 10:03:54.954166 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 13 10:03:55 crc kubenswrapper[4632]: I0313 10:03:55.159798 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:03:55 crc kubenswrapper[4632]: I0313 10:03:55.201242 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:55 crc kubenswrapper[4632]: I0313 10:03:55.201242 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:55 crc kubenswrapper[4632]: I0313 10:03:55.201243 4632 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 10:03:55 crc kubenswrapper[4632]: I0313 10:03:55.201448 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:55 crc kubenswrapper[4632]: I0313 10:03:55.202180 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:55 crc kubenswrapper[4632]: I0313 10:03:55.202220 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:55 crc kubenswrapper[4632]: I0313 10:03:55.202236 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:55 crc kubenswrapper[4632]: I0313 10:03:55.202265 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:55 crc kubenswrapper[4632]: I0313 10:03:55.202285 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:55 crc kubenswrapper[4632]: I0313 10:03:55.202294 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:55 crc kubenswrapper[4632]: I0313 10:03:55.202452 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:55 crc kubenswrapper[4632]: I0313 10:03:55.202470 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:55 crc kubenswrapper[4632]: I0313 10:03:55.202479 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:56 crc kubenswrapper[4632]: I0313 10:03:56.088702 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:03:56 crc kubenswrapper[4632]: I0313 10:03:56.191920 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 13 10:03:56 crc kubenswrapper[4632]: I0313 10:03:56.207007 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:56 crc kubenswrapper[4632]: I0313 10:03:56.207186 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:56 crc kubenswrapper[4632]: I0313 10:03:56.207882 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:56 crc kubenswrapper[4632]: I0313 10:03:56.207909 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:56 crc kubenswrapper[4632]: I0313 10:03:56.207919 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:56 crc kubenswrapper[4632]: I0313 10:03:56.208860 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:56 crc kubenswrapper[4632]: I0313 10:03:56.208881 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:56 crc kubenswrapper[4632]: I0313 10:03:56.208889 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:57 crc kubenswrapper[4632]: I0313 10:03:57.209826 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:57 crc kubenswrapper[4632]: I0313 10:03:57.211122 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:57 crc kubenswrapper[4632]: I0313 10:03:57.211179 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:57 crc kubenswrapper[4632]: I0313 10:03:57.211203 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:58 crc kubenswrapper[4632]: E0313 10:03:58.131323 4632 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 10:03:58 crc kubenswrapper[4632]: I0313 10:03:58.160038 4632 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:03:58 crc kubenswrapper[4632]: I0313 10:03:58.160153 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:03:59 crc kubenswrapper[4632]: I0313 10:03:59.463997 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:03:59 crc kubenswrapper[4632]: I0313 10:03:59.464287 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:03:59 crc kubenswrapper[4632]: I0313 10:03:59.468022 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:03:59 crc kubenswrapper[4632]: I0313 10:03:59.468097 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:03:59 crc kubenswrapper[4632]: I0313 10:03:59.468129 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:03:59 crc kubenswrapper[4632]: I0313 10:03:59.474003 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:04:00 crc kubenswrapper[4632]: I0313 10:04:00.218000 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:00 crc kubenswrapper[4632]: I0313 10:04:00.219863 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:00 crc kubenswrapper[4632]: I0313 10:04:00.219973 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:00 crc kubenswrapper[4632]: I0313 10:04:00.219991 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:00 crc kubenswrapper[4632]: I0313 10:04:00.222810 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:04:01 crc kubenswrapper[4632]: I0313 10:04:01.220760 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:01 crc kubenswrapper[4632]: I0313 10:04:01.222091 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:01 crc kubenswrapper[4632]: I0313 10:04:01.222151 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:01 crc kubenswrapper[4632]: I0313 10:04:01.222164 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:02 crc kubenswrapper[4632]: W0313 10:04:02.403085 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 13 10:04:02 crc kubenswrapper[4632]: I0313 10:04:02.403278 4632 trace.go:236] Trace[1189007437]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Mar-2026 10:03:52.401) (total time: 10001ms): Mar 13 10:04:02 crc kubenswrapper[4632]: Trace[1189007437]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (10:04:02.403) Mar 13 10:04:02 crc kubenswrapper[4632]: Trace[1189007437]: [10.001998026s] [10.001998026s] END Mar 13 10:04:02 crc kubenswrapper[4632]: E0313 10:04:02.403323 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 13 10:04:02 crc kubenswrapper[4632]: I0313 10:04:02.434546 4632 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:04:02 crc kubenswrapper[4632]: I0313 10:04:02.434977 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:04:02 crc kubenswrapper[4632]: I0313 10:04:02.976632 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 13 10:04:03 crc kubenswrapper[4632]: I0313 10:04:03.226421 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 13 10:04:03 crc kubenswrapper[4632]: I0313 10:04:03.227842 4632 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7b911e0926eac089864b65a40a27d5996412dc7dd93176dc1472b5e6fee82ee0" exitCode=255 Mar 13 10:04:03 crc kubenswrapper[4632]: I0313 10:04:03.228018 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7b911e0926eac089864b65a40a27d5996412dc7dd93176dc1472b5e6fee82ee0"} Mar 13 10:04:03 crc kubenswrapper[4632]: I0313 10:04:03.228246 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:03 crc kubenswrapper[4632]: I0313 10:04:03.229120 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:03 crc kubenswrapper[4632]: I0313 10:04:03.229234 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:03 crc kubenswrapper[4632]: I0313 10:04:03.229326 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:03 crc kubenswrapper[4632]: I0313 10:04:03.230065 4632 scope.go:117] "RemoveContainer" containerID="7b911e0926eac089864b65a40a27d5996412dc7dd93176dc1472b5e6fee82ee0" Mar 13 10:04:04 crc kubenswrapper[4632]: E0313 10:04:04.065306 4632 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 13 10:04:04 crc kubenswrapper[4632]: E0313 10:04:04.203243 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="6.4s" Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.232666 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.234670 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d83208b114211c941c04a286b532683c256ad92512d1e3fac27e249095b31d4a"} Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.234817 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.235731 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.235757 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.235765 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:04 crc kubenswrapper[4632]: E0313 10:04:04.469410 4632 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": net/http: TLS handshake timeout" node="crc" Mar 13 10:04:04 crc kubenswrapper[4632]: W0313 10:04:04.806568 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:04Z is after 2026-02-23T05:33:13Z Mar 13 10:04:04 crc kubenswrapper[4632]: E0313 10:04:04.806660 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:04Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 10:04:04 crc kubenswrapper[4632]: W0313 10:04:04.811442 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:04Z is after 2026-02-23T05:33:13Z Mar 13 10:04:04 crc kubenswrapper[4632]: E0313 10:04:04.811827 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:04Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 10:04:04 crc kubenswrapper[4632]: W0313 10:04:04.813258 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:04Z is after 2026-02-23T05:33:13Z Mar 13 10:04:04 crc kubenswrapper[4632]: E0313 10:04:04.813364 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:04Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.814769 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:04Z is after 2026-02-23T05:33:13Z Mar 13 10:04:04 crc kubenswrapper[4632]: E0313 10:04:04.818452 4632 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:04Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c5e79366017e8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:47.9734538 +0000 UTC m=+1.995983933,LastTimestamp:2026-03-13 10:03:47.9734538 +0000 UTC m=+1.995983933,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.859436 4632 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.859839 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.987995 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.988591 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.990363 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.990412 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.990421 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:04 crc kubenswrapper[4632]: I0313 10:04:04.993970 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:04Z is after 2026-02-23T05:33:13Z Mar 13 10:04:05 crc kubenswrapper[4632]: I0313 10:04:05.004695 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 13 10:04:05 crc kubenswrapper[4632]: I0313 10:04:05.247418 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:05 crc kubenswrapper[4632]: I0313 10:04:05.248517 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:05 crc kubenswrapper[4632]: I0313 10:04:05.248677 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:05 crc kubenswrapper[4632]: I0313 10:04:05.248750 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:05 crc kubenswrapper[4632]: I0313 10:04:05.977860 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:05Z is after 2026-02-23T05:33:13Z Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.088751 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.089039 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.090300 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.090419 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.090527 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.252279 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.253545 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.255931 4632 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d83208b114211c941c04a286b532683c256ad92512d1e3fac27e249095b31d4a" exitCode=255 Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.256002 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d83208b114211c941c04a286b532683c256ad92512d1e3fac27e249095b31d4a"} Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.256040 4632 scope.go:117] "RemoveContainer" containerID="7b911e0926eac089864b65a40a27d5996412dc7dd93176dc1472b5e6fee82ee0" Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.256498 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.258248 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.258311 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.258325 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.259545 4632 scope.go:117] "RemoveContainer" containerID="d83208b114211c941c04a286b532683c256ad92512d1e3fac27e249095b31d4a" Mar 13 10:04:06 crc kubenswrapper[4632]: E0313 10:04:06.259984 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 10:04:06 crc kubenswrapper[4632]: I0313 10:04:06.978694 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:06Z is after 2026-02-23T05:33:13Z Mar 13 10:04:07 crc kubenswrapper[4632]: I0313 10:04:07.261439 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 10:04:07 crc kubenswrapper[4632]: I0313 10:04:07.441894 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:04:07 crc kubenswrapper[4632]: I0313 10:04:07.442148 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:07 crc kubenswrapper[4632]: I0313 10:04:07.443502 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:07 crc kubenswrapper[4632]: I0313 10:04:07.443538 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:07 crc kubenswrapper[4632]: I0313 10:04:07.443550 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:07 crc kubenswrapper[4632]: I0313 10:04:07.444272 4632 scope.go:117] "RemoveContainer" containerID="d83208b114211c941c04a286b532683c256ad92512d1e3fac27e249095b31d4a" Mar 13 10:04:07 crc kubenswrapper[4632]: E0313 10:04:07.444455 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 10:04:07 crc kubenswrapper[4632]: I0313 10:04:07.447398 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:04:07 crc kubenswrapper[4632]: W0313 10:04:07.901792 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:07Z is after 2026-02-23T05:33:13Z Mar 13 10:04:07 crc kubenswrapper[4632]: E0313 10:04:07.901903 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:07Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 10:04:07 crc kubenswrapper[4632]: I0313 10:04:07.978303 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:07Z is after 2026-02-23T05:33:13Z Mar 13 10:04:08 crc kubenswrapper[4632]: E0313 10:04:08.131586 4632 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 10:04:08 crc kubenswrapper[4632]: I0313 10:04:08.164291 4632 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:04:08 crc kubenswrapper[4632]: I0313 10:04:08.164380 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:04:08 crc kubenswrapper[4632]: I0313 10:04:08.266204 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:08 crc kubenswrapper[4632]: I0313 10:04:08.267473 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:08 crc kubenswrapper[4632]: I0313 10:04:08.267544 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:08 crc kubenswrapper[4632]: I0313 10:04:08.267579 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:08 crc kubenswrapper[4632]: I0313 10:04:08.268147 4632 scope.go:117] "RemoveContainer" containerID="d83208b114211c941c04a286b532683c256ad92512d1e3fac27e249095b31d4a" Mar 13 10:04:08 crc kubenswrapper[4632]: E0313 10:04:08.268309 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 10:04:08 crc kubenswrapper[4632]: I0313 10:04:08.979660 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:08Z is after 2026-02-23T05:33:13Z Mar 13 10:04:09 crc kubenswrapper[4632]: I0313 10:04:09.977555 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:09Z is after 2026-02-23T05:33:13Z Mar 13 10:04:10 crc kubenswrapper[4632]: E0313 10:04:10.606863 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:10Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 10:04:10 crc kubenswrapper[4632]: I0313 10:04:10.870248 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:10 crc kubenswrapper[4632]: I0313 10:04:10.871669 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:10 crc kubenswrapper[4632]: I0313 10:04:10.871709 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:10 crc kubenswrapper[4632]: I0313 10:04:10.871719 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:10 crc kubenswrapper[4632]: I0313 10:04:10.871746 4632 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 10:04:10 crc kubenswrapper[4632]: E0313 10:04:10.875571 4632 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:10Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 10:04:10 crc kubenswrapper[4632]: I0313 10:04:10.978860 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:04:10Z is after 2026-02-23T05:33:13Z Mar 13 10:04:11 crc kubenswrapper[4632]: W0313 10:04:11.420698 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 13 10:04:11 crc kubenswrapper[4632]: E0313 10:04:11.421033 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 13 10:04:11 crc kubenswrapper[4632]: I0313 10:04:11.979671 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:12 crc kubenswrapper[4632]: I0313 10:04:12.189607 4632 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 10:04:12 crc kubenswrapper[4632]: I0313 10:04:12.210985 4632 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 13 10:04:12 crc kubenswrapper[4632]: W0313 10:04:12.864606 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:12 crc kubenswrapper[4632]: E0313 10:04:12.864682 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 13 10:04:12 crc kubenswrapper[4632]: I0313 10:04:12.978865 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:13 crc kubenswrapper[4632]: I0313 10:04:13.979344 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.824701 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79366017e8 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:47.9734538 +0000 UTC m=+1.995983933,LastTimestamp:2026-03-13 10:03:47.9734538 +0000 UTC m=+1.995983933,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.830225 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397edfe0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.02580272 +0000 UTC m=+2.048332843,LastTimestamp:2026-03-13 10:03:48.02580272 +0000 UTC m=+2.048332843,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.831765 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f2ff5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025823221 +0000 UTC m=+2.048353354,LastTimestamp:2026-03-13 10:03:48.025823221 +0000 UTC m=+2.048353354,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.837061 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f5877 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025833591 +0000 UTC m=+2.048363724,LastTimestamp:2026-03-13 10:03:48.025833591 +0000 UTC m=+2.048363724,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.842283 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e793f688428 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.125000744 +0000 UTC m=+2.147530877,LastTimestamp:2026-03-13 10:03:48.125000744 +0000 UTC m=+2.147530877,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.847574 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397edfe0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397edfe0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.02580272 +0000 UTC m=+2.048332843,LastTimestamp:2026-03-13 10:03:48.14510877 +0000 UTC m=+2.167638903,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.852547 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397f2ff5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f2ff5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025823221 +0000 UTC m=+2.048353354,LastTimestamp:2026-03-13 10:03:48.1451216 +0000 UTC m=+2.167651733,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.858765 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397f5877\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f5877 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025833591 +0000 UTC m=+2.048363724,LastTimestamp:2026-03-13 10:03:48.145129211 +0000 UTC m=+2.167659344,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.863981 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397edfe0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397edfe0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.02580272 +0000 UTC m=+2.048332843,LastTimestamp:2026-03-13 10:03:48.146382802 +0000 UTC m=+2.168912935,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.869984 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397f2ff5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f2ff5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025823221 +0000 UTC m=+2.048353354,LastTimestamp:2026-03-13 10:03:48.146391232 +0000 UTC m=+2.168921365,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.874808 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397f5877\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f5877 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025833591 +0000 UTC m=+2.048363724,LastTimestamp:2026-03-13 10:03:48.146398402 +0000 UTC m=+2.168928535,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.879355 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397edfe0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397edfe0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.02580272 +0000 UTC m=+2.048332843,LastTimestamp:2026-03-13 10:03:48.146910214 +0000 UTC m=+2.169440337,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.884482 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397f2ff5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f2ff5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025823221 +0000 UTC m=+2.048353354,LastTimestamp:2026-03-13 10:03:48.146926325 +0000 UTC m=+2.169456458,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.889126 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397f5877\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f5877 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025833591 +0000 UTC m=+2.048363724,LastTimestamp:2026-03-13 10:03:48.146956875 +0000 UTC m=+2.169486998,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.896289 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397edfe0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397edfe0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.02580272 +0000 UTC m=+2.048332843,LastTimestamp:2026-03-13 10:03:48.147313284 +0000 UTC m=+2.169843417,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.902553 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397f2ff5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f2ff5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025823221 +0000 UTC m=+2.048353354,LastTimestamp:2026-03-13 10:03:48.147326084 +0000 UTC m=+2.169856217,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.909793 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397f5877\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f5877 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025833591 +0000 UTC m=+2.048363724,LastTimestamp:2026-03-13 10:03:48.147333314 +0000 UTC m=+2.169863447,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.916131 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397edfe0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397edfe0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.02580272 +0000 UTC m=+2.048332843,LastTimestamp:2026-03-13 10:03:48.147398086 +0000 UTC m=+2.169928209,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.920521 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397f2ff5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f2ff5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025823221 +0000 UTC m=+2.048353354,LastTimestamp:2026-03-13 10:03:48.147450927 +0000 UTC m=+2.169981060,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.925465 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397f5877\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f5877 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025833591 +0000 UTC m=+2.048363724,LastTimestamp:2026-03-13 10:03:48.147464837 +0000 UTC m=+2.169994960,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.930599 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397edfe0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397edfe0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.02580272 +0000 UTC m=+2.048332843,LastTimestamp:2026-03-13 10:03:48.148020562 +0000 UTC m=+2.170550695,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.935402 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397f2ff5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f2ff5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025823221 +0000 UTC m=+2.048353354,LastTimestamp:2026-03-13 10:03:48.148036082 +0000 UTC m=+2.170566215,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.940082 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397f5877\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f5877 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025833591 +0000 UTC m=+2.048363724,LastTimestamp:2026-03-13 10:03:48.148044682 +0000 UTC m=+2.170574815,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.948511 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397edfe0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397edfe0 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.02580272 +0000 UTC m=+2.048332843,LastTimestamp:2026-03-13 10:03:48.148560545 +0000 UTC m=+2.171090678,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.952770 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5e79397f2ff5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5e79397f2ff5 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.025823221 +0000 UTC m=+2.048353354,LastTimestamp:2026-03-13 10:03:48.148581795 +0000 UTC m=+2.171111928,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.958641 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5e795950ab80 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.559645568 +0000 UTC m=+2.582175701,LastTimestamp:2026-03-13 10:03:48.559645568 +0000 UTC m=+2.582175701,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.964130 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e795951112e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.559671598 +0000 UTC m=+2.582201731,LastTimestamp:2026-03-13 10:03:48.559671598 +0000 UTC m=+2.582201731,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.968636 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e795a3382cc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.57451182 +0000 UTC m=+2.597041943,LastTimestamp:2026-03-13 10:03:48.57451182 +0000 UTC m=+2.597041943,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.973208 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e795a4077c9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.575360969 +0000 UTC m=+2.597891102,LastTimestamp:2026-03-13 10:03:48.575360969 +0000 UTC m=+2.597891102,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: I0313 10:04:14.980332 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.980919 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5e795b989d15 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:48.597914901 +0000 UTC m=+2.620445034,LastTimestamp:2026-03-13 10:03:48.597914901 +0000 UTC m=+2.620445034,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.987069 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5e798179cdf1 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.233430001 +0000 UTC m=+3.255960134,LastTimestamp:2026-03-13 10:03:49.233430001 +0000 UTC m=+3.255960134,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.991511 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7981984065 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.235425381 +0000 UTC m=+3.257955514,LastTimestamp:2026-03-13 10:03:49.235425381 +0000 UTC m=+3.257955514,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:14 crc kubenswrapper[4632]: E0313 10:04:14.995728 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e7981985221 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.235429921 +0000 UTC m=+3.257960054,LastTimestamp:2026-03-13 10:03:49.235429921 +0000 UTC m=+3.257960054,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.000195 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e79819bd53e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.235660094 +0000 UTC m=+3.258190227,LastTimestamp:2026-03-13 10:03:49.235660094 +0000 UTC m=+3.258190227,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.005466 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5e7981a62fef openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.236338671 +0000 UTC m=+3.258868804,LastTimestamp:2026-03-13 10:03:49.236338671 +0000 UTC m=+3.258868804,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.010439 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5e798285710f openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.250969871 +0000 UTC m=+3.273500004,LastTimestamp:2026-03-13 10:03:49.250969871 +0000 UTC m=+3.273500004,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.015754 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e7982958adb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.252025051 +0000 UTC m=+3.274555184,LastTimestamp:2026-03-13 10:03:49.252025051 +0000 UTC m=+3.274555184,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.021883 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e79829857b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.252208563 +0000 UTC m=+3.274738696,LastTimestamp:2026-03-13 10:03:49.252208563 +0000 UTC m=+3.274738696,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.026300 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e79829e7e2c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.252611628 +0000 UTC m=+3.275141761,LastTimestamp:2026-03-13 10:03:49.252611628 +0000 UTC m=+3.275141761,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: I0313 10:04:15.031048 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.031083 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5e79829fc215 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.252694549 +0000 UTC m=+3.275224692,LastTimestamp:2026-03-13 10:03:49.252694549 +0000 UTC m=+3.275224692,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: I0313 10:04:15.031267 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:15 crc kubenswrapper[4632]: I0313 10:04:15.032496 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:15 crc kubenswrapper[4632]: I0313 10:04:15.032650 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:15 crc kubenswrapper[4632]: I0313 10:04:15.032761 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:15 crc kubenswrapper[4632]: I0313 10:04:15.033547 4632 scope.go:117] "RemoveContainer" containerID="d83208b114211c941c04a286b532683c256ad92512d1e3fac27e249095b31d4a" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.033845 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.036915 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e7982b87027 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.254311975 +0000 UTC m=+3.276842108,LastTimestamp:2026-03-13 10:03:49.254311975 +0000 UTC m=+3.276842108,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.041845 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e799377a0ed openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.535277293 +0000 UTC m=+3.557807426,LastTimestamp:2026-03-13 10:03:49.535277293 +0000 UTC m=+3.557807426,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.046857 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e7993fba23d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.543928381 +0000 UTC m=+3.566458514,LastTimestamp:2026-03-13 10:03:49.543928381 +0000 UTC m=+3.566458514,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.051402 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e79940a3838 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.54488428 +0000 UTC m=+3.567414413,LastTimestamp:2026-03-13 10:03:49.54488428 +0000 UTC m=+3.567414413,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.055708 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e799e3100d1 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.715198161 +0000 UTC m=+3.737728294,LastTimestamp:2026-03-13 10:03:49.715198161 +0000 UTC m=+3.737728294,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.059838 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e799f3a6138 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.73258988 +0000 UTC m=+3.755120013,LastTimestamp:2026-03-13 10:03:49.73258988 +0000 UTC m=+3.755120013,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.065064 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e799f513f39 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.734088505 +0000 UTC m=+3.756618668,LastTimestamp:2026-03-13 10:03:49.734088505 +0000 UTC m=+3.756618668,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.069466 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e79ab0f25ab openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.931083179 +0000 UTC m=+3.953613312,LastTimestamp:2026-03-13 10:03:49.931083179 +0000 UTC m=+3.953613312,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.074382 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e79abd022fb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.943730939 +0000 UTC m=+3.966261072,LastTimestamp:2026-03-13 10:03:49.943730939 +0000 UTC m=+3.966261072,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.079687 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5e79b36b9370 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.07135832 +0000 UTC m=+4.093888453,LastTimestamp:2026-03-13 10:03:50.07135832 +0000 UTC m=+4.093888453,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.084416 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5e79b3da235d openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.078604125 +0000 UTC m=+4.101134258,LastTimestamp:2026-03-13 10:03:50.078604125 +0000 UTC m=+4.101134258,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.088705 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e79b46640e9 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.087786729 +0000 UTC m=+4.110316862,LastTimestamp:2026-03-13 10:03:50.087786729 +0000 UTC m=+4.110316862,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.093383 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e79b4aed344 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.092542788 +0000 UTC m=+4.115072921,LastTimestamp:2026-03-13 10:03:50.092542788 +0000 UTC m=+4.115072921,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.097126 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5e79c251aaf9 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.321318649 +0000 UTC m=+4.343848782,LastTimestamp:2026-03-13 10:03:50.321318649 +0000 UTC m=+4.343848782,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.101477 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5e79c26e2c82 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.323186818 +0000 UTC m=+4.345716951,LastTimestamp:2026-03-13 10:03:50.323186818 +0000 UTC m=+4.345716951,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.106033 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e79c29b6968 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.326151528 +0000 UTC m=+4.348681661,LastTimestamp:2026-03-13 10:03:50.326151528 +0000 UTC m=+4.348681661,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.110084 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e79c29b6e86 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.326152838 +0000 UTC m=+4.348682971,LastTimestamp:2026-03-13 10:03:50.326152838 +0000 UTC m=+4.348682971,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.113538 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e79c41abdb4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.351273396 +0000 UTC m=+4.373803529,LastTimestamp:2026-03-13 10:03:50.351273396 +0000 UTC m=+4.373803529,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.117308 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e79c431eb52 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.352792402 +0000 UTC m=+4.375322535,LastTimestamp:2026-03-13 10:03:50.352792402 +0000 UTC m=+4.375322535,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.121198 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5e79c45b65de openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.35551075 +0000 UTC m=+4.378040883,LastTimestamp:2026-03-13 10:03:50.35551075 +0000 UTC m=+4.378040883,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.125584 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5e79c47fbf64 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.357892964 +0000 UTC m=+4.380423107,LastTimestamp:2026-03-13 10:03:50.357892964 +0000 UTC m=+4.380423107,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.131806 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e79c488c1e7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.358483431 +0000 UTC m=+4.381013564,LastTimestamp:2026-03-13 10:03:50.358483431 +0000 UTC m=+4.381013564,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.135763 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5e79c4bfb9e2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.362085858 +0000 UTC m=+4.384615991,LastTimestamp:2026-03-13 10:03:50.362085858 +0000 UTC m=+4.384615991,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.139912 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e79d27605b7 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.592136631 +0000 UTC m=+4.614666764,LastTimestamp:2026-03-13 10:03:50.592136631 +0000 UTC m=+4.614666764,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.142017 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5e79d2958489 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.594200713 +0000 UTC m=+4.616730836,LastTimestamp:2026-03-13 10:03:50.594200713 +0000 UTC m=+4.616730836,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.144054 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e79d35fd0af openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.607458479 +0000 UTC m=+4.629988612,LastTimestamp:2026-03-13 10:03:50.607458479 +0000 UTC m=+4.629988612,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.147987 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5e79d371df95 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.608641941 +0000 UTC m=+4.631172074,LastTimestamp:2026-03-13 10:03:50.608641941 +0000 UTC m=+4.631172074,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.151808 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e79d3853e76 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.609911414 +0000 UTC m=+4.632441547,LastTimestamp:2026-03-13 10:03:50.609911414 +0000 UTC m=+4.632441547,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.155834 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5e79d3867521 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.609990945 +0000 UTC m=+4.632521078,LastTimestamp:2026-03-13 10:03:50.609990945 +0000 UTC m=+4.632521078,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.160611 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e79dec011c2 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.79831597 +0000 UTC m=+4.820846103,LastTimestamp:2026-03-13 10:03:50.79831597 +0000 UTC m=+4.820846103,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.164588 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e79e0ae4f97 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.830706583 +0000 UTC m=+4.853236716,LastTimestamp:2026-03-13 10:03:50.830706583 +0000 UTC m=+4.853236716,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.170489 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e79e0ce3826 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.832797734 +0000 UTC m=+4.855327857,LastTimestamp:2026-03-13 10:03:50.832797734 +0000 UTC m=+4.855327857,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.175593 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5e79e3ca15ef openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.882858479 +0000 UTC m=+4.905388612,LastTimestamp:2026-03-13 10:03:50.882858479 +0000 UTC m=+4.905388612,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.179901 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5e79e5721b18 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:50.910647064 +0000 UTC m=+4.933177197,LastTimestamp:2026-03-13 10:03:50.910647064 +0000 UTC m=+4.933177197,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.185611 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e79f3d5d759 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:51.152064345 +0000 UTC m=+5.174594478,LastTimestamp:2026-03-13 10:03:51.152064345 +0000 UTC m=+5.174594478,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.191214 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e79fce736c4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:51.304197828 +0000 UTC m=+5.326727971,LastTimestamp:2026-03-13 10:03:51.304197828 +0000 UTC m=+5.326727971,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.196056 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e79fdc3ac6a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:51.318645866 +0000 UTC m=+5.341176009,LastTimestamp:2026-03-13 10:03:51.318645866 +0000 UTC m=+5.341176009,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.200118 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e79fe0a6dba openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:51.323282874 +0000 UTC m=+5.345813027,LastTimestamp:2026-03-13 10:03:51.323282874 +0000 UTC m=+5.345813027,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.205333 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a0663e194 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:51.463362964 +0000 UTC m=+5.485893097,LastTimestamp:2026-03-13 10:03:51.463362964 +0000 UTC m=+5.485893097,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.210632 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a07d9a447 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:51.487857735 +0000 UTC m=+5.510387868,LastTimestamp:2026-03-13 10:03:51.487857735 +0000 UTC m=+5.510387868,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.215342 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e7a0a4e98df openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:51.529076959 +0000 UTC m=+5.551607092,LastTimestamp:2026-03-13 10:03:51.529076959 +0000 UTC m=+5.551607092,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.220731 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e7a0b1d59ed openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:51.542626797 +0000 UTC m=+5.565156930,LastTimestamp:2026-03-13 10:03:51.542626797 +0000 UTC m=+5.565156930,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.227054 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a319410b5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:52.187941045 +0000 UTC m=+6.210471178,LastTimestamp:2026-03-13 10:03:52.187941045 +0000 UTC m=+6.210471178,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.233036 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a4005fd86 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:52.430288262 +0000 UTC m=+6.452818395,LastTimestamp:2026-03-13 10:03:52.430288262 +0000 UTC m=+6.452818395,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.238361 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a40ac380b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:52.441182219 +0000 UTC m=+6.463712352,LastTimestamp:2026-03-13 10:03:52.441182219 +0000 UTC m=+6.463712352,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.242666 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a40c5253f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:52.442815807 +0000 UTC m=+6.465345940,LastTimestamp:2026-03-13 10:03:52.442815807 +0000 UTC m=+6.465345940,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.247066 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a4c0ec186 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:52.632189318 +0000 UTC m=+6.654719451,LastTimestamp:2026-03-13 10:03:52.632189318 +0000 UTC m=+6.654719451,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.251305 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a4cd5725b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:52.645210715 +0000 UTC m=+6.667740838,LastTimestamp:2026-03-13 10:03:52.645210715 +0000 UTC m=+6.667740838,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.256470 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a4ce5b0db openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:52.646275291 +0000 UTC m=+6.668805424,LastTimestamp:2026-03-13 10:03:52.646275291 +0000 UTC m=+6.668805424,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.261217 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a57999173 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:52.825835891 +0000 UTC m=+6.848366014,LastTimestamp:2026-03-13 10:03:52.825835891 +0000 UTC m=+6.848366014,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.265767 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a584d82de openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:52.837628638 +0000 UTC m=+6.860158791,LastTimestamp:2026-03-13 10:03:52.837628638 +0000 UTC m=+6.860158791,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.271435 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a587ba633 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:52.840652339 +0000 UTC m=+6.863182482,LastTimestamp:2026-03-13 10:03:52.840652339 +0000 UTC m=+6.863182482,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.277291 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a6325ad2d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:53.019567405 +0000 UTC m=+7.042097538,LastTimestamp:2026-03-13 10:03:53.019567405 +0000 UTC m=+7.042097538,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.284841 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a679dc730 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:53.094547248 +0000 UTC m=+7.117077381,LastTimestamp:2026-03-13 10:03:53.094547248 +0000 UTC m=+7.117077381,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.291135 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a67b1f94f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:53.095870799 +0000 UTC m=+7.118400922,LastTimestamp:2026-03-13 10:03:53.095870799 +0000 UTC m=+7.118400922,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.299151 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a73018ae6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:53.28563479 +0000 UTC m=+7.308164923,LastTimestamp:2026-03-13 10:03:53.28563479 +0000 UTC m=+7.308164923,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.302682 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5e7a73d77cd2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:53.29965589 +0000 UTC m=+7.322186023,LastTimestamp:2026-03-13 10:03:53.29965589 +0000 UTC m=+7.322186023,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.312243 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 10:04:15 crc kubenswrapper[4632]: &Event{ObjectMeta:{kube-controller-manager-crc.189c5e7b958c683c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 10:04:15 crc kubenswrapper[4632]: body: Mar 13 10:04:15 crc kubenswrapper[4632]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:58.16012806 +0000 UTC m=+12.182658183,LastTimestamp:2026-03-13 10:03:58.16012806 +0000 UTC m=+12.182658183,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 10:04:15 crc kubenswrapper[4632]: > Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.335919 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e7b958da514 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:58.160209172 +0000 UTC m=+12.182739325,LastTimestamp:2026-03-13 10:03:58.160209172 +0000 UTC m=+12.182739325,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.340302 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 13 10:04:15 crc kubenswrapper[4632]: &Event{ObjectMeta:{kube-apiserver-crc.189c5e7c94589d41 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:6443/livez": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 10:04:15 crc kubenswrapper[4632]: body: Mar 13 10:04:15 crc kubenswrapper[4632]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:04:02.434923841 +0000 UTC m=+16.457453974,LastTimestamp:2026-03-13 10:04:02.434923841 +0000 UTC m=+16.457453974,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 10:04:15 crc kubenswrapper[4632]: > Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.344022 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e7c945af72f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:6443/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:04:02.435077935 +0000 UTC m=+16.457608068,LastTimestamp:2026-03-13 10:04:02.435077935 +0000 UTC m=+16.457608068,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.348584 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c5e79fe0a6dba\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e79fe0a6dba openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:51.323282874 +0000 UTC m=+5.345813027,LastTimestamp:2026-03-13 10:04:03.231962788 +0000 UTC m=+17.254492931,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.351883 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c5e7a0a4e98df\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e7a0a4e98df openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:51.529076959 +0000 UTC m=+5.551607092,LastTimestamp:2026-03-13 10:04:03.56217343 +0000 UTC m=+17.584703563,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.355172 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c5e7a0b1d59ed\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e7a0b1d59ed openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:51.542626797 +0000 UTC m=+5.565156930,LastTimestamp:2026-03-13 10:04:03.573111098 +0000 UTC m=+17.595641231,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.358802 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 13 10:04:15 crc kubenswrapper[4632]: &Event{ObjectMeta:{kube-apiserver-crc.189c5e7d24e13615 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 13 10:04:15 crc kubenswrapper[4632]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 10:04:15 crc kubenswrapper[4632]: Mar 13 10:04:15 crc kubenswrapper[4632]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:04:04.859794965 +0000 UTC m=+18.882325098,LastTimestamp:2026-03-13 10:04:04.859794965 +0000 UTC m=+18.882325098,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 10:04:15 crc kubenswrapper[4632]: > Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.362218 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5e7d24e3d879 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:04:04.859967609 +0000 UTC m=+18.882497742,LastTimestamp:2026-03-13 10:04:04.859967609 +0000 UTC m=+18.882497742,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.366289 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5e7b958c683c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 10:04:15 crc kubenswrapper[4632]: &Event{ObjectMeta:{kube-controller-manager-crc.189c5e7b958c683c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 10:04:15 crc kubenswrapper[4632]: body: Mar 13 10:04:15 crc kubenswrapper[4632]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:58.16012806 +0000 UTC m=+12.182658183,LastTimestamp:2026-03-13 10:04:08.164353209 +0000 UTC m=+22.186883342,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 10:04:15 crc kubenswrapper[4632]: > Mar 13 10:04:15 crc kubenswrapper[4632]: E0313 10:04:15.369591 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5e7b958da514\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e7b958da514 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:58.160209172 +0000 UTC m=+12.182739325,LastTimestamp:2026-03-13 10:04:08.16441919 +0000 UTC m=+22.186949323,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:15 crc kubenswrapper[4632]: I0313 10:04:15.979280 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:16 crc kubenswrapper[4632]: W0313 10:04:16.742063 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 13 10:04:16 crc kubenswrapper[4632]: E0313 10:04:16.742162 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 13 10:04:16 crc kubenswrapper[4632]: I0313 10:04:16.979394 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:17 crc kubenswrapper[4632]: E0313 10:04:17.612904 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 10:04:17 crc kubenswrapper[4632]: I0313 10:04:17.876687 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:17 crc kubenswrapper[4632]: I0313 10:04:17.879079 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:17 crc kubenswrapper[4632]: I0313 10:04:17.879149 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:17 crc kubenswrapper[4632]: I0313 10:04:17.879166 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:17 crc kubenswrapper[4632]: I0313 10:04:17.879195 4632 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 10:04:17 crc kubenswrapper[4632]: E0313 10:04:17.901487 4632 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 10:04:17 crc kubenswrapper[4632]: I0313 10:04:17.980253 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:18 crc kubenswrapper[4632]: E0313 10:04:18.131735 4632 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 10:04:18 crc kubenswrapper[4632]: I0313 10:04:18.161319 4632 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:04:18 crc kubenswrapper[4632]: I0313 10:04:18.161406 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:04:18 crc kubenswrapper[4632]: I0313 10:04:18.161484 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:04:18 crc kubenswrapper[4632]: I0313 10:04:18.161664 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:18 crc kubenswrapper[4632]: I0313 10:04:18.162968 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:18 crc kubenswrapper[4632]: I0313 10:04:18.163020 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:18 crc kubenswrapper[4632]: I0313 10:04:18.163046 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:18 crc kubenswrapper[4632]: I0313 10:04:18.163771 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 13 10:04:18 crc kubenswrapper[4632]: I0313 10:04:18.163988 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac" gracePeriod=30 Mar 13 10:04:18 crc kubenswrapper[4632]: E0313 10:04:18.184838 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5e7b958c683c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 10:04:18 crc kubenswrapper[4632]: &Event{ObjectMeta:{kube-controller-manager-crc.189c5e7b958c683c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 10:04:18 crc kubenswrapper[4632]: body: Mar 13 10:04:18 crc kubenswrapper[4632]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:58.16012806 +0000 UTC m=+12.182658183,LastTimestamp:2026-03-13 10:04:18.161384938 +0000 UTC m=+32.183915071,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 10:04:18 crc kubenswrapper[4632]: > Mar 13 10:04:18 crc kubenswrapper[4632]: E0313 10:04:18.191203 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5e7b958da514\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e7b958da514 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:58.160209172 +0000 UTC m=+12.182739325,LastTimestamp:2026-03-13 10:04:18.16143983 +0000 UTC m=+32.183969963,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:18 crc kubenswrapper[4632]: E0313 10:04:18.203854 4632 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e803ddec832 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:04:18.16396805 +0000 UTC m=+32.186498193,LastTimestamp:2026-03-13 10:04:18.16396805 +0000 UTC m=+32.186498193,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:18 crc kubenswrapper[4632]: E0313 10:04:18.292187 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5e7982b87027\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e7982b87027 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.254311975 +0000 UTC m=+3.276842108,LastTimestamp:2026-03-13 10:04:18.286345496 +0000 UTC m=+32.308875629,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:18 crc kubenswrapper[4632]: I0313 10:04:18.301349 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 10:04:18 crc kubenswrapper[4632]: I0313 10:04:18.303052 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac"} Mar 13 10:04:18 crc kubenswrapper[4632]: I0313 10:04:18.303678 4632 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac" exitCode=255 Mar 13 10:04:18 crc kubenswrapper[4632]: E0313 10:04:18.463702 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5e799377a0ed\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e799377a0ed openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.535277293 +0000 UTC m=+3.557807426,LastTimestamp:2026-03-13 10:04:18.45786721 +0000 UTC m=+32.480397363,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:18 crc kubenswrapper[4632]: E0313 10:04:18.476675 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5e7993fba23d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e7993fba23d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:49.543928381 +0000 UTC m=+3.566458514,LastTimestamp:2026-03-13 10:04:18.471551981 +0000 UTC m=+32.494082114,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:18 crc kubenswrapper[4632]: I0313 10:04:18.979715 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:19 crc kubenswrapper[4632]: I0313 10:04:19.309702 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 10:04:19 crc kubenswrapper[4632]: I0313 10:04:19.310220 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"951a71f1358e164fefc6fbe9404cf1d0e387d58b0bc1d060eeca64a85e0fc08e"} Mar 13 10:04:19 crc kubenswrapper[4632]: I0313 10:04:19.310368 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:19 crc kubenswrapper[4632]: I0313 10:04:19.311979 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:19 crc kubenswrapper[4632]: I0313 10:04:19.312025 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:19 crc kubenswrapper[4632]: I0313 10:04:19.312044 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:19 crc kubenswrapper[4632]: I0313 10:04:19.979880 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:20 crc kubenswrapper[4632]: W0313 10:04:20.251851 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 13 10:04:20 crc kubenswrapper[4632]: E0313 10:04:20.252415 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 13 10:04:20 crc kubenswrapper[4632]: I0313 10:04:20.312791 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:20 crc kubenswrapper[4632]: I0313 10:04:20.314087 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:20 crc kubenswrapper[4632]: I0313 10:04:20.314184 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:20 crc kubenswrapper[4632]: I0313 10:04:20.314203 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:20 crc kubenswrapper[4632]: I0313 10:04:20.981234 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:21 crc kubenswrapper[4632]: I0313 10:04:21.980137 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:22 crc kubenswrapper[4632]: I0313 10:04:22.979441 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:23 crc kubenswrapper[4632]: I0313 10:04:23.980020 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:24 crc kubenswrapper[4632]: I0313 10:04:24.525029 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:04:24 crc kubenswrapper[4632]: I0313 10:04:24.525294 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:24 crc kubenswrapper[4632]: I0313 10:04:24.526795 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:24 crc kubenswrapper[4632]: I0313 10:04:24.526840 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:24 crc kubenswrapper[4632]: I0313 10:04:24.526850 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:24 crc kubenswrapper[4632]: E0313 10:04:24.618503 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 10:04:24 crc kubenswrapper[4632]: I0313 10:04:24.902287 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:24 crc kubenswrapper[4632]: I0313 10:04:24.904126 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:24 crc kubenswrapper[4632]: I0313 10:04:24.904181 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:24 crc kubenswrapper[4632]: I0313 10:04:24.904198 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:24 crc kubenswrapper[4632]: I0313 10:04:24.904229 4632 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 10:04:24 crc kubenswrapper[4632]: E0313 10:04:24.909309 4632 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 10:04:24 crc kubenswrapper[4632]: I0313 10:04:24.981133 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:25 crc kubenswrapper[4632]: I0313 10:04:25.160702 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:04:25 crc kubenswrapper[4632]: I0313 10:04:25.333212 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:25 crc kubenswrapper[4632]: I0313 10:04:25.334263 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:25 crc kubenswrapper[4632]: I0313 10:04:25.334321 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:25 crc kubenswrapper[4632]: I0313 10:04:25.334333 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:25 crc kubenswrapper[4632]: I0313 10:04:25.979346 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:26 crc kubenswrapper[4632]: I0313 10:04:26.043697 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:26 crc kubenswrapper[4632]: I0313 10:04:26.045990 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:26 crc kubenswrapper[4632]: I0313 10:04:26.046068 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:26 crc kubenswrapper[4632]: I0313 10:04:26.046081 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:26 crc kubenswrapper[4632]: I0313 10:04:26.046797 4632 scope.go:117] "RemoveContainer" containerID="d83208b114211c941c04a286b532683c256ad92512d1e3fac27e249095b31d4a" Mar 13 10:04:26 crc kubenswrapper[4632]: I0313 10:04:26.338642 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 10:04:26 crc kubenswrapper[4632]: I0313 10:04:26.340415 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6b2d0d7e12e3475e3d8c895d2175381ba90230c85db30b582d19853ce1e4ba54"} Mar 13 10:04:26 crc kubenswrapper[4632]: I0313 10:04:26.340540 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:26 crc kubenswrapper[4632]: I0313 10:04:26.341288 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:26 crc kubenswrapper[4632]: I0313 10:04:26.341312 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:26 crc kubenswrapper[4632]: I0313 10:04:26.341320 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:26 crc kubenswrapper[4632]: I0313 10:04:26.979757 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:27 crc kubenswrapper[4632]: I0313 10:04:27.346033 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 10:04:27 crc kubenswrapper[4632]: I0313 10:04:27.346497 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 10:04:27 crc kubenswrapper[4632]: I0313 10:04:27.348443 4632 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6b2d0d7e12e3475e3d8c895d2175381ba90230c85db30b582d19853ce1e4ba54" exitCode=255 Mar 13 10:04:27 crc kubenswrapper[4632]: I0313 10:04:27.348492 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"6b2d0d7e12e3475e3d8c895d2175381ba90230c85db30b582d19853ce1e4ba54"} Mar 13 10:04:27 crc kubenswrapper[4632]: I0313 10:04:27.348546 4632 scope.go:117] "RemoveContainer" containerID="d83208b114211c941c04a286b532683c256ad92512d1e3fac27e249095b31d4a" Mar 13 10:04:27 crc kubenswrapper[4632]: I0313 10:04:27.348670 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:27 crc kubenswrapper[4632]: I0313 10:04:27.349900 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:27 crc kubenswrapper[4632]: I0313 10:04:27.349961 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:27 crc kubenswrapper[4632]: I0313 10:04:27.350000 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:27 crc kubenswrapper[4632]: I0313 10:04:27.350786 4632 scope.go:117] "RemoveContainer" containerID="6b2d0d7e12e3475e3d8c895d2175381ba90230c85db30b582d19853ce1e4ba54" Mar 13 10:04:27 crc kubenswrapper[4632]: E0313 10:04:27.351028 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 10:04:27 crc kubenswrapper[4632]: I0313 10:04:27.979602 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:28 crc kubenswrapper[4632]: E0313 10:04:28.132173 4632 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 10:04:28 crc kubenswrapper[4632]: I0313 10:04:28.161046 4632 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:04:28 crc kubenswrapper[4632]: I0313 10:04:28.161162 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:04:28 crc kubenswrapper[4632]: E0313 10:04:28.166654 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5e7b958c683c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 10:04:28 crc kubenswrapper[4632]: &Event{ObjectMeta:{kube-controller-manager-crc.189c5e7b958c683c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 10:04:28 crc kubenswrapper[4632]: body: Mar 13 10:04:28 crc kubenswrapper[4632]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:58.16012806 +0000 UTC m=+12.182658183,LastTimestamp:2026-03-13 10:04:28.161125237 +0000 UTC m=+42.183655380,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 10:04:28 crc kubenswrapper[4632]: > Mar 13 10:04:28 crc kubenswrapper[4632]: E0313 10:04:28.171758 4632 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5e7b958da514\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5e7b958da514 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:03:58.160209172 +0000 UTC m=+12.182739325,LastTimestamp:2026-03-13 10:04:28.161198848 +0000 UTC m=+42.183729001,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:04:28 crc kubenswrapper[4632]: I0313 10:04:28.359516 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 10:04:28 crc kubenswrapper[4632]: I0313 10:04:28.981173 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:29 crc kubenswrapper[4632]: I0313 10:04:29.980610 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:30 crc kubenswrapper[4632]: I0313 10:04:30.981214 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:31 crc kubenswrapper[4632]: E0313 10:04:31.624726 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 10:04:31 crc kubenswrapper[4632]: I0313 10:04:31.910350 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:31 crc kubenswrapper[4632]: I0313 10:04:31.912163 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:31 crc kubenswrapper[4632]: I0313 10:04:31.912230 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:31 crc kubenswrapper[4632]: I0313 10:04:31.912255 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:31 crc kubenswrapper[4632]: I0313 10:04:31.912309 4632 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 10:04:31 crc kubenswrapper[4632]: E0313 10:04:31.917879 4632 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 10:04:31 crc kubenswrapper[4632]: I0313 10:04:31.983223 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:32 crc kubenswrapper[4632]: W0313 10:04:32.145577 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:32 crc kubenswrapper[4632]: E0313 10:04:32.145653 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 13 10:04:32 crc kubenswrapper[4632]: I0313 10:04:32.980233 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:34 crc kubenswrapper[4632]: I0313 10:04:34.010654 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:34 crc kubenswrapper[4632]: I0313 10:04:34.979471 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:35 crc kubenswrapper[4632]: I0313 10:04:35.032131 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:04:35 crc kubenswrapper[4632]: I0313 10:04:35.032372 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:35 crc kubenswrapper[4632]: I0313 10:04:35.034166 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:35 crc kubenswrapper[4632]: I0313 10:04:35.034207 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:35 crc kubenswrapper[4632]: I0313 10:04:35.034219 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:35 crc kubenswrapper[4632]: I0313 10:04:35.034822 4632 scope.go:117] "RemoveContainer" containerID="6b2d0d7e12e3475e3d8c895d2175381ba90230c85db30b582d19853ce1e4ba54" Mar 13 10:04:35 crc kubenswrapper[4632]: E0313 10:04:35.035079 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 10:04:35 crc kubenswrapper[4632]: W0313 10:04:35.415562 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 13 10:04:35 crc kubenswrapper[4632]: E0313 10:04:35.415641 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 13 10:04:35 crc kubenswrapper[4632]: I0313 10:04:35.980714 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.089144 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.089379 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.091344 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.091715 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.091742 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.092681 4632 scope.go:117] "RemoveContainer" containerID="6b2d0d7e12e3475e3d8c895d2175381ba90230c85db30b582d19853ce1e4ba54" Mar 13 10:04:36 crc kubenswrapper[4632]: E0313 10:04:36.092878 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.142356 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.142540 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.144113 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.144151 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.144162 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.146720 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.387355 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.388541 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.388587 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.388603 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:36 crc kubenswrapper[4632]: I0313 10:04:36.979587 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:37 crc kubenswrapper[4632]: W0313 10:04:37.124968 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 13 10:04:37 crc kubenswrapper[4632]: E0313 10:04:37.125054 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 13 10:04:37 crc kubenswrapper[4632]: I0313 10:04:37.995242 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:38 crc kubenswrapper[4632]: E0313 10:04:38.133367 4632 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 10:04:38 crc kubenswrapper[4632]: E0313 10:04:38.755145 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 10:04:38 crc kubenswrapper[4632]: I0313 10:04:38.918299 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:38 crc kubenswrapper[4632]: I0313 10:04:38.920076 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:38 crc kubenswrapper[4632]: I0313 10:04:38.920248 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:38 crc kubenswrapper[4632]: I0313 10:04:38.920346 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:38 crc kubenswrapper[4632]: I0313 10:04:38.920446 4632 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 10:04:38 crc kubenswrapper[4632]: E0313 10:04:38.972205 4632 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 10:04:38 crc kubenswrapper[4632]: I0313 10:04:38.980007 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:39 crc kubenswrapper[4632]: I0313 10:04:39.987409 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:40 crc kubenswrapper[4632]: I0313 10:04:40.980743 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:41 crc kubenswrapper[4632]: I0313 10:04:41.358757 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 10:04:41 crc kubenswrapper[4632]: I0313 10:04:41.358906 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:41 crc kubenswrapper[4632]: I0313 10:04:41.369804 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:41 crc kubenswrapper[4632]: I0313 10:04:41.369885 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:41 crc kubenswrapper[4632]: I0313 10:04:41.369902 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:41 crc kubenswrapper[4632]: I0313 10:04:41.982792 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:43 crc kubenswrapper[4632]: I0313 10:04:43.039739 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:43 crc kubenswrapper[4632]: W0313 10:04:43.180366 4632 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 13 10:04:43 crc kubenswrapper[4632]: E0313 10:04:43.180431 4632 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 13 10:04:43 crc kubenswrapper[4632]: I0313 10:04:43.980585 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:44 crc kubenswrapper[4632]: I0313 10:04:44.991597 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:45 crc kubenswrapper[4632]: E0313 10:04:45.761887 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 10:04:45 crc kubenswrapper[4632]: I0313 10:04:45.972563 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:45 crc kubenswrapper[4632]: I0313 10:04:45.974751 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:45 crc kubenswrapper[4632]: I0313 10:04:45.974831 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:45 crc kubenswrapper[4632]: I0313 10:04:45.974844 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:45 crc kubenswrapper[4632]: I0313 10:04:45.974881 4632 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 10:04:45 crc kubenswrapper[4632]: I0313 10:04:45.978997 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:45 crc kubenswrapper[4632]: E0313 10:04:45.979349 4632 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 10:04:46 crc kubenswrapper[4632]: I0313 10:04:46.979405 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:47 crc kubenswrapper[4632]: I0313 10:04:47.979414 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:48 crc kubenswrapper[4632]: E0313 10:04:48.134489 4632 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 10:04:48 crc kubenswrapper[4632]: I0313 10:04:48.980333 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:49 crc kubenswrapper[4632]: I0313 10:04:49.982397 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:50 crc kubenswrapper[4632]: I0313 10:04:50.044206 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:50 crc kubenswrapper[4632]: I0313 10:04:50.046086 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:50 crc kubenswrapper[4632]: I0313 10:04:50.046133 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:50 crc kubenswrapper[4632]: I0313 10:04:50.046147 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:50 crc kubenswrapper[4632]: I0313 10:04:50.046808 4632 scope.go:117] "RemoveContainer" containerID="6b2d0d7e12e3475e3d8c895d2175381ba90230c85db30b582d19853ce1e4ba54" Mar 13 10:04:50 crc kubenswrapper[4632]: I0313 10:04:50.979917 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:51 crc kubenswrapper[4632]: I0313 10:04:51.435253 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 10:04:51 crc kubenswrapper[4632]: I0313 10:04:51.438321 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae"} Mar 13 10:04:51 crc kubenswrapper[4632]: I0313 10:04:51.438815 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:51 crc kubenswrapper[4632]: I0313 10:04:51.440211 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:51 crc kubenswrapper[4632]: I0313 10:04:51.440436 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:51 crc kubenswrapper[4632]: I0313 10:04:51.440571 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:51 crc kubenswrapper[4632]: I0313 10:04:51.979668 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.443982 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.445584 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.448534 4632 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae" exitCode=255 Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.448601 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae"} Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.448668 4632 scope.go:117] "RemoveContainer" containerID="6b2d0d7e12e3475e3d8c895d2175381ba90230c85db30b582d19853ce1e4ba54" Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.448966 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.450007 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.450041 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.450051 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.450676 4632 scope.go:117] "RemoveContainer" containerID="9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae" Mar 13 10:04:52 crc kubenswrapper[4632]: E0313 10:04:52.450999 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 10:04:52 crc kubenswrapper[4632]: E0313 10:04:52.768114 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.979441 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.980112 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.980665 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.980727 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.980747 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:52 crc kubenswrapper[4632]: I0313 10:04:52.980819 4632 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 10:04:52 crc kubenswrapper[4632]: E0313 10:04:52.988231 4632 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 10:04:53 crc kubenswrapper[4632]: I0313 10:04:53.453726 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 10:04:53 crc kubenswrapper[4632]: I0313 10:04:53.979330 4632 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 10:04:54 crc kubenswrapper[4632]: I0313 10:04:54.124029 4632 csr.go:261] certificate signing request csr-f8d22 is approved, waiting to be issued Mar 13 10:04:54 crc kubenswrapper[4632]: I0313 10:04:54.134050 4632 csr.go:257] certificate signing request csr-f8d22 is issued Mar 13 10:04:54 crc kubenswrapper[4632]: I0313 10:04:54.163778 4632 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 13 10:04:54 crc kubenswrapper[4632]: I0313 10:04:54.643539 4632 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 13 10:04:55 crc kubenswrapper[4632]: I0313 10:04:55.031921 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:04:55 crc kubenswrapper[4632]: I0313 10:04:55.032473 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:55 crc kubenswrapper[4632]: I0313 10:04:55.034265 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:55 crc kubenswrapper[4632]: I0313 10:04:55.034331 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:55 crc kubenswrapper[4632]: I0313 10:04:55.034345 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:55 crc kubenswrapper[4632]: I0313 10:04:55.035185 4632 scope.go:117] "RemoveContainer" containerID="9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae" Mar 13 10:04:55 crc kubenswrapper[4632]: E0313 10:04:55.035396 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 10:04:55 crc kubenswrapper[4632]: I0313 10:04:55.135216 4632 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-13 20:31:44.592872983 +0000 UTC Mar 13 10:04:55 crc kubenswrapper[4632]: I0313 10:04:55.135331 4632 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6610h26m49.457544381s for next certificate rotation Mar 13 10:04:56 crc kubenswrapper[4632]: I0313 10:04:56.088911 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:04:56 crc kubenswrapper[4632]: I0313 10:04:56.089159 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:56 crc kubenswrapper[4632]: I0313 10:04:56.090381 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:56 crc kubenswrapper[4632]: I0313 10:04:56.090422 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:56 crc kubenswrapper[4632]: I0313 10:04:56.090434 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:56 crc kubenswrapper[4632]: I0313 10:04:56.090984 4632 scope.go:117] "RemoveContainer" containerID="9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae" Mar 13 10:04:56 crc kubenswrapper[4632]: E0313 10:04:56.091171 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 10:04:58 crc kubenswrapper[4632]: E0313 10:04:58.135647 4632 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 10:04:59 crc kubenswrapper[4632]: I0313 10:04:59.988602 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:04:59 crc kubenswrapper[4632]: I0313 10:04:59.990106 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:04:59 crc kubenswrapper[4632]: I0313 10:04:59.990143 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:04:59 crc kubenswrapper[4632]: I0313 10:04:59.990153 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:04:59 crc kubenswrapper[4632]: I0313 10:04:59.990241 4632 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 10:04:59 crc kubenswrapper[4632]: I0313 10:04:59.998230 4632 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 13 10:04:59 crc kubenswrapper[4632]: I0313 10:04:59.998661 4632 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 13 10:04:59 crc kubenswrapper[4632]: E0313 10:04:59.998685 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.002690 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.002745 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.002758 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.002775 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.002787 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:00Z","lastTransitionTime":"2026-03-13T10:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.018527 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.027545 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.027614 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.027629 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.027652 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.027668 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:00Z","lastTransitionTime":"2026-03-13T10:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.050859 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.059971 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.060015 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.060026 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.060044 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.060055 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:00Z","lastTransitionTime":"2026-03-13T10:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.071292 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.077505 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.077558 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.077574 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.077600 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:00 crc kubenswrapper[4632]: I0313 10:05:00.077616 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:00Z","lastTransitionTime":"2026-03-13T10:05:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.087376 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.087528 4632 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.087562 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.188129 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.288966 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.390169 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.490995 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.592238 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.693388 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.794090 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.895483 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:00 crc kubenswrapper[4632]: E0313 10:05:00.996930 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:01 crc kubenswrapper[4632]: E0313 10:05:01.097443 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:01 crc kubenswrapper[4632]: E0313 10:05:01.198262 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:01 crc kubenswrapper[4632]: E0313 10:05:01.301600 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:01 crc kubenswrapper[4632]: E0313 10:05:01.401878 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:01 crc kubenswrapper[4632]: E0313 10:05:01.502818 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:01 crc kubenswrapper[4632]: E0313 10:05:01.603282 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:01 crc kubenswrapper[4632]: E0313 10:05:01.703553 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:01 crc kubenswrapper[4632]: E0313 10:05:01.804332 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:01 crc kubenswrapper[4632]: E0313 10:05:01.905392 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:02 crc kubenswrapper[4632]: E0313 10:05:02.006631 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:02 crc kubenswrapper[4632]: E0313 10:05:02.107756 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:02 crc kubenswrapper[4632]: E0313 10:05:02.208285 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:02 crc kubenswrapper[4632]: E0313 10:05:02.308869 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:02 crc kubenswrapper[4632]: E0313 10:05:02.409341 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:02 crc kubenswrapper[4632]: E0313 10:05:02.510048 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:02 crc kubenswrapper[4632]: E0313 10:05:02.610564 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:02 crc kubenswrapper[4632]: E0313 10:05:02.710840 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:02 crc kubenswrapper[4632]: E0313 10:05:02.811655 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:02 crc kubenswrapper[4632]: E0313 10:05:02.912774 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:03 crc kubenswrapper[4632]: E0313 10:05:03.013565 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:03 crc kubenswrapper[4632]: E0313 10:05:03.115150 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:03 crc kubenswrapper[4632]: E0313 10:05:03.216223 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:03 crc kubenswrapper[4632]: E0313 10:05:03.317798 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:03 crc kubenswrapper[4632]: E0313 10:05:03.418475 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:03 crc kubenswrapper[4632]: E0313 10:05:03.519075 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:03 crc kubenswrapper[4632]: E0313 10:05:03.619692 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:03 crc kubenswrapper[4632]: E0313 10:05:03.720824 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:03 crc kubenswrapper[4632]: E0313 10:05:03.821795 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:03 crc kubenswrapper[4632]: E0313 10:05:03.922488 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:04 crc kubenswrapper[4632]: E0313 10:05:04.023331 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:04 crc kubenswrapper[4632]: E0313 10:05:04.123885 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:04 crc kubenswrapper[4632]: E0313 10:05:04.224384 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:04 crc kubenswrapper[4632]: E0313 10:05:04.324889 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:04 crc kubenswrapper[4632]: E0313 10:05:04.425715 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:04 crc kubenswrapper[4632]: E0313 10:05:04.526400 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:04 crc kubenswrapper[4632]: E0313 10:05:04.626981 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:04 crc kubenswrapper[4632]: E0313 10:05:04.727561 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:04 crc kubenswrapper[4632]: E0313 10:05:04.827833 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:04 crc kubenswrapper[4632]: E0313 10:05:04.927981 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:05 crc kubenswrapper[4632]: E0313 10:05:05.028686 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:05 crc kubenswrapper[4632]: E0313 10:05:05.129849 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:05 crc kubenswrapper[4632]: E0313 10:05:05.230986 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:05 crc kubenswrapper[4632]: E0313 10:05:05.331524 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:05 crc kubenswrapper[4632]: E0313 10:05:05.431851 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:05 crc kubenswrapper[4632]: E0313 10:05:05.532820 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:05 crc kubenswrapper[4632]: E0313 10:05:05.633058 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:05 crc kubenswrapper[4632]: E0313 10:05:05.734179 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:05 crc kubenswrapper[4632]: E0313 10:05:05.834642 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:05 crc kubenswrapper[4632]: E0313 10:05:05.935520 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:06 crc kubenswrapper[4632]: E0313 10:05:06.036159 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:06 crc kubenswrapper[4632]: E0313 10:05:06.136866 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:06 crc kubenswrapper[4632]: E0313 10:05:06.237094 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:06 crc kubenswrapper[4632]: E0313 10:05:06.337669 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:06 crc kubenswrapper[4632]: E0313 10:05:06.438895 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:06 crc kubenswrapper[4632]: E0313 10:05:06.539670 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:06 crc kubenswrapper[4632]: E0313 10:05:06.640820 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:06 crc kubenswrapper[4632]: E0313 10:05:06.741816 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:06 crc kubenswrapper[4632]: E0313 10:05:06.843121 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:06 crc kubenswrapper[4632]: E0313 10:05:06.943498 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:07 crc kubenswrapper[4632]: E0313 10:05:07.044096 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:07 crc kubenswrapper[4632]: E0313 10:05:07.144888 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:07 crc kubenswrapper[4632]: E0313 10:05:07.245071 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:07 crc kubenswrapper[4632]: E0313 10:05:07.345321 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:07 crc kubenswrapper[4632]: E0313 10:05:07.445844 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:07 crc kubenswrapper[4632]: E0313 10:05:07.546514 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:07 crc kubenswrapper[4632]: E0313 10:05:07.647426 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:07 crc kubenswrapper[4632]: E0313 10:05:07.747622 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:07 crc kubenswrapper[4632]: E0313 10:05:07.848526 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:07 crc kubenswrapper[4632]: E0313 10:05:07.948924 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:08 crc kubenswrapper[4632]: I0313 10:05:08.044071 4632 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 10:05:08 crc kubenswrapper[4632]: I0313 10:05:08.045592 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:08 crc kubenswrapper[4632]: I0313 10:05:08.045650 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:08 crc kubenswrapper[4632]: I0313 10:05:08.045665 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:08 crc kubenswrapper[4632]: E0313 10:05:08.049066 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:08 crc kubenswrapper[4632]: E0313 10:05:08.136067 4632 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 10:05:08 crc kubenswrapper[4632]: E0313 10:05:08.150149 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:08 crc kubenswrapper[4632]: E0313 10:05:08.250505 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:08 crc kubenswrapper[4632]: E0313 10:05:08.351616 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:08 crc kubenswrapper[4632]: E0313 10:05:08.452708 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:08 crc kubenswrapper[4632]: E0313 10:05:08.553055 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:08 crc kubenswrapper[4632]: E0313 10:05:08.653604 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:08 crc kubenswrapper[4632]: E0313 10:05:08.754449 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:08 crc kubenswrapper[4632]: E0313 10:05:08.855192 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:08 crc kubenswrapper[4632]: E0313 10:05:08.956022 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:09 crc kubenswrapper[4632]: E0313 10:05:09.056595 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:09 crc kubenswrapper[4632]: E0313 10:05:09.157484 4632 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.249331 4632 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.259965 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.260026 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.260041 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.260062 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.260076 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:09Z","lastTransitionTime":"2026-03-13T10:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.363394 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.363441 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.363454 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.363472 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.363484 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:09Z","lastTransitionTime":"2026-03-13T10:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.466521 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.466556 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.466565 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.466579 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.466588 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:09Z","lastTransitionTime":"2026-03-13T10:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.570353 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.570427 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.570443 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.570468 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.570483 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:09Z","lastTransitionTime":"2026-03-13T10:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.674091 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.674170 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.674191 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.674219 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.674242 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:09Z","lastTransitionTime":"2026-03-13T10:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.777507 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.777575 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.777584 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.777598 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.777630 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:09Z","lastTransitionTime":"2026-03-13T10:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.880686 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.880786 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.880806 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.880826 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.880836 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:09Z","lastTransitionTime":"2026-03-13T10:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.983410 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.983473 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.983484 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.983498 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:09 crc kubenswrapper[4632]: I0313 10:05:09.983507 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:09Z","lastTransitionTime":"2026-03-13T10:05:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.050352 4632 apiserver.go:52] "Watching apiserver" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.086512 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.086576 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.086589 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.086612 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.086625 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:10Z","lastTransitionTime":"2026-03-13T10:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.087653 4632 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.088052 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2","openshift-multus/multus-additional-cni-plugins-qlc8m","openshift-multus/multus-gqf22","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-n55jt","openshift-multus/network-metrics-daemon-z2vlz","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-operator/iptables-alerter-4ln5h","openshift-ovn-kubernetes/ovnkube-node-qb725","openshift-image-registry/node-ca-zwlc8","openshift-machine-config-operator/machine-config-daemon-zkscb","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.088410 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.088453 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.088491 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.088576 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.088581 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.088527 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.088971 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.089060 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.089095 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.089997 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.090813 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-n55jt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.090842 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.091365 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.091423 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zwlc8" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.091478 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.091876 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.091963 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.092046 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.096915 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.097183 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.097396 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.097918 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.101607 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.102081 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.102146 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.102081 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.102679 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.102689 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.102683 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.102847 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.103126 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.103192 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.103397 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.103421 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.103677 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.104158 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.104357 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.104527 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.104760 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.104880 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.105098 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.105162 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.105170 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.105237 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.105378 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.105391 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.105555 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.105705 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.105804 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.105805 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.105990 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.106038 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.106096 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.106443 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.106643 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.123558 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.138063 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.148818 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.164649 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.169155 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.169191 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.169204 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.169222 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.169233 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:10Z","lastTransitionTime":"2026-03-13T10:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.179213 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.179261 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.179603 4632 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.183206 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.183241 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.183257 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.183272 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.183282 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:10Z","lastTransitionTime":"2026-03-13T10:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.192321 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.195632 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.195663 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.195671 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.195684 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.195694 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:10Z","lastTransitionTime":"2026-03-13T10:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.196765 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.204465 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.204544 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.207667 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.207708 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.207719 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.207736 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.207746 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:10Z","lastTransitionTime":"2026-03-13T10:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.214263 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.217778 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.221124 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.221159 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.221172 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.221188 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.221199 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:10Z","lastTransitionTime":"2026-03-13T10:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.230080 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.230091 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.230722 4632 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.233403 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.233434 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.233447 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.233468 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.233481 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:10Z","lastTransitionTime":"2026-03-13T10:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.239015 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.248648 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.258253 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.264029 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.264263 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.264373 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.264482 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.264576 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.264680 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.264792 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.264892 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.265022 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.265128 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.265219 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.265321 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.265419 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.265519 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.265613 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.265772 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.265896 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.266049 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.266554 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.266987 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.267288 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.267584 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.267692 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.267785 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.267879 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.268369 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.268834 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.269115 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.269308 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.269844 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.270546 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.270675 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.270800 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.270907 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.264680 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.264775 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.264930 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.265203 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.265626 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.265914 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.270917 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.266125 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.266203 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.266306 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.266518 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.266713 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.267117 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.267955 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.267930 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.267973 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.268202 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.268559 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.268600 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.268957 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.269221 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.269385 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.269449 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.269490 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.269513 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.270207 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.270497 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.270503 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.271650 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.271684 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.271706 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.271728 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.271751 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.271776 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.271801 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.271820 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.271837 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.271853 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.271869 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.271890 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.271932 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272038 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272059 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272076 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272093 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272111 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272141 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272217 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272248 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272270 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272293 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272312 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272327 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272352 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272369 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272387 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272409 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272427 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272443 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272460 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272476 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272492 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272509 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272525 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272540 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272556 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272573 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272587 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272601 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272616 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272631 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272822 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272875 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272903 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272923 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272969 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.272994 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273011 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273030 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273049 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273066 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273081 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273104 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273127 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273151 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273175 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273198 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273218 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273239 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273265 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273293 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273321 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273363 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273389 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273406 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273430 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273445 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273461 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273477 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273499 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273530 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273561 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273586 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273604 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273620 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273636 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273651 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273668 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273724 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273741 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273757 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273774 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273791 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273807 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273830 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273853 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273877 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.273901 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274196 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274234 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274261 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274293 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274319 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274344 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274370 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274408 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274435 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274461 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274487 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274543 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274600 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274667 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274699 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274725 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274752 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274778 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274803 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274829 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274853 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274878 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274903 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274927 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.274969 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275017 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275044 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275072 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275100 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275126 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275151 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275176 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275200 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275222 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275251 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275277 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275419 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275453 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275479 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275514 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275545 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275570 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275598 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275623 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275651 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275676 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275699 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275725 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275750 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275777 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275804 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275830 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275856 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275881 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275908 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275962 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.275988 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276013 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276091 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276121 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276150 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276177 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276206 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276233 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276261 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276290 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276315 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276359 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276386 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276413 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276437 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276513 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d77b18a7-7ad9-4bf5-bff5-da45878af7f4-rootfs\") pod \"machine-config-daemon-zkscb\" (UID: \"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\") " pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276545 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnh6w\" (UniqueName: \"kubernetes.io/projected/d77b18a7-7ad9-4bf5-bff5-da45878af7f4-kube-api-access-vnh6w\") pod \"machine-config-daemon-zkscb\" (UID: \"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\") " pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276573 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b054ca08-1d09-4eca-a608-eb5b9323959a-cni-binary-copy\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276596 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b054ca08-1d09-4eca-a608-eb5b9323959a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276627 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276655 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b0c542d5-8c38-4243-8af7-cfc0d8e22773-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-kbtt2\" (UID: \"b0c542d5-8c38-4243-8af7-cfc0d8e22773\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276711 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276739 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-run-netns\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276762 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-openvswitch\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276798 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-ovnkube-config\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276821 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3b40c6b3-0061-4224-82d5-3ccf67998722-ovn-node-metrics-cert\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276845 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276871 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9a50974e-f938-40f7-ace5-2a3b4cb1f3e7-serviceca\") pod \"node-ca-zwlc8\" (UID: \"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\") " pod="openshift-image-registry/node-ca-zwlc8" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276896 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9s5s\" (UniqueName: \"kubernetes.io/projected/b054ca08-1d09-4eca-a608-eb5b9323959a-kube-api-access-l9s5s\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276918 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-log-socket\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276964 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.276992 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-var-lib-openvswitch\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277018 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj6cl\" (UniqueName: \"kubernetes.io/projected/3b40c6b3-0061-4224-82d5-3ccf67998722-kube-api-access-dj6cl\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277053 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277080 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277108 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277161 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-var-lib-cni-bin\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277191 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277219 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b0c542d5-8c38-4243-8af7-cfc0d8e22773-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-kbtt2\" (UID: \"b0c542d5-8c38-4243-8af7-cfc0d8e22773\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277244 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-system-cni-dir\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277305 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b054ca08-1d09-4eca-a608-eb5b9323959a-os-release\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277328 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-cni-netd\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277353 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mq5zl\" (UniqueName: \"kubernetes.io/projected/9a50974e-f938-40f7-ace5-2a3b4cb1f3e7-kube-api-access-mq5zl\") pod \"node-ca-zwlc8\" (UID: \"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\") " pod="openshift-image-registry/node-ca-zwlc8" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277569 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-multus-cni-dir\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277646 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277757 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ec8e301-3037-4de0-94d2-32c49709660e-cni-binary-copy\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277787 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-multus-socket-dir-parent\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277814 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-var-lib-cni-multus\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277842 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277867 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-node-log\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277887 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9a50974e-f938-40f7-ace5-2a3b4cb1f3e7-host\") pod \"node-ca-zwlc8\" (UID: \"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\") " pod="openshift-image-registry/node-ca-zwlc8" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277909 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d5c4\" (UniqueName: \"kubernetes.io/projected/4ec8e301-3037-4de0-94d2-32c49709660e-kube-api-access-8d5c4\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.277995 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d77b18a7-7ad9-4bf5-bff5-da45878af7f4-mcd-auth-proxy-config\") pod \"machine-config-daemon-zkscb\" (UID: \"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\") " pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.278053 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-etc-openvswitch\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.278086 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-cni-bin\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.278595 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7bcc\" (UniqueName: \"kubernetes.io/projected/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-kube-api-access-q7bcc\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.278639 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.278670 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b054ca08-1d09-4eca-a608-eb5b9323959a-system-cni-dir\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.278695 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b054ca08-1d09-4eca-a608-eb5b9323959a-cnibin\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.278760 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.278829 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-systemd-units\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.278861 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-systemd\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.278889 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4ec8e301-3037-4de0-94d2-32c49709660e-multus-daemon-config\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.278963 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-ovnkube-script-lib\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.278995 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-var-lib-kubelet\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279025 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279052 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b29b9ad7-8cc9-434f-8731-a86265c383fd-hosts-file\") pod \"node-resolver-n55jt\" (UID: \"b29b9ad7-8cc9-434f-8731-a86265c383fd\") " pod="openshift-dns/node-resolver-n55jt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279077 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-run-k8s-cni-cncf-io\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279129 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffbwk\" (UniqueName: \"kubernetes.io/projected/b0c542d5-8c38-4243-8af7-cfc0d8e22773-kube-api-access-ffbwk\") pod \"ovnkube-control-plane-749d76644c-kbtt2\" (UID: \"b0c542d5-8c38-4243-8af7-cfc0d8e22773\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279153 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-run-multus-certs\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279175 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-kubelet\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279197 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-ovn\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279218 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-run-ovn-kubernetes\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279241 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b0c542d5-8c38-4243-8af7-cfc0d8e22773-env-overrides\") pod \"ovnkube-control-plane-749d76644c-kbtt2\" (UID: \"b0c542d5-8c38-4243-8af7-cfc0d8e22773\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279262 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d77b18a7-7ad9-4bf5-bff5-da45878af7f4-proxy-tls\") pod \"machine-config-daemon-zkscb\" (UID: \"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\") " pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279446 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-run-netns\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279474 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-os-release\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279547 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-hostroot\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279619 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-multus-conf-dir\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279653 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279679 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-slash\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279703 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-env-overrides\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279726 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279778 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-etc-kubernetes\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279829 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgn9f\" (UniqueName: \"kubernetes.io/projected/b29b9ad7-8cc9-434f-8731-a86265c383fd-kube-api-access-pgn9f\") pod \"node-resolver-n55jt\" (UID: \"b29b9ad7-8cc9-434f-8731-a86265c383fd\") " pod="openshift-dns/node-resolver-n55jt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279859 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.279923 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-cnibin\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.280840 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b054ca08-1d09-4eca-a608-eb5b9323959a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.280977 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.280996 4632 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281013 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281027 4632 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281043 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281059 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281073 4632 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281087 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281100 4632 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281114 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281126 4632 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281138 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281151 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281163 4632 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281176 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281189 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281202 4632 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281214 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281228 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281241 4632 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281254 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281266 4632 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281278 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281290 4632 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281302 4632 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281316 4632 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.281330 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.282385 4632 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.282601 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.283051 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:10.783028261 +0000 UTC m=+84.805558394 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.288449 4632 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.288550 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:10.788533611 +0000 UTC m=+84.811063744 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.289626 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.289688 4632 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.289780 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.289908 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:05:10.789893293 +0000 UTC m=+84.812423436 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.290183 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.293538 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.294120 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.294673 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.295666 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.296292 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.296909 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.298864 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.303275 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.304214 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.304246 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.304582 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.304730 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.305349 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.307141 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.307427 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.307455 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.307667 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.308597 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.308659 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.308820 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.309010 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.309119 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.309643 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.309661 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.310362 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.310390 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.310638 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.310655 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.310754 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.310817 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.310875 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.310891 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.311666 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.311748 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.311782 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.312248 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.312378 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.315216 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.315233 4632 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.315279 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.315301 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:10.815276866 +0000 UTC m=+84.837806999 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.315489 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.315750 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.315790 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.315910 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.315969 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.312671 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.312998 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.313011 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.313460 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.313527 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.313628 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.313646 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.313872 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.314111 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.314130 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.314380 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.314477 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.314742 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.314818 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.315853 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.316279 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.316435 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.311119 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.315867 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.316694 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.316771 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.316782 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.316850 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.316873 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.317203 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.312640 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.317250 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.317312 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.317413 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.317450 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.317487 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.318307 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.318327 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.318382 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.318410 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.318441 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.318548 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.318680 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.318791 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.318891 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.318930 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.319115 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.319129 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.319190 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.319233 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.319241 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.319255 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.319365 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.319456 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.319531 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.319703 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.320154 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.320243 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.320269 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.320282 4632 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.320318 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:10.8203068 +0000 UTC m=+84.842836933 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.320379 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.321067 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.321182 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.321654 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.325692 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.326018 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.326342 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.326488 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.326729 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.328097 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.328363 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.328769 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.328768 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.328878 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.328974 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.329017 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.329038 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.329208 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.329271 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.329280 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.329290 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.329670 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.329778 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.330220 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.330231 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.330546 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.330594 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.330776 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.330864 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.330938 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.331072 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.331183 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.331672 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.331862 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.332127 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.332159 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.332332 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.332358 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.332359 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.332681 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.332814 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.332902 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.333014 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.333097 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.333172 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.333742 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.333738 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.334299 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.334357 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.335395 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.335416 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.335481 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.335609 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.335792 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.336076 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.336108 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.336216 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.336411 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.336503 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.337765 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.337801 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.337816 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.337831 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.337842 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:10Z","lastTransitionTime":"2026-03-13T10:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.338355 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.338541 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.338577 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.338892 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.339534 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.339771 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.339882 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.340119 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.340319 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.340430 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.340473 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.341195 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.342713 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.344264 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.345087 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.345368 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.345545 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.355505 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.362072 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.363281 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.373028 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.382694 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-log-socket\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.382725 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.382742 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9a50974e-f938-40f7-ace5-2a3b4cb1f3e7-serviceca\") pod \"node-ca-zwlc8\" (UID: \"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\") " pod="openshift-image-registry/node-ca-zwlc8" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.382772 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9s5s\" (UniqueName: \"kubernetes.io/projected/b054ca08-1d09-4eca-a608-eb5b9323959a-kube-api-access-l9s5s\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.382793 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-log-socket\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.382796 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-var-lib-openvswitch\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.382847 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj6cl\" (UniqueName: \"kubernetes.io/projected/3b40c6b3-0061-4224-82d5-3ccf67998722-kube-api-access-dj6cl\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.382865 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-system-cni-dir\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.382880 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-var-lib-cni-bin\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.382924 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b0c542d5-8c38-4243-8af7-cfc0d8e22773-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-kbtt2\" (UID: \"b0c542d5-8c38-4243-8af7-cfc0d8e22773\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.382967 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b054ca08-1d09-4eca-a608-eb5b9323959a-os-release\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.382989 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-cni-netd\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.383009 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mq5zl\" (UniqueName: \"kubernetes.io/projected/9a50974e-f938-40f7-ace5-2a3b4cb1f3e7-kube-api-access-mq5zl\") pod \"node-ca-zwlc8\" (UID: \"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\") " pod="openshift-image-registry/node-ca-zwlc8" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.383034 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.383085 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-system-cni-dir\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.383104 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-var-lib-openvswitch\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.383372 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/b054ca08-1d09-4eca-a608-eb5b9323959a-os-release\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.383413 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-cni-netd\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.383448 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-var-lib-cni-bin\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.383798 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-multus-cni-dir\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.383834 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-multus-socket-dir-parent\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.383852 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9a50974e-f938-40f7-ace5-2a3b4cb1f3e7-serviceca\") pod \"node-ca-zwlc8\" (UID: \"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\") " pod="openshift-image-registry/node-ca-zwlc8" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.383883 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ec8e301-3037-4de0-94d2-32c49709660e-cni-binary-copy\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.383903 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-var-lib-cni-multus\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.384043 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-multus-socket-dir-parent\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.384073 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-var-lib-cni-multus\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.384059 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-multus-cni-dir\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.383924 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d77b18a7-7ad9-4bf5-bff5-da45878af7f4-mcd-auth-proxy-config\") pod \"machine-config-daemon-zkscb\" (UID: \"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\") " pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.384683 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-node-log\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.384829 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9a50974e-f938-40f7-ace5-2a3b4cb1f3e7-host\") pod \"node-ca-zwlc8\" (UID: \"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\") " pod="openshift-image-registry/node-ca-zwlc8" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.384855 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d5c4\" (UniqueName: \"kubernetes.io/projected/4ec8e301-3037-4de0-94d2-32c49709660e-kube-api-access-8d5c4\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.385004 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-etc-openvswitch\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.385147 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-cni-bin\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.385177 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7bcc\" (UniqueName: \"kubernetes.io/projected/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-kube-api-access-q7bcc\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.385267 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4ec8e301-3037-4de0-94d2-32c49709660e-cni-binary-copy\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.385309 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4ec8e301-3037-4de0-94d2-32c49709660e-multus-daemon-config\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.385427 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b054ca08-1d09-4eca-a608-eb5b9323959a-system-cni-dir\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.385566 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b054ca08-1d09-4eca-a608-eb5b9323959a-cnibin\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.385605 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-systemd-units\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.386411 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-systemd\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.386440 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-ovnkube-script-lib\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.386583 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-run-k8s-cni-cncf-io\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.386611 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-var-lib-kubelet\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.386743 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-systemd-units\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.386757 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b29b9ad7-8cc9-434f-8731-a86265c383fd-hosts-file\") pod \"node-resolver-n55jt\" (UID: \"b29b9ad7-8cc9-434f-8731-a86265c383fd\") " pod="openshift-dns/node-resolver-n55jt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.386896 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b0c542d5-8c38-4243-8af7-cfc0d8e22773-env-overrides\") pod \"ovnkube-control-plane-749d76644c-kbtt2\" (UID: \"b0c542d5-8c38-4243-8af7-cfc0d8e22773\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.386965 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffbwk\" (UniqueName: \"kubernetes.io/projected/b0c542d5-8c38-4243-8af7-cfc0d8e22773-kube-api-access-ffbwk\") pod \"ovnkube-control-plane-749d76644c-kbtt2\" (UID: \"b0c542d5-8c38-4243-8af7-cfc0d8e22773\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.386994 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-run-multus-certs\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387071 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b29b9ad7-8cc9-434f-8731-a86265c383fd-hosts-file\") pod \"node-resolver-n55jt\" (UID: \"b29b9ad7-8cc9-434f-8731-a86265c383fd\") " pod="openshift-dns/node-resolver-n55jt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387067 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-kubelet\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387136 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-kubelet\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387159 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4ec8e301-3037-4de0-94d2-32c49709660e-multus-daemon-config\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387172 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/b054ca08-1d09-4eca-a608-eb5b9323959a-system-cni-dir\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387217 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/b054ca08-1d09-4eca-a608-eb5b9323959a-cnibin\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387232 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-ovn\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387269 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-run-ovn-kubernetes\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.384853 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-node-log\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.384764 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/d77b18a7-7ad9-4bf5-bff5-da45878af7f4-mcd-auth-proxy-config\") pod \"machine-config-daemon-zkscb\" (UID: \"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\") " pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.384888 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9a50974e-f938-40f7-ace5-2a3b4cb1f3e7-host\") pod \"node-ca-zwlc8\" (UID: \"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\") " pod="openshift-image-registry/node-ca-zwlc8" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387473 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-multus-conf-dir\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387485 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-run-ovn-kubernetes\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387508 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d77b18a7-7ad9-4bf5-bff5-da45878af7f4-proxy-tls\") pod \"machine-config-daemon-zkscb\" (UID: \"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\") " pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387585 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-run-netns\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387704 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-multus-conf-dir\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387825 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-run-k8s-cni-cncf-io\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387612 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-os-release\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387871 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-hostroot\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.387930 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-ovn\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388144 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-etc-openvswitch\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388210 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-systemd\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388248 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-cni-bin\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388507 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b0c542d5-8c38-4243-8af7-cfc0d8e22773-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-kbtt2\" (UID: \"b0c542d5-8c38-4243-8af7-cfc0d8e22773\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388570 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-run-multus-certs\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388605 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgn9f\" (UniqueName: \"kubernetes.io/projected/b29b9ad7-8cc9-434f-8731-a86265c383fd-kube-api-access-pgn9f\") pod \"node-resolver-n55jt\" (UID: \"b29b9ad7-8cc9-434f-8731-a86265c383fd\") " pod="openshift-dns/node-resolver-n55jt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388659 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-slash\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388719 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-env-overrides\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388811 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-os-release\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388850 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-var-lib-kubelet\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388879 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-run-netns\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388744 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388899 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-slash\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388912 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-hostroot\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.388974 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-etc-kubernetes\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389020 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b054ca08-1d09-4eca-a608-eb5b9323959a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389046 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389071 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-cnibin\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389124 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-run-netns\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389148 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d77b18a7-7ad9-4bf5-bff5-da45878af7f4-rootfs\") pod \"machine-config-daemon-zkscb\" (UID: \"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\") " pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389173 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnh6w\" (UniqueName: \"kubernetes.io/projected/d77b18a7-7ad9-4bf5-bff5-da45878af7f4-kube-api-access-vnh6w\") pod \"machine-config-daemon-zkscb\" (UID: \"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\") " pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389195 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b054ca08-1d09-4eca-a608-eb5b9323959a-cni-binary-copy\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389217 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b054ca08-1d09-4eca-a608-eb5b9323959a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389243 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389718 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-env-overrides\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389723 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-ovnkube-script-lib\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389801 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-host-run-netns\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389806 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b0c542d5-8c38-4243-8af7-cfc0d8e22773-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-kbtt2\" (UID: \"b0c542d5-8c38-4243-8af7-cfc0d8e22773\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389843 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3b40c6b3-0061-4224-82d5-3ccf67998722-ovn-node-metrics-cert\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389869 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-openvswitch\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.389894 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-ovnkube-config\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.390790 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b0c542d5-8c38-4243-8af7-cfc0d8e22773-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-kbtt2\" (UID: \"b0c542d5-8c38-4243-8af7-cfc0d8e22773\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.390867 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.390893 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-ovnkube-config\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.390952 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-cnibin\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.391671 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/b054ca08-1d09-4eca-a608-eb5b9323959a-tuning-conf-dir\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.391700 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-openvswitch\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.391934 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/b054ca08-1d09-4eca-a608-eb5b9323959a-cni-binary-copy\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.397671 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.397979 4632 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.398097 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs podName:ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad nodeName:}" failed. No retries permitted until 2026-03-13 10:05:10.898062949 +0000 UTC m=+84.920593082 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs") pod "network-metrics-daemon-z2vlz" (UID: "ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.398547 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/d77b18a7-7ad9-4bf5-bff5-da45878af7f4-rootfs\") pod \"machine-config-daemon-zkscb\" (UID: \"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\") " pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.399029 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b0c542d5-8c38-4243-8af7-cfc0d8e22773-env-overrides\") pod \"ovnkube-control-plane-749d76644c-kbtt2\" (UID: \"b0c542d5-8c38-4243-8af7-cfc0d8e22773\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.399100 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/4ec8e301-3037-4de0-94d2-32c49709660e-etc-kubernetes\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.399691 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/b054ca08-1d09-4eca-a608-eb5b9323959a-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400137 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400170 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400185 4632 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400200 4632 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400212 4632 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400230 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400256 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400352 4632 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400370 4632 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400389 4632 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400433 4632 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400617 4632 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400638 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400654 4632 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400698 4632 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400883 4632 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400907 4632 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400962 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.400978 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.401224 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.401278 4632 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.401297 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.401311 4632 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.401328 4632 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.401553 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.401574 4632 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.401588 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402098 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402140 4632 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402165 4632 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402179 4632 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402198 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402211 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402224 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402238 4632 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402316 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/d77b18a7-7ad9-4bf5-bff5-da45878af7f4-proxy-tls\") pod \"machine-config-daemon-zkscb\" (UID: \"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\") " pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402249 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402358 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402390 4632 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402412 4632 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402613 4632 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402629 4632 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402644 4632 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402660 4632 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402680 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402694 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402708 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402727 4632 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402742 4632 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402818 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402835 4632 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402853 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402867 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.402987 4632 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403005 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403026 4632 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403040 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403129 4632 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403143 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403160 4632 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403174 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403190 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403208 4632 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403221 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403235 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403247 4632 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403268 4632 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403281 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403294 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403310 4632 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403328 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403340 4632 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403352 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403370 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403384 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403398 4632 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403410 4632 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403429 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403443 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403563 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403578 4632 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403596 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403608 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403621 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403634 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403654 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403667 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403680 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403697 4632 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403709 4632 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403722 4632 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403737 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403755 4632 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403768 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403780 4632 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403792 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403809 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403823 4632 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403836 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403850 4632 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403868 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403884 4632 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403896 4632 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403915 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403929 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403962 4632 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403976 4632 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.403997 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404011 4632 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404025 4632 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404038 4632 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404057 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404069 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404082 4632 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404099 4632 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404111 4632 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404123 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404135 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404151 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404159 4632 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404169 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404182 4632 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404198 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404211 4632 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404222 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404235 4632 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404252 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404266 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404279 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404295 4632 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404306 4632 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404318 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404328 4632 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404345 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404358 4632 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404371 4632 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404382 4632 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404397 4632 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404410 4632 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404421 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404436 4632 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404448 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404462 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404474 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404490 4632 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404503 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404518 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404532 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404551 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404563 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404576 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404588 4632 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404605 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404618 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404631 4632 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404646 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404660 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404674 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404687 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404705 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404717 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404729 4632 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404743 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404761 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404774 4632 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404787 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404802 4632 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.404818 4632 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.407051 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.407518 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj6cl\" (UniqueName: \"kubernetes.io/projected/3b40c6b3-0061-4224-82d5-3ccf67998722-kube-api-access-dj6cl\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.408265 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7bcc\" (UniqueName: \"kubernetes.io/projected/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-kube-api-access-q7bcc\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.412211 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3b40c6b3-0061-4224-82d5-3ccf67998722-ovn-node-metrics-cert\") pod \"ovnkube-node-qb725\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.412733 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mq5zl\" (UniqueName: \"kubernetes.io/projected/9a50974e-f938-40f7-ace5-2a3b4cb1f3e7-kube-api-access-mq5zl\") pod \"node-ca-zwlc8\" (UID: \"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\") " pod="openshift-image-registry/node-ca-zwlc8" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.414195 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d5c4\" (UniqueName: \"kubernetes.io/projected/4ec8e301-3037-4de0-94d2-32c49709660e-kube-api-access-8d5c4\") pod \"multus-gqf22\" (UID: \"4ec8e301-3037-4de0-94d2-32c49709660e\") " pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.414425 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9s5s\" (UniqueName: \"kubernetes.io/projected/b054ca08-1d09-4eca-a608-eb5b9323959a-kube-api-access-l9s5s\") pod \"multus-additional-cni-plugins-qlc8m\" (UID: \"b054ca08-1d09-4eca-a608-eb5b9323959a\") " pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.416653 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffbwk\" (UniqueName: \"kubernetes.io/projected/b0c542d5-8c38-4243-8af7-cfc0d8e22773-kube-api-access-ffbwk\") pod \"ovnkube-control-plane-749d76644c-kbtt2\" (UID: \"b0c542d5-8c38-4243-8af7-cfc0d8e22773\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.418931 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnh6w\" (UniqueName: \"kubernetes.io/projected/d77b18a7-7ad9-4bf5-bff5-da45878af7f4-kube-api-access-vnh6w\") pod \"machine-config-daemon-zkscb\" (UID: \"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\") " pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.419426 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgn9f\" (UniqueName: \"kubernetes.io/projected/b29b9ad7-8cc9-434f-8731-a86265c383fd-kube-api-access-pgn9f\") pod \"node-resolver-n55jt\" (UID: \"b29b9ad7-8cc9-434f-8731-a86265c383fd\") " pod="openshift-dns/node-resolver-n55jt" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.420713 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.425649 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 13 10:05:10 crc kubenswrapper[4632]: set -o allexport Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: source /etc/kubernetes/apiserver-url.env Mar 13 10:05:10 crc kubenswrapper[4632]: else Mar 13 10:05:10 crc kubenswrapper[4632]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 13 10:05:10 crc kubenswrapper[4632]: exit 1 Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 13 10:05:10 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.426958 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.431137 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.432730 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ -f "/env/_master" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: set -o allexport Mar 13 10:05:10 crc kubenswrapper[4632]: source "/env/_master" Mar 13 10:05:10 crc kubenswrapper[4632]: set +o allexport Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 13 10:05:10 crc kubenswrapper[4632]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 13 10:05:10 crc kubenswrapper[4632]: ho_enable="--enable-hybrid-overlay" Mar 13 10:05:10 crc kubenswrapper[4632]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 13 10:05:10 crc kubenswrapper[4632]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 13 10:05:10 crc kubenswrapper[4632]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 13 10:05:10 crc kubenswrapper[4632]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 13 10:05:10 crc kubenswrapper[4632]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 13 10:05:10 crc kubenswrapper[4632]: --webhook-host=127.0.0.1 \ Mar 13 10:05:10 crc kubenswrapper[4632]: --webhook-port=9743 \ Mar 13 10:05:10 crc kubenswrapper[4632]: ${ho_enable} \ Mar 13 10:05:10 crc kubenswrapper[4632]: --enable-interconnect \ Mar 13 10:05:10 crc kubenswrapper[4632]: --disable-approver \ Mar 13 10:05:10 crc kubenswrapper[4632]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 13 10:05:10 crc kubenswrapper[4632]: --wait-for-kubernetes-api=200s \ Mar 13 10:05:10 crc kubenswrapper[4632]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 13 10:05:10 crc kubenswrapper[4632]: --loglevel="${LOGLEVEL}" Mar 13 10:05:10 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.437817 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ -f "/env/_master" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: set -o allexport Mar 13 10:05:10 crc kubenswrapper[4632]: source "/env/_master" Mar 13 10:05:10 crc kubenswrapper[4632]: set +o allexport Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 13 10:05:10 crc kubenswrapper[4632]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 13 10:05:10 crc kubenswrapper[4632]: --disable-webhook \ Mar 13 10:05:10 crc kubenswrapper[4632]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 13 10:05:10 crc kubenswrapper[4632]: --loglevel="${LOGLEVEL}" Mar 13 10:05:10 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.438959 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.439420 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-gqf22" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.440646 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.440701 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.440713 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.440772 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.440793 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:10Z","lastTransitionTime":"2026-03-13T10:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:10 crc kubenswrapper[4632]: W0313 10:05:10.443738 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-8b7cb438273be81daa37dee852b13428c2abbaceb49c8c8e2ae84fffcf8cf261 WatchSource:0}: Error finding container 8b7cb438273be81daa37dee852b13428c2abbaceb49c8c8e2ae84fffcf8cf261: Status 404 returned error can't find the container with id 8b7cb438273be81daa37dee852b13428c2abbaceb49c8c8e2ae84fffcf8cf261 Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.447834 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-n55jt" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.448373 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.450929 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.459727 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.460254 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Mar 13 10:05:10 crc kubenswrapper[4632]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Mar 13 10:05:10 crc kubenswrapper[4632]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8d5c4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-gqf22_openshift-multus(4ec8e301-3037-4de0-94d2-32c49709660e): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.461547 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-gqf22" podUID="4ec8e301-3037-4de0-94d2-32c49709660e" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.465821 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" Mar 13 10:05:10 crc kubenswrapper[4632]: W0313 10:05:10.466442 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb29b9ad7_8cc9_434f_8731_a86265c383fd.slice/crio-12fc45120502c7a3991aa07e36d421f9b60ca97adfa86accdf7d5f45f50ccde5 WatchSource:0}: Error finding container 12fc45120502c7a3991aa07e36d421f9b60ca97adfa86accdf7d5f45f50ccde5: Status 404 returned error can't find the container with id 12fc45120502c7a3991aa07e36d421f9b60ca97adfa86accdf7d5f45f50ccde5 Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.469128 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/bin/bash -c #!/bin/bash Mar 13 10:05:10 crc kubenswrapper[4632]: set -uo pipefail Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Mar 13 10:05:10 crc kubenswrapper[4632]: HOSTS_FILE="/etc/hosts" Mar 13 10:05:10 crc kubenswrapper[4632]: TEMP_FILE="/etc/hosts.tmp" Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: IFS=', ' read -r -a services <<< "${SERVICES}" Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: # Make a temporary file with the old hosts file's attributes. Mar 13 10:05:10 crc kubenswrapper[4632]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Mar 13 10:05:10 crc kubenswrapper[4632]: echo "Failed to preserve hosts file. Exiting." Mar 13 10:05:10 crc kubenswrapper[4632]: exit 1 Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: while true; do Mar 13 10:05:10 crc kubenswrapper[4632]: declare -A svc_ips Mar 13 10:05:10 crc kubenswrapper[4632]: for svc in "${services[@]}"; do Mar 13 10:05:10 crc kubenswrapper[4632]: # Fetch service IP from cluster dns if present. We make several tries Mar 13 10:05:10 crc kubenswrapper[4632]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Mar 13 10:05:10 crc kubenswrapper[4632]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Mar 13 10:05:10 crc kubenswrapper[4632]: # support UDP loadbalancers and require reaching DNS through TCP. Mar 13 10:05:10 crc kubenswrapper[4632]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 13 10:05:10 crc kubenswrapper[4632]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 13 10:05:10 crc kubenswrapper[4632]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 13 10:05:10 crc kubenswrapper[4632]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Mar 13 10:05:10 crc kubenswrapper[4632]: for i in ${!cmds[*]} Mar 13 10:05:10 crc kubenswrapper[4632]: do Mar 13 10:05:10 crc kubenswrapper[4632]: ips=($(eval "${cmds[i]}")) Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: svc_ips["${svc}"]="${ips[@]}" Mar 13 10:05:10 crc kubenswrapper[4632]: break Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: # Update /etc/hosts only if we get valid service IPs Mar 13 10:05:10 crc kubenswrapper[4632]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Mar 13 10:05:10 crc kubenswrapper[4632]: # Stale entries could exist in /etc/hosts if the service is deleted Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ -n "${svc_ips[*]-}" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Mar 13 10:05:10 crc kubenswrapper[4632]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Mar 13 10:05:10 crc kubenswrapper[4632]: # Only continue rebuilding the hosts entries if its original content is preserved Mar 13 10:05:10 crc kubenswrapper[4632]: sleep 60 & wait Mar 13 10:05:10 crc kubenswrapper[4632]: continue Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: # Append resolver entries for services Mar 13 10:05:10 crc kubenswrapper[4632]: rc=0 Mar 13 10:05:10 crc kubenswrapper[4632]: for svc in "${!svc_ips[@]}"; do Mar 13 10:05:10 crc kubenswrapper[4632]: for ip in ${svc_ips[${svc}]}; do Mar 13 10:05:10 crc kubenswrapper[4632]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ $rc -ne 0 ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: sleep 60 & wait Mar 13 10:05:10 crc kubenswrapper[4632]: continue Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Mar 13 10:05:10 crc kubenswrapper[4632]: # Replace /etc/hosts with our modified version if needed Mar 13 10:05:10 crc kubenswrapper[4632]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Mar 13 10:05:10 crc kubenswrapper[4632]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: sleep 60 & wait Mar 13 10:05:10 crc kubenswrapper[4632]: unset svc_ips Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pgn9f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-n55jt_openshift-dns(b29b9ad7-8cc9-434f-8731-a86265c383fd): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.470380 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-n55jt" podUID="b29b9ad7-8cc9-434f-8731-a86265c383fd" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.472895 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-zwlc8" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.480574 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.482773 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.18.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vnh6w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.489378 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.490122 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vnh6w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.491886 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.499327 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l9s5s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-qlc8m_openshift-multus(b054ca08-1d09-4eca-a608-eb5b9323959a): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.500489 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: container &Container{Name:node-ca,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Mar 13 10:05:10 crc kubenswrapper[4632]: while [ true ]; Mar 13 10:05:10 crc kubenswrapper[4632]: do Mar 13 10:05:10 crc kubenswrapper[4632]: for f in $(ls /tmp/serviceca); do Mar 13 10:05:10 crc kubenswrapper[4632]: echo $f Mar 13 10:05:10 crc kubenswrapper[4632]: ca_file_path="/tmp/serviceca/${f}" Mar 13 10:05:10 crc kubenswrapper[4632]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Mar 13 10:05:10 crc kubenswrapper[4632]: reg_dir_path="/etc/docker/certs.d/${f}" Mar 13 10:05:10 crc kubenswrapper[4632]: if [ -e "${reg_dir_path}" ]; then Mar 13 10:05:10 crc kubenswrapper[4632]: cp -u $ca_file_path $reg_dir_path/ca.crt Mar 13 10:05:10 crc kubenswrapper[4632]: else Mar 13 10:05:10 crc kubenswrapper[4632]: mkdir $reg_dir_path Mar 13 10:05:10 crc kubenswrapper[4632]: cp $ca_file_path $reg_dir_path/ca.crt Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: for d in $(ls /etc/docker/certs.d); do Mar 13 10:05:10 crc kubenswrapper[4632]: echo $d Mar 13 10:05:10 crc kubenswrapper[4632]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Mar 13 10:05:10 crc kubenswrapper[4632]: reg_conf_path="/tmp/serviceca/${dp}" Mar 13 10:05:10 crc kubenswrapper[4632]: if [ ! -e "${reg_conf_path}" ]; then Mar 13 10:05:10 crc kubenswrapper[4632]: rm -rf /etc/docker/certs.d/$d Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: sleep 60 & wait ${!} Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mq5zl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-ca-zwlc8_openshift-image-registry(9a50974e-f938-40f7-ace5-2a3b4cb1f3e7): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.500581 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" podUID="b054ca08-1d09-4eca-a608-eb5b9323959a" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.501651 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-zwlc8" podUID="9a50974e-f938-40f7-ace5-2a3b4cb1f3e7" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.504197 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"b6c337d9d2e58ca0167b79bc2a82693282b889a5339b46a5201132b60dec013f"} Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.506390 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-n55jt" event={"ID":"b29b9ad7-8cc9-434f-8731-a86265c383fd","Type":"ContainerStarted","Data":"12fc45120502c7a3991aa07e36d421f9b60ca97adfa86accdf7d5f45f50ccde5"} Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.506510 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:machine-config-daemon,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a,Command:[/usr/bin/machine-config-daemon],Args:[start --payload-version=4.18.1],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:health,HostPort:8798,ContainerPort:8798,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:rootfs,ReadOnly:false,MountPath:/rootfs,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vnh6w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/health,Port:{0 8798 },Host:127.0.0.1,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:120,TimeoutSeconds:1,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.508610 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: container &Container{Name:dns-node-resolver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/bin/bash -c #!/bin/bash Mar 13 10:05:10 crc kubenswrapper[4632]: set -uo pipefail Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: trap 'jobs -p | xargs kill || true; wait; exit 0' TERM Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: OPENSHIFT_MARKER="openshift-generated-node-resolver" Mar 13 10:05:10 crc kubenswrapper[4632]: HOSTS_FILE="/etc/hosts" Mar 13 10:05:10 crc kubenswrapper[4632]: TEMP_FILE="/etc/hosts.tmp" Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: IFS=', ' read -r -a services <<< "${SERVICES}" Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: # Make a temporary file with the old hosts file's attributes. Mar 13 10:05:10 crc kubenswrapper[4632]: if ! cp -f --attributes-only "${HOSTS_FILE}" "${TEMP_FILE}"; then Mar 13 10:05:10 crc kubenswrapper[4632]: echo "Failed to preserve hosts file. Exiting." Mar 13 10:05:10 crc kubenswrapper[4632]: exit 1 Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: while true; do Mar 13 10:05:10 crc kubenswrapper[4632]: declare -A svc_ips Mar 13 10:05:10 crc kubenswrapper[4632]: for svc in "${services[@]}"; do Mar 13 10:05:10 crc kubenswrapper[4632]: # Fetch service IP from cluster dns if present. We make several tries Mar 13 10:05:10 crc kubenswrapper[4632]: # to do it: IPv4, IPv6, IPv4 over TCP and IPv6 over TCP. The two last ones Mar 13 10:05:10 crc kubenswrapper[4632]: # are for deployments with Kuryr on older OpenStack (OSP13) - those do not Mar 13 10:05:10 crc kubenswrapper[4632]: # support UDP loadbalancers and require reaching DNS through TCP. Mar 13 10:05:10 crc kubenswrapper[4632]: cmds=('dig -t A @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 13 10:05:10 crc kubenswrapper[4632]: 'dig -t AAAA @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 13 10:05:10 crc kubenswrapper[4632]: 'dig -t A +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"' Mar 13 10:05:10 crc kubenswrapper[4632]: 'dig -t AAAA +tcp +retry=0 @"${NAMESERVER}" +short "${svc}.${CLUSTER_DOMAIN}"|grep -v "^;"') Mar 13 10:05:10 crc kubenswrapper[4632]: for i in ${!cmds[*]} Mar 13 10:05:10 crc kubenswrapper[4632]: do Mar 13 10:05:10 crc kubenswrapper[4632]: ips=($(eval "${cmds[i]}")) Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ "$?" -eq 0 && "${#ips[@]}" -ne 0 ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: svc_ips["${svc}"]="${ips[@]}" Mar 13 10:05:10 crc kubenswrapper[4632]: break Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: # Update /etc/hosts only if we get valid service IPs Mar 13 10:05:10 crc kubenswrapper[4632]: # We will not update /etc/hosts when there is coredns service outage or api unavailability Mar 13 10:05:10 crc kubenswrapper[4632]: # Stale entries could exist in /etc/hosts if the service is deleted Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ -n "${svc_ips[*]-}" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: # Build a new hosts file from /etc/hosts with our custom entries filtered out Mar 13 10:05:10 crc kubenswrapper[4632]: if ! sed --silent "/# ${OPENSHIFT_MARKER}/d; w ${TEMP_FILE}" "${HOSTS_FILE}"; then Mar 13 10:05:10 crc kubenswrapper[4632]: # Only continue rebuilding the hosts entries if its original content is preserved Mar 13 10:05:10 crc kubenswrapper[4632]: sleep 60 & wait Mar 13 10:05:10 crc kubenswrapper[4632]: continue Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: # Append resolver entries for services Mar 13 10:05:10 crc kubenswrapper[4632]: rc=0 Mar 13 10:05:10 crc kubenswrapper[4632]: for svc in "${!svc_ips[@]}"; do Mar 13 10:05:10 crc kubenswrapper[4632]: for ip in ${svc_ips[${svc}]}; do Mar 13 10:05:10 crc kubenswrapper[4632]: echo "${ip} ${svc} ${svc}.${CLUSTER_DOMAIN} # ${OPENSHIFT_MARKER}" >> "${TEMP_FILE}" || rc=$? Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ $rc -ne 0 ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: sleep 60 & wait Mar 13 10:05:10 crc kubenswrapper[4632]: continue Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: # TODO: Update /etc/hosts atomically to avoid any inconsistent behavior Mar 13 10:05:10 crc kubenswrapper[4632]: # Replace /etc/hosts with our modified version if needed Mar 13 10:05:10 crc kubenswrapper[4632]: cmp "${TEMP_FILE}" "${HOSTS_FILE}" || cp -f "${TEMP_FILE}" "${HOSTS_FILE}" Mar 13 10:05:10 crc kubenswrapper[4632]: # TEMP_FILE is not removed to avoid file create/delete and attributes copy churn Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: sleep 60 & wait Mar 13 10:05:10 crc kubenswrapper[4632]: unset svc_ips Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:SERVICES,Value:image-registry.openshift-image-registry.svc,ValueFrom:nil,},EnvVar{Name:NAMESERVER,Value:10.217.4.10,ValueFrom:nil,},EnvVar{Name:CLUSTER_DOMAIN,Value:cluster.local,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{22020096 0} {} 21Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:hosts-file,ReadOnly:false,MountPath:/etc/hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pgn9f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-resolver-n55jt_openshift-dns(b29b9ad7-8cc9-434f-8731-a86265c383fd): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.509700 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dns-node-resolver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-dns/node-resolver-n55jt" podUID="b29b9ad7-8cc9-434f-8731-a86265c383fd" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.516563 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[],Args:[--secure-listen-address=0.0.0.0:9001 --config-file=/etc/kube-rbac-proxy/config-file.yaml --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12 --upstream=http://127.0.0.1:8797 --logtostderr=true --tls-cert-file=/etc/tls/private/tls.crt --tls-private-key-file=/etc/tls/private/tls.key],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:9001,ContainerPort:9001,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:proxy-tls,ReadOnly:false,MountPath:/etc/tls/private,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:mcd-auth-proxy-config,ReadOnly:false,MountPath:/etc/kube-rbac-proxy,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-vnh6w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.516629 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[/bin/bash -c #!/bin/bash Mar 13 10:05:10 crc kubenswrapper[4632]: set -euo pipefail Mar 13 10:05:10 crc kubenswrapper[4632]: TLS_PK=/etc/pki/tls/metrics-cert/tls.key Mar 13 10:05:10 crc kubenswrapper[4632]: TLS_CERT=/etc/pki/tls/metrics-cert/tls.crt Mar 13 10:05:10 crc kubenswrapper[4632]: # As the secret mount is optional we must wait for the files to be present. Mar 13 10:05:10 crc kubenswrapper[4632]: # The service is created in monitor.yaml and this is created in sdn.yaml. Mar 13 10:05:10 crc kubenswrapper[4632]: TS=$(date +%s) Mar 13 10:05:10 crc kubenswrapper[4632]: WARN_TS=$(( ${TS} + $(( 20 * 60)) )) Mar 13 10:05:10 crc kubenswrapper[4632]: HAS_LOGGED_INFO=0 Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: log_missing_certs(){ Mar 13 10:05:10 crc kubenswrapper[4632]: CUR_TS=$(date +%s) Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ "${CUR_TS}" -gt "WARN_TS" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: echo $(date -Iseconds) WARN: ovn-control-plane-metrics-cert not mounted after 20 minutes. Mar 13 10:05:10 crc kubenswrapper[4632]: elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then Mar 13 10:05:10 crc kubenswrapper[4632]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-cert not mounted. Waiting 20 minutes. Mar 13 10:05:10 crc kubenswrapper[4632]: HAS_LOGGED_INFO=1 Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: } Mar 13 10:05:10 crc kubenswrapper[4632]: while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do Mar 13 10:05:10 crc kubenswrapper[4632]: log_missing_certs Mar 13 10:05:10 crc kubenswrapper[4632]: sleep 5 Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-certs mounted, starting kube-rbac-proxy Mar 13 10:05:10 crc kubenswrapper[4632]: exec /usr/bin/kube-rbac-proxy \ Mar 13 10:05:10 crc kubenswrapper[4632]: --logtostderr \ Mar 13 10:05:10 crc kubenswrapper[4632]: --secure-listen-address=:9108 \ Mar 13 10:05:10 crc kubenswrapper[4632]: --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 \ Mar 13 10:05:10 crc kubenswrapper[4632]: --upstream=http://127.0.0.1:29108/ \ Mar 13 10:05:10 crc kubenswrapper[4632]: --tls-private-key-file=${TLS_PK} \ Mar 13 10:05:10 crc kubenswrapper[4632]: --tls-cert-file=${TLS_CERT} Mar 13 10:05:10 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:9108,ContainerPort:9108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovn-control-plane-metrics-cert,ReadOnly:true,MountPath:/etc/pki/tls/metrics-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ffbwk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-kbtt2_openshift-ovn-kubernetes(b0c542d5-8c38-4243-8af7-cfc0d8e22773): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.517430 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.518181 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"machine-config-daemon\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.518867 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gqf22" event={"ID":"4ec8e301-3037-4de0-94d2-32c49709660e","Type":"ContainerStarted","Data":"59b4c4e96c104a54f846e291518c93a1b3e3a63ef11982c82cf3ed1b26b885f0"} Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.520826 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: container &Container{Name:ovnkube-cluster-manager,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ -f "/env/_master" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: set -o allexport Mar 13 10:05:10 crc kubenswrapper[4632]: source "/env/_master" Mar 13 10:05:10 crc kubenswrapper[4632]: set +o allexport Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: ovn_v4_join_subnet_opt= Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ "" != "" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: ovn_v4_join_subnet_opt="--gateway-v4-join-subnet " Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: ovn_v6_join_subnet_opt= Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ "" != "" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: ovn_v6_join_subnet_opt="--gateway-v6-join-subnet " Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: ovn_v4_transit_switch_subnet_opt= Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ "" != "" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: ovn_v4_transit_switch_subnet_opt="--cluster-manager-v4-transit-switch-subnet " Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: ovn_v6_transit_switch_subnet_opt= Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ "" != "" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: ovn_v6_transit_switch_subnet_opt="--cluster-manager-v6-transit-switch-subnet " Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: dns_name_resolver_enabled_flag= Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ "false" == "true" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: dns_name_resolver_enabled_flag="--enable-dns-name-resolver" Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: persistent_ips_enabled_flag= Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ "true" == "true" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: persistent_ips_enabled_flag="--enable-persistent-ips" Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: # This is needed so that converting clusters from GA to TP Mar 13 10:05:10 crc kubenswrapper[4632]: # will rollout control plane pods as well Mar 13 10:05:10 crc kubenswrapper[4632]: network_segmentation_enabled_flag= Mar 13 10:05:10 crc kubenswrapper[4632]: multi_network_enabled_flag= Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ "true" == "true" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: multi_network_enabled_flag="--enable-multi-network" Mar 13 10:05:10 crc kubenswrapper[4632]: network_segmentation_enabled_flag="--enable-network-segmentation" Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: echo "I$(date "+%m%d %H:%M:%S.%N") - ovnkube-control-plane - start ovnkube --init-cluster-manager ${K8S_NODE}" Mar 13 10:05:10 crc kubenswrapper[4632]: exec /usr/bin/ovnkube \ Mar 13 10:05:10 crc kubenswrapper[4632]: --enable-interconnect \ Mar 13 10:05:10 crc kubenswrapper[4632]: --init-cluster-manager "${K8S_NODE}" \ Mar 13 10:05:10 crc kubenswrapper[4632]: --config-file=/run/ovnkube-config/ovnkube.conf \ Mar 13 10:05:10 crc kubenswrapper[4632]: --loglevel "${OVN_KUBE_LOG_LEVEL}" \ Mar 13 10:05:10 crc kubenswrapper[4632]: --metrics-bind-address "127.0.0.1:29108" \ Mar 13 10:05:10 crc kubenswrapper[4632]: --metrics-enable-pprof \ Mar 13 10:05:10 crc kubenswrapper[4632]: --metrics-enable-config-duration \ Mar 13 10:05:10 crc kubenswrapper[4632]: ${ovn_v4_join_subnet_opt} \ Mar 13 10:05:10 crc kubenswrapper[4632]: ${ovn_v6_join_subnet_opt} \ Mar 13 10:05:10 crc kubenswrapper[4632]: ${ovn_v4_transit_switch_subnet_opt} \ Mar 13 10:05:10 crc kubenswrapper[4632]: ${ovn_v6_transit_switch_subnet_opt} \ Mar 13 10:05:10 crc kubenswrapper[4632]: ${dns_name_resolver_enabled_flag} \ Mar 13 10:05:10 crc kubenswrapper[4632]: ${persistent_ips_enabled_flag} \ Mar 13 10:05:10 crc kubenswrapper[4632]: ${multi_network_enabled_flag} \ Mar 13 10:05:10 crc kubenswrapper[4632]: ${network_segmentation_enabled_flag} Mar 13 10:05:10 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics-port,HostPort:29108,ContainerPort:29108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OVN_KUBE_LOG_LEVEL,Value:4,ValueFrom:nil,},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{314572800 0} {} 300Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovnkube-config,ReadOnly:false,MountPath:/run/ovnkube-config/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ffbwk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-kbtt2_openshift-ovn-kubernetes(b0c542d5-8c38-4243-8af7-cfc0d8e22773): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.521188 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Mar 13 10:05:10 crc kubenswrapper[4632]: apiVersion: v1 Mar 13 10:05:10 crc kubenswrapper[4632]: clusters: Mar 13 10:05:10 crc kubenswrapper[4632]: - cluster: Mar 13 10:05:10 crc kubenswrapper[4632]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Mar 13 10:05:10 crc kubenswrapper[4632]: server: https://api-int.crc.testing:6443 Mar 13 10:05:10 crc kubenswrapper[4632]: name: default-cluster Mar 13 10:05:10 crc kubenswrapper[4632]: contexts: Mar 13 10:05:10 crc kubenswrapper[4632]: - context: Mar 13 10:05:10 crc kubenswrapper[4632]: cluster: default-cluster Mar 13 10:05:10 crc kubenswrapper[4632]: namespace: default Mar 13 10:05:10 crc kubenswrapper[4632]: user: default-auth Mar 13 10:05:10 crc kubenswrapper[4632]: name: default-context Mar 13 10:05:10 crc kubenswrapper[4632]: current-context: default-context Mar 13 10:05:10 crc kubenswrapper[4632]: kind: Config Mar 13 10:05:10 crc kubenswrapper[4632]: preferences: {} Mar 13 10:05:10 crc kubenswrapper[4632]: users: Mar 13 10:05:10 crc kubenswrapper[4632]: - name: default-auth Mar 13 10:05:10 crc kubenswrapper[4632]: user: Mar 13 10:05:10 crc kubenswrapper[4632]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 13 10:05:10 crc kubenswrapper[4632]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 13 10:05:10 crc kubenswrapper[4632]: EOF Mar 13 10:05:10 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dj6cl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.521502 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8b7cb438273be81daa37dee852b13428c2abbaceb49c8c8e2ae84fffcf8cf261"} Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.521813 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: container &Container{Name:kube-multus,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,Command:[/bin/bash -ec --],Args:[MULTUS_DAEMON_OPT="" Mar 13 10:05:10 crc kubenswrapper[4632]: /entrypoint/cnibincopy.sh; exec /usr/src/multus-cni/bin/multus-daemon $MULTUS_DAEMON_OPT Mar 13 10:05:10 crc kubenswrapper[4632]: ],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/multus-cni/bin/,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_PORT,Value:6443,ValueFrom:nil,},EnvVar{Name:KUBERNETES_SERVICE_HOST,Value:api-int.crc.testing,ValueFrom:nil,},EnvVar{Name:MULTUS_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:false,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:system-cni-dir,ReadOnly:false,MountPath:/host/etc/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-cni-dir,ReadOnly:false,MountPath:/host/run/multus/cni/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-socket-dir-parent,ReadOnly:false,MountPath:/host/run/multus,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-k8s-cni-cncf-io,ReadOnly:false,MountPath:/run/k8s.cni.cncf.io,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-netns,ReadOnly:false,MountPath:/run/netns,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-bin,ReadOnly:false,MountPath:/var/lib/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-cni-multus,ReadOnly:false,MountPath:/var/lib/cni/multus,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-var-lib-kubelet,ReadOnly:false,MountPath:/var/lib/kubelet,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:hostroot,ReadOnly:false,MountPath:/hostroot,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-conf-dir,ReadOnly:false,MountPath:/etc/cni/multus/net.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:multus-daemon-config,ReadOnly:true,MountPath:/etc/cni/net.d/multus.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-run-multus-certs,ReadOnly:false,MountPath:/etc/cni/multus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:etc-kubernetes,ReadOnly:false,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8d5c4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-gqf22_openshift-multus(4ec8e301-3037-4de0-94d2-32c49709660e): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.521986 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"ovnkube-cluster-manager\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" podUID="b0c542d5-8c38-4243-8af7-cfc0d8e22773" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.522604 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.522733 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:iptables-alerter,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,Command:[/iptables-alerter/iptables-alerter.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONTAINER_RUNTIME_ENDPOINT,Value:unix:///run/crio/crio.sock,ValueFrom:nil,},EnvVar{Name:ALERTER_POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{68157440 0} {} 65Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:iptables-alerter-script,ReadOnly:false,MountPath:/iptables-alerter,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host-slash,ReadOnly:true,MountPath:/host,SubPath:,MountPropagation:*HostToContainer,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rczfb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod iptables-alerter-4ln5h_openshift-network-operator(d75a4c96-2883-4a0b-bab2-0fab2b6c0b49): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.523813 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"iptables-alerter\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/iptables-alerter-4ln5h" podUID="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.523846 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-gqf22" podUID="4ec8e301-3037-4de0-94d2-32c49709660e" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.525747 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zwlc8" event={"ID":"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7","Type":"ContainerStarted","Data":"4c92132358db0c70647bacc66decb6a4c6c62b231a7ed9f697e887c3f84c7787"} Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.527554 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: container &Container{Name:node-ca,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f,Command:[/bin/sh -c trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM Mar 13 10:05:10 crc kubenswrapper[4632]: while [ true ]; Mar 13 10:05:10 crc kubenswrapper[4632]: do Mar 13 10:05:10 crc kubenswrapper[4632]: for f in $(ls /tmp/serviceca); do Mar 13 10:05:10 crc kubenswrapper[4632]: echo $f Mar 13 10:05:10 crc kubenswrapper[4632]: ca_file_path="/tmp/serviceca/${f}" Mar 13 10:05:10 crc kubenswrapper[4632]: f=$(echo $f | sed -r 's/(.*)\.\./\1:/') Mar 13 10:05:10 crc kubenswrapper[4632]: reg_dir_path="/etc/docker/certs.d/${f}" Mar 13 10:05:10 crc kubenswrapper[4632]: if [ -e "${reg_dir_path}" ]; then Mar 13 10:05:10 crc kubenswrapper[4632]: cp -u $ca_file_path $reg_dir_path/ca.crt Mar 13 10:05:10 crc kubenswrapper[4632]: else Mar 13 10:05:10 crc kubenswrapper[4632]: mkdir $reg_dir_path Mar 13 10:05:10 crc kubenswrapper[4632]: cp $ca_file_path $reg_dir_path/ca.crt Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: for d in $(ls /etc/docker/certs.d); do Mar 13 10:05:10 crc kubenswrapper[4632]: echo $d Mar 13 10:05:10 crc kubenswrapper[4632]: dp=$(echo $d | sed -r 's/(.*):/\1\.\./') Mar 13 10:05:10 crc kubenswrapper[4632]: reg_conf_path="/tmp/serviceca/${dp}" Mar 13 10:05:10 crc kubenswrapper[4632]: if [ ! -e "${reg_conf_path}" ]; then Mar 13 10:05:10 crc kubenswrapper[4632]: rm -rf /etc/docker/certs.d/$d Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: sleep 60 & wait ${!} Mar 13 10:05:10 crc kubenswrapper[4632]: done Mar 13 10:05:10 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{10485760 0} {} 10Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:serviceca,ReadOnly:false,MountPath:/tmp/serviceca,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:host,ReadOnly:false,MountPath:/etc/docker/certs.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mq5zl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*1001,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*0,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod node-ca-zwlc8_openshift-image-registry(9a50974e-f938-40f7-ace5-2a3b4cb1f3e7): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.528711 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"node-ca\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-image-registry/node-ca-zwlc8" podUID="9a50974e-f938-40f7-ace5-2a3b4cb1f3e7" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.529767 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" event={"ID":"b054ca08-1d09-4eca-a608-eb5b9323959a","Type":"ContainerStarted","Data":"b6d89645c2bf89bdd72981c912fdc942c220dbb780f6f99cfc8bc5a2bbbf55cf"} Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.530208 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.531030 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6f22aa92075ec9a6d25de41e0ea229738f486d80dffdd54030cd9441e1bc535a"} Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.533538 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: container &Container{Name:webhook,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ -f "/env/_master" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: set -o allexport Mar 13 10:05:10 crc kubenswrapper[4632]: source "/env/_master" Mar 13 10:05:10 crc kubenswrapper[4632]: set +o allexport Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: # OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled. Mar 13 10:05:10 crc kubenswrapper[4632]: # https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791 Mar 13 10:05:10 crc kubenswrapper[4632]: ho_enable="--enable-hybrid-overlay" Mar 13 10:05:10 crc kubenswrapper[4632]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start webhook" Mar 13 10:05:10 crc kubenswrapper[4632]: # extra-allowed-user: service account `ovn-kubernetes-control-plane` Mar 13 10:05:10 crc kubenswrapper[4632]: # sets pod annotations in multi-homing layer3 network controller (cluster-manager) Mar 13 10:05:10 crc kubenswrapper[4632]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 13 10:05:10 crc kubenswrapper[4632]: --webhook-cert-dir="/etc/webhook-cert" \ Mar 13 10:05:10 crc kubenswrapper[4632]: --webhook-host=127.0.0.1 \ Mar 13 10:05:10 crc kubenswrapper[4632]: --webhook-port=9743 \ Mar 13 10:05:10 crc kubenswrapper[4632]: ${ho_enable} \ Mar 13 10:05:10 crc kubenswrapper[4632]: --enable-interconnect \ Mar 13 10:05:10 crc kubenswrapper[4632]: --disable-approver \ Mar 13 10:05:10 crc kubenswrapper[4632]: --extra-allowed-user="system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane" \ Mar 13 10:05:10 crc kubenswrapper[4632]: --wait-for-kubernetes-api=200s \ Mar 13 10:05:10 crc kubenswrapper[4632]: --pod-admission-conditions="/var/run/ovnkube-identity-config/additional-pod-admission-cond.json" \ Mar 13 10:05:10 crc kubenswrapper[4632]: --loglevel="${LOGLEVEL}" Mar 13 10:05:10 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:2,ValueFrom:nil,},EnvVar{Name:KUBERNETES_NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:webhook-cert,ReadOnly:false,MountPath:/etc/webhook-cert/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.533841 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a73720ed844a5d744dc836e0fdd3b9ba936013cb457caf6f23070e9ade4d0cbf"} Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.534741 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:egress-router-binary-copy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,Command:[/entrypoint/cnibincopy.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:RHEL8_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel8/bin/,ValueFrom:nil,},EnvVar{Name:RHEL9_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/rhel9/bin/,ValueFrom:nil,},EnvVar{Name:DEFAULT_SOURCE_DIRECTORY,Value:/usr/src/egress-router-cni/bin/,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cni-binary-copy,ReadOnly:false,MountPath:/entrypoint,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:cnibin,ReadOnly:false,MountPath:/host/opt/cni/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:os-release,ReadOnly:true,MountPath:/host/etc/os-release,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l9s5s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod multus-additional-cni-plugins-qlc8m_openshift-multus(b054ca08-1d09-4eca-a608-eb5b9323959a): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars" logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.535839 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"egress-router-binary-copy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" podUID="b054ca08-1d09-4eca-a608-eb5b9323959a" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.535968 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: container &Container{Name:approver,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ -f "/env/_master" ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: set -o allexport Mar 13 10:05:10 crc kubenswrapper[4632]: source "/env/_master" Mar 13 10:05:10 crc kubenswrapper[4632]: set +o allexport Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: Mar 13 10:05:10 crc kubenswrapper[4632]: echo "I$(date "+%m%d %H:%M:%S.%N") - network-node-identity - start approver" Mar 13 10:05:10 crc kubenswrapper[4632]: exec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.crc.testing:6443 \ Mar 13 10:05:10 crc kubenswrapper[4632]: --disable-webhook \ Mar 13 10:05:10 crc kubenswrapper[4632]: --csr-acceptance-conditions="/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json" \ Mar 13 10:05:10 crc kubenswrapper[4632]: --loglevel="${LOGLEVEL}" Mar 13 10:05:10 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LOGLEVEL,Value:4,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovnkube-identity-cm,ReadOnly:false,MountPath:/var/run/ovnkube-identity-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s2kz5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000470000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-node-identity-vrzqb_openshift-network-node-identity(ef543e1b-8068-4ea3-b32a-61027b32e95d): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.536115 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:10 crc kubenswrapper[4632]: container &Container{Name:network-operator,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,Command:[/bin/bash -c #!/bin/bash Mar 13 10:05:10 crc kubenswrapper[4632]: set -o allexport Mar 13 10:05:10 crc kubenswrapper[4632]: if [[ -f /etc/kubernetes/apiserver-url.env ]]; then Mar 13 10:05:10 crc kubenswrapper[4632]: source /etc/kubernetes/apiserver-url.env Mar 13 10:05:10 crc kubenswrapper[4632]: else Mar 13 10:05:10 crc kubenswrapper[4632]: echo "Error: /etc/kubernetes/apiserver-url.env is missing" Mar 13 10:05:10 crc kubenswrapper[4632]: exit 1 Mar 13 10:05:10 crc kubenswrapper[4632]: fi Mar 13 10:05:10 crc kubenswrapper[4632]: exec /usr/bin/cluster-network-operator start --listen=0.0.0.0:9104 Mar 13 10:05:10 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:cno,HostPort:9104,ContainerPort:9104,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:RELEASE_VERSION,Value:4.18.1,ValueFrom:nil,},EnvVar{Name:KUBE_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b97554198294bf544fbc116c94a0a1fb2ec8a4de0e926bf9d9e320135f0bee6f,ValueFrom:nil,},EnvVar{Name:KUBE_RBAC_PROXY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,ValueFrom:nil,},EnvVar{Name:MULTUS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26,ValueFrom:nil,},EnvVar{Name:MULTUS_ADMISSION_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317,ValueFrom:nil,},EnvVar{Name:CNI_PLUGINS_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc,ValueFrom:nil,},EnvVar{Name:BOND_CNI_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78,ValueFrom:nil,},EnvVar{Name:WHEREABOUTS_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4,ValueFrom:nil,},EnvVar{Name:ROUTE_OVERRRIDE_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa,ValueFrom:nil,},EnvVar{Name:MULTUS_NETWORKPOLICY_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:23f833d3738d68706eb2f2868bd76bd71cee016cffa6faf5f045a60cc8c6eddd,ValueFrom:nil,},EnvVar{Name:OVN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,ValueFrom:nil,},EnvVar{Name:OVN_NB_RAFT_ELECTION_TIMER,Value:10,ValueFrom:nil,},EnvVar{Name:OVN_SB_RAFT_ELECTION_TIMER,Value:16,ValueFrom:nil,},EnvVar{Name:OVN_NORTHD_PROBE_INTERVAL,Value:10000,ValueFrom:nil,},EnvVar{Name:OVN_CONTROLLER_INACTIVITY_PROBE,Value:180000,ValueFrom:nil,},EnvVar{Name:OVN_NB_INACTIVITY_PROBE,Value:60000,ValueFrom:nil,},EnvVar{Name:EGRESS_ROUTER_CNI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c,ValueFrom:nil,},EnvVar{Name:NETWORK_METRICS_DAEMON_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_SOURCE_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_CHECK_TARGET_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:NETWORK_OPERATOR_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b,ValueFrom:nil,},EnvVar{Name:CLOUD_NETWORK_CONFIG_CONTROLLER_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8048f1cb0be521f09749c0a489503cd56d85b68c6ca93380e082cfd693cd97a8,ValueFrom:nil,},EnvVar{Name:CLI_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2,ValueFrom:nil,},EnvVar{Name:FRR_K8S_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5dbf844e49bb46b78586930149e5e5f5dc121014c8afd10fe36f3651967cc256,ValueFrom:nil,},EnvVar{Name:NETWORKING_CONSOLE_PLUGIN_IMAGE,Value:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd,ValueFrom:nil,},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host-etc-kube,ReadOnly:true,MountPath:/etc/kubernetes,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:metrics-tls,ReadOnly:false,MountPath:/var/run/secrets/serving-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rdwmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod network-operator-58b4c7f79c-55gtf_openshift-network-operator(37a5e44f-9a88-4405-be8a-b645485e7312): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:10 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.537678 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"webhook\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"approver\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-network-node-identity/network-node-identity-vrzqb" podUID="ef543e1b-8068-4ea3-b32a-61027b32e95d" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.537721 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"network-operator\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" podUID="37a5e44f-9a88-4405-be8a-b645485e7312" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.542523 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.544428 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.544457 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.544515 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.544539 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.544704 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:10Z","lastTransitionTime":"2026-03-13T10:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.551613 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.566545 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.578186 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.586099 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.594639 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.604703 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.616641 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.625626 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.634756 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.644396 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.647155 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.647179 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.647212 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.647232 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.647243 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:10Z","lastTransitionTime":"2026-03-13T10:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.655302 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.667390 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.678131 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.687150 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.695713 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.705758 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.716810 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.725505 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.734470 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.743557 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.749871 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.749952 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.749972 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.749987 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.749996 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:10Z","lastTransitionTime":"2026-03-13T10:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.751178 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.757579 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.787216 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.809096 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.809260 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:05:11.809238229 +0000 UTC m=+85.831768362 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.809378 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.809540 4632 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.809609 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.809769 4632 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.809782 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:11.809764925 +0000 UTC m=+85.832295058 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.809894 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:11.809871859 +0000 UTC m=+85.832402012 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.820209 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.853561 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.853623 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.853636 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.853662 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.853677 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:10Z","lastTransitionTime":"2026-03-13T10:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.859696 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.911020 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.911099 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.911143 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.911314 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.911384 4632 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.911474 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs podName:ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad nodeName:}" failed. No retries permitted until 2026-03-13 10:05:11.911449812 +0000 UTC m=+85.933979985 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs") pod "network-metrics-daemon-z2vlz" (UID: "ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.911389 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.911528 4632 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.911505 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.911615 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.911625 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:11.911596106 +0000 UTC m=+85.934126399 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.911637 4632 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:10 crc kubenswrapper[4632]: E0313 10:05:10.911731 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:11.911703559 +0000 UTC m=+85.934233882 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.958079 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.958153 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.958167 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.958195 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:10 crc kubenswrapper[4632]: I0313 10:05:10.958211 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:10Z","lastTransitionTime":"2026-03-13T10:05:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.065424 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.065798 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.065880 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.065981 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.066075 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:11Z","lastTransitionTime":"2026-03-13T10:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.066157 4632 scope.go:117] "RemoveContainer" containerID="9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae" Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.067493 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.067748 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.169164 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.169225 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.169240 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.169260 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.169273 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:11Z","lastTransitionTime":"2026-03-13T10:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.271956 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.272016 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.272031 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.272057 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.272072 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:11Z","lastTransitionTime":"2026-03-13T10:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.375027 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.375081 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.375096 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.375115 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.375129 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:11Z","lastTransitionTime":"2026-03-13T10:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.478231 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.478305 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.478322 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.478343 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.478358 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:11Z","lastTransitionTime":"2026-03-13T10:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.537824 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" event={"ID":"b0c542d5-8c38-4243-8af7-cfc0d8e22773","Type":"ContainerStarted","Data":"475af06ee8970aef910c45ce81dbd4a1179474c0a41501893d0b8f7aa65229e2"} Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.539476 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:11 crc kubenswrapper[4632]: container &Container{Name:kube-rbac-proxy,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09,Command:[/bin/bash -c #!/bin/bash Mar 13 10:05:11 crc kubenswrapper[4632]: set -euo pipefail Mar 13 10:05:11 crc kubenswrapper[4632]: TLS_PK=/etc/pki/tls/metrics-cert/tls.key Mar 13 10:05:11 crc kubenswrapper[4632]: TLS_CERT=/etc/pki/tls/metrics-cert/tls.crt Mar 13 10:05:11 crc kubenswrapper[4632]: # As the secret mount is optional we must wait for the files to be present. Mar 13 10:05:11 crc kubenswrapper[4632]: # The service is created in monitor.yaml and this is created in sdn.yaml. Mar 13 10:05:11 crc kubenswrapper[4632]: TS=$(date +%s) Mar 13 10:05:11 crc kubenswrapper[4632]: WARN_TS=$(( ${TS} + $(( 20 * 60)) )) Mar 13 10:05:11 crc kubenswrapper[4632]: HAS_LOGGED_INFO=0 Mar 13 10:05:11 crc kubenswrapper[4632]: Mar 13 10:05:11 crc kubenswrapper[4632]: log_missing_certs(){ Mar 13 10:05:11 crc kubenswrapper[4632]: CUR_TS=$(date +%s) Mar 13 10:05:11 crc kubenswrapper[4632]: if [[ "${CUR_TS}" -gt "WARN_TS" ]]; then Mar 13 10:05:11 crc kubenswrapper[4632]: echo $(date -Iseconds) WARN: ovn-control-plane-metrics-cert not mounted after 20 minutes. Mar 13 10:05:11 crc kubenswrapper[4632]: elif [[ "${HAS_LOGGED_INFO}" -eq 0 ]] ; then Mar 13 10:05:11 crc kubenswrapper[4632]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-cert not mounted. Waiting 20 minutes. Mar 13 10:05:11 crc kubenswrapper[4632]: HAS_LOGGED_INFO=1 Mar 13 10:05:11 crc kubenswrapper[4632]: fi Mar 13 10:05:11 crc kubenswrapper[4632]: } Mar 13 10:05:11 crc kubenswrapper[4632]: while [[ ! -f "${TLS_PK}" || ! -f "${TLS_CERT}" ]] ; do Mar 13 10:05:11 crc kubenswrapper[4632]: log_missing_certs Mar 13 10:05:11 crc kubenswrapper[4632]: sleep 5 Mar 13 10:05:11 crc kubenswrapper[4632]: done Mar 13 10:05:11 crc kubenswrapper[4632]: Mar 13 10:05:11 crc kubenswrapper[4632]: echo $(date -Iseconds) INFO: ovn-control-plane-metrics-certs mounted, starting kube-rbac-proxy Mar 13 10:05:11 crc kubenswrapper[4632]: exec /usr/bin/kube-rbac-proxy \ Mar 13 10:05:11 crc kubenswrapper[4632]: --logtostderr \ Mar 13 10:05:11 crc kubenswrapper[4632]: --secure-listen-address=:9108 \ Mar 13 10:05:11 crc kubenswrapper[4632]: --tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256 \ Mar 13 10:05:11 crc kubenswrapper[4632]: --upstream=http://127.0.0.1:29108/ \ Mar 13 10:05:11 crc kubenswrapper[4632]: --tls-private-key-file=${TLS_PK} \ Mar 13 10:05:11 crc kubenswrapper[4632]: --tls-cert-file=${TLS_CERT} Mar 13 10:05:11 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:9108,ContainerPort:9108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{20971520 0} {} 20Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovn-control-plane-metrics-cert,ReadOnly:true,MountPath:/etc/pki/tls/metrics-cert,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ffbwk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-kbtt2_openshift-ovn-kubernetes(b0c542d5-8c38-4243-8af7-cfc0d8e22773): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:11 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.539838 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerStarted","Data":"5f99589c0e329dc2bea211f1582fe2ff509c48ed7460521bac851a5b63796f30"} Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.540154 4632 scope.go:117] "RemoveContainer" containerID="9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae" Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.540323 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.543181 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:11 crc kubenswrapper[4632]: init container &Container{Name:kubecfg-setup,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c cat << EOF > /etc/ovn/kubeconfig Mar 13 10:05:11 crc kubenswrapper[4632]: apiVersion: v1 Mar 13 10:05:11 crc kubenswrapper[4632]: clusters: Mar 13 10:05:11 crc kubenswrapper[4632]: - cluster: Mar 13 10:05:11 crc kubenswrapper[4632]: certificate-authority: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt Mar 13 10:05:11 crc kubenswrapper[4632]: server: https://api-int.crc.testing:6443 Mar 13 10:05:11 crc kubenswrapper[4632]: name: default-cluster Mar 13 10:05:11 crc kubenswrapper[4632]: contexts: Mar 13 10:05:11 crc kubenswrapper[4632]: - context: Mar 13 10:05:11 crc kubenswrapper[4632]: cluster: default-cluster Mar 13 10:05:11 crc kubenswrapper[4632]: namespace: default Mar 13 10:05:11 crc kubenswrapper[4632]: user: default-auth Mar 13 10:05:11 crc kubenswrapper[4632]: name: default-context Mar 13 10:05:11 crc kubenswrapper[4632]: current-context: default-context Mar 13 10:05:11 crc kubenswrapper[4632]: kind: Config Mar 13 10:05:11 crc kubenswrapper[4632]: preferences: {} Mar 13 10:05:11 crc kubenswrapper[4632]: users: Mar 13 10:05:11 crc kubenswrapper[4632]: - name: default-auth Mar 13 10:05:11 crc kubenswrapper[4632]: user: Mar 13 10:05:11 crc kubenswrapper[4632]: client-certificate: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 13 10:05:11 crc kubenswrapper[4632]: client-key: /etc/ovn/ovnkube-node-certs/ovnkube-client-current.pem Mar 13 10:05:11 crc kubenswrapper[4632]: EOF Mar 13 10:05:11 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-openvswitch,ReadOnly:false,MountPath:/etc/ovn/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dj6cl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:11 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.543275 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:05:11 crc kubenswrapper[4632]: container &Container{Name:ovnkube-cluster-manager,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2,Command:[/bin/bash -c set -xe Mar 13 10:05:11 crc kubenswrapper[4632]: if [[ -f "/env/_master" ]]; then Mar 13 10:05:11 crc kubenswrapper[4632]: set -o allexport Mar 13 10:05:11 crc kubenswrapper[4632]: source "/env/_master" Mar 13 10:05:11 crc kubenswrapper[4632]: set +o allexport Mar 13 10:05:11 crc kubenswrapper[4632]: fi Mar 13 10:05:11 crc kubenswrapper[4632]: Mar 13 10:05:11 crc kubenswrapper[4632]: ovn_v4_join_subnet_opt= Mar 13 10:05:11 crc kubenswrapper[4632]: if [[ "" != "" ]]; then Mar 13 10:05:11 crc kubenswrapper[4632]: ovn_v4_join_subnet_opt="--gateway-v4-join-subnet " Mar 13 10:05:11 crc kubenswrapper[4632]: fi Mar 13 10:05:11 crc kubenswrapper[4632]: ovn_v6_join_subnet_opt= Mar 13 10:05:11 crc kubenswrapper[4632]: if [[ "" != "" ]]; then Mar 13 10:05:11 crc kubenswrapper[4632]: ovn_v6_join_subnet_opt="--gateway-v6-join-subnet " Mar 13 10:05:11 crc kubenswrapper[4632]: fi Mar 13 10:05:11 crc kubenswrapper[4632]: Mar 13 10:05:11 crc kubenswrapper[4632]: ovn_v4_transit_switch_subnet_opt= Mar 13 10:05:11 crc kubenswrapper[4632]: if [[ "" != "" ]]; then Mar 13 10:05:11 crc kubenswrapper[4632]: ovn_v4_transit_switch_subnet_opt="--cluster-manager-v4-transit-switch-subnet " Mar 13 10:05:11 crc kubenswrapper[4632]: fi Mar 13 10:05:11 crc kubenswrapper[4632]: ovn_v6_transit_switch_subnet_opt= Mar 13 10:05:11 crc kubenswrapper[4632]: if [[ "" != "" ]]; then Mar 13 10:05:11 crc kubenswrapper[4632]: ovn_v6_transit_switch_subnet_opt="--cluster-manager-v6-transit-switch-subnet " Mar 13 10:05:11 crc kubenswrapper[4632]: fi Mar 13 10:05:11 crc kubenswrapper[4632]: Mar 13 10:05:11 crc kubenswrapper[4632]: dns_name_resolver_enabled_flag= Mar 13 10:05:11 crc kubenswrapper[4632]: if [[ "false" == "true" ]]; then Mar 13 10:05:11 crc kubenswrapper[4632]: dns_name_resolver_enabled_flag="--enable-dns-name-resolver" Mar 13 10:05:11 crc kubenswrapper[4632]: fi Mar 13 10:05:11 crc kubenswrapper[4632]: Mar 13 10:05:11 crc kubenswrapper[4632]: persistent_ips_enabled_flag= Mar 13 10:05:11 crc kubenswrapper[4632]: if [[ "true" == "true" ]]; then Mar 13 10:05:11 crc kubenswrapper[4632]: persistent_ips_enabled_flag="--enable-persistent-ips" Mar 13 10:05:11 crc kubenswrapper[4632]: fi Mar 13 10:05:11 crc kubenswrapper[4632]: Mar 13 10:05:11 crc kubenswrapper[4632]: # This is needed so that converting clusters from GA to TP Mar 13 10:05:11 crc kubenswrapper[4632]: # will rollout control plane pods as well Mar 13 10:05:11 crc kubenswrapper[4632]: network_segmentation_enabled_flag= Mar 13 10:05:11 crc kubenswrapper[4632]: multi_network_enabled_flag= Mar 13 10:05:11 crc kubenswrapper[4632]: if [[ "true" == "true" ]]; then Mar 13 10:05:11 crc kubenswrapper[4632]: multi_network_enabled_flag="--enable-multi-network" Mar 13 10:05:11 crc kubenswrapper[4632]: network_segmentation_enabled_flag="--enable-network-segmentation" Mar 13 10:05:11 crc kubenswrapper[4632]: fi Mar 13 10:05:11 crc kubenswrapper[4632]: Mar 13 10:05:11 crc kubenswrapper[4632]: echo "I$(date "+%m%d %H:%M:%S.%N") - ovnkube-control-plane - start ovnkube --init-cluster-manager ${K8S_NODE}" Mar 13 10:05:11 crc kubenswrapper[4632]: exec /usr/bin/ovnkube \ Mar 13 10:05:11 crc kubenswrapper[4632]: --enable-interconnect \ Mar 13 10:05:11 crc kubenswrapper[4632]: --init-cluster-manager "${K8S_NODE}" \ Mar 13 10:05:11 crc kubenswrapper[4632]: --config-file=/run/ovnkube-config/ovnkube.conf \ Mar 13 10:05:11 crc kubenswrapper[4632]: --loglevel "${OVN_KUBE_LOG_LEVEL}" \ Mar 13 10:05:11 crc kubenswrapper[4632]: --metrics-bind-address "127.0.0.1:29108" \ Mar 13 10:05:11 crc kubenswrapper[4632]: --metrics-enable-pprof \ Mar 13 10:05:11 crc kubenswrapper[4632]: --metrics-enable-config-duration \ Mar 13 10:05:11 crc kubenswrapper[4632]: ${ovn_v4_join_subnet_opt} \ Mar 13 10:05:11 crc kubenswrapper[4632]: ${ovn_v6_join_subnet_opt} \ Mar 13 10:05:11 crc kubenswrapper[4632]: ${ovn_v4_transit_switch_subnet_opt} \ Mar 13 10:05:11 crc kubenswrapper[4632]: ${ovn_v6_transit_switch_subnet_opt} \ Mar 13 10:05:11 crc kubenswrapper[4632]: ${dns_name_resolver_enabled_flag} \ Mar 13 10:05:11 crc kubenswrapper[4632]: ${persistent_ips_enabled_flag} \ Mar 13 10:05:11 crc kubenswrapper[4632]: ${multi_network_enabled_flag} \ Mar 13 10:05:11 crc kubenswrapper[4632]: ${network_segmentation_enabled_flag} Mar 13 10:05:11 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics-port,HostPort:29108,ContainerPort:29108,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OVN_KUBE_LOG_LEVEL,Value:4,ValueFrom:nil,},EnvVar{Name:K8S_NODE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{314572800 0} {} 300Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:ovnkube-config,ReadOnly:false,MountPath:/run/ovnkube-config/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:env-overrides,ReadOnly:false,MountPath:/env,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-ffbwk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovnkube-control-plane-749d76644c-kbtt2_openshift-ovn-kubernetes(b0c542d5-8c38-4243-8af7-cfc0d8e22773): CreateContainerConfigError: services have not yet been read at least once, cannot construct envvars Mar 13 10:05:11 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.544337 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kubecfg-setup\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.544507 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"kube-rbac-proxy\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\", failed to \"StartContainer\" for \"ovnkube-cluster-manager\" with CreateContainerConfigError: \"services have not yet been read at least once, cannot construct envvars\"]" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" podUID="b0c542d5-8c38-4243-8af7-cfc0d8e22773" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.552711 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.564793 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.576730 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.580365 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.580401 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.580410 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.580426 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.580435 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:11Z","lastTransitionTime":"2026-03-13T10:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.594118 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.610523 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.622289 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.636535 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.650235 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.663657 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.674518 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.685304 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.685346 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.685355 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.685374 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.685383 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:11Z","lastTransitionTime":"2026-03-13T10:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.686622 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.696892 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.707373 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.716191 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.733604 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.746037 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.756904 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.769302 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.782431 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.788326 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.788381 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.788402 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.788420 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.788431 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:11Z","lastTransitionTime":"2026-03-13T10:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.796455 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.807071 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.820059 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.821362 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.821671 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:05:13.821633341 +0000 UTC m=+87.844163594 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.821767 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.821868 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.822010 4632 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.822070 4632 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.822100 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:13.822079535 +0000 UTC m=+87.844609668 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.822233 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:13.822210009 +0000 UTC m=+87.844740272 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.833651 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.844880 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.876417 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.893396 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.893457 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.893470 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.893494 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.893511 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:11Z","lastTransitionTime":"2026-03-13T10:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.918807 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.923504 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.923626 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.923749 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.923695 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.923832 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.923851 4632 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.923717 4632 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.923907 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:13.923889345 +0000 UTC m=+87.946419478 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.923961 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs podName:ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad nodeName:}" failed. No retries permitted until 2026-03-13 10:05:13.923933066 +0000 UTC m=+87.946463209 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs") pod "network-metrics-daemon-z2vlz" (UID: "ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.923987 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.924007 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.924018 4632 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:11 crc kubenswrapper[4632]: E0313 10:05:11.924082 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:13.92407047 +0000 UTC m=+87.946600603 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.960341 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.998240 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.998428 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.998464 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.998477 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.998499 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:11 crc kubenswrapper[4632]: I0313 10:05:11.998514 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:11Z","lastTransitionTime":"2026-03-13T10:05:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.038230 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.043646 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.043787 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.043833 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:12 crc kubenswrapper[4632]: E0313 10:05:12.043897 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.043921 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:12 crc kubenswrapper[4632]: E0313 10:05:12.044089 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:12 crc kubenswrapper[4632]: E0313 10:05:12.044221 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:12 crc kubenswrapper[4632]: E0313 10:05:12.044304 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.050740 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.051491 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.052458 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.053238 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.054092 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.054713 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.055397 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.055980 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.056683 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.057238 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.060217 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.061012 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.061970 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.062538 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.063113 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.064112 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.064728 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.065688 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.066301 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.066873 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.067806 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.068425 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.068888 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.070218 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.070700 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.071780 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.072478 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.073407 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.074142 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.075026 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.075522 4632 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.075637 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.077728 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.078395 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.078797 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.080351 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.081366 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.082076 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.082092 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.083291 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.083960 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.084794 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.085472 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.086552 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.087602 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.088167 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.088850 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.090133 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.090969 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.091808 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.092313 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.093258 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.093756 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.094402 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.095277 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.100598 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.100699 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.100713 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.100736 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.100752 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:12Z","lastTransitionTime":"2026-03-13T10:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.203925 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.204059 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.204074 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.204098 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.204114 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:12Z","lastTransitionTime":"2026-03-13T10:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.312468 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.312819 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.313586 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.313628 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.313641 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:12Z","lastTransitionTime":"2026-03-13T10:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.416724 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.416801 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.416814 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.416837 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.416854 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:12Z","lastTransitionTime":"2026-03-13T10:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.518919 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.518999 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.519013 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.519030 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.519048 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:12Z","lastTransitionTime":"2026-03-13T10:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.621764 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.621808 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.621817 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.621831 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.621839 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:12Z","lastTransitionTime":"2026-03-13T10:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.725320 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.725370 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.725381 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.725399 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.725410 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:12Z","lastTransitionTime":"2026-03-13T10:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.828623 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.828667 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.828714 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.828733 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.828742 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:12Z","lastTransitionTime":"2026-03-13T10:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.932086 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.932186 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.932205 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.932223 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:12 crc kubenswrapper[4632]: I0313 10:05:12.932233 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:12Z","lastTransitionTime":"2026-03-13T10:05:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.034853 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.034912 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.034928 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.034994 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.035017 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:13Z","lastTransitionTime":"2026-03-13T10:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.137806 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.138016 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.138038 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.138072 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.138090 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:13Z","lastTransitionTime":"2026-03-13T10:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.239877 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.239955 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.239973 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.239989 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.240001 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:13Z","lastTransitionTime":"2026-03-13T10:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.342715 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.342755 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.342773 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.342791 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.342801 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:13Z","lastTransitionTime":"2026-03-13T10:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.445623 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.445698 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.445714 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.445731 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.445750 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:13Z","lastTransitionTime":"2026-03-13T10:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.547739 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.547802 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.547818 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.547840 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.547855 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:13Z","lastTransitionTime":"2026-03-13T10:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.652243 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.652339 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.652353 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.652375 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.652389 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:13Z","lastTransitionTime":"2026-03-13T10:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.754894 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.754965 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.754979 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.754997 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.755010 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:13Z","lastTransitionTime":"2026-03-13T10:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.841713 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.841817 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.841876 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.841978 4632 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.842011 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:05:17.841969098 +0000 UTC m=+91.864499241 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.842069 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:17.84205792 +0000 UTC m=+91.864588103 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.842079 4632 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.842168 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:17.842147643 +0000 UTC m=+91.864677776 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.856766 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.856802 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.856813 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.856826 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.856835 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:13Z","lastTransitionTime":"2026-03-13T10:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.943454 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.943753 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.943911 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.943929 4632 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.943856 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.944030 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:17.944004913 +0000 UTC m=+91.966535056 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.944093 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.944235 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.944260 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.944272 4632 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.944325 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:17.944308564 +0000 UTC m=+91.966838797 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.944523 4632 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:05:13 crc kubenswrapper[4632]: E0313 10:05:13.944669 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs podName:ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad nodeName:}" failed. No retries permitted until 2026-03-13 10:05:17.944654704 +0000 UTC m=+91.967184927 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs") pod "network-metrics-daemon-z2vlz" (UID: "ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.959498 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.959824 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.959930 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.960081 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:13 crc kubenswrapper[4632]: I0313 10:05:13.960392 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:13Z","lastTransitionTime":"2026-03-13T10:05:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.043453 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.043460 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:14 crc kubenswrapper[4632]: E0313 10:05:14.043589 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.043474 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:14 crc kubenswrapper[4632]: E0313 10:05:14.043716 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:14 crc kubenswrapper[4632]: E0313 10:05:14.043761 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.044242 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:14 crc kubenswrapper[4632]: E0313 10:05:14.044417 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.064487 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.064544 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.064559 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.064580 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.064593 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:14Z","lastTransitionTime":"2026-03-13T10:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.167729 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.167785 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.167798 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.167816 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.167830 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:14Z","lastTransitionTime":"2026-03-13T10:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.271083 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.271137 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.271149 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.271168 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.271180 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:14Z","lastTransitionTime":"2026-03-13T10:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.374868 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.374927 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.374969 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.374988 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.375001 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:14Z","lastTransitionTime":"2026-03-13T10:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.477647 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.477733 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.477744 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.477765 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.477776 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:14Z","lastTransitionTime":"2026-03-13T10:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.580300 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.580621 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.580747 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.580868 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.581082 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:14Z","lastTransitionTime":"2026-03-13T10:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.684872 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.685320 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.685435 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.685549 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.685658 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:14Z","lastTransitionTime":"2026-03-13T10:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.789023 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.789068 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.789083 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.789103 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.789119 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:14Z","lastTransitionTime":"2026-03-13T10:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.891698 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.891726 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.891735 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.891749 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.891758 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:14Z","lastTransitionTime":"2026-03-13T10:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.994685 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.995030 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.995110 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.995193 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:14 crc kubenswrapper[4632]: I0313 10:05:14.995254 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:14Z","lastTransitionTime":"2026-03-13T10:05:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.097967 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.098245 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.098325 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.098428 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.098537 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:15Z","lastTransitionTime":"2026-03-13T10:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.201610 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.201650 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.201662 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.201682 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.201694 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:15Z","lastTransitionTime":"2026-03-13T10:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.303650 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.303698 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.303708 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.303720 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.303729 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:15Z","lastTransitionTime":"2026-03-13T10:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.406644 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.406682 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.406691 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.406707 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.406718 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:15Z","lastTransitionTime":"2026-03-13T10:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.509925 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.510003 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.510017 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.510033 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.510044 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:15Z","lastTransitionTime":"2026-03-13T10:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.613825 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.614190 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.614330 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.614435 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.614524 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:15Z","lastTransitionTime":"2026-03-13T10:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.718579 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.718630 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.718643 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.718665 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.718676 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:15Z","lastTransitionTime":"2026-03-13T10:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.835339 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.835404 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.835416 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.835438 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.835451 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:15Z","lastTransitionTime":"2026-03-13T10:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.938914 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.939167 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.939232 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.939296 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:15 crc kubenswrapper[4632]: I0313 10:05:15.939355 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:15Z","lastTransitionTime":"2026-03-13T10:05:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.041617 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.041656 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.041667 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.041684 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.041694 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:16Z","lastTransitionTime":"2026-03-13T10:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.044112 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.044128 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.044150 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:16 crc kubenswrapper[4632]: E0313 10:05:16.044202 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.044112 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:16 crc kubenswrapper[4632]: E0313 10:05:16.044288 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:16 crc kubenswrapper[4632]: E0313 10:05:16.044344 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:16 crc kubenswrapper[4632]: E0313 10:05:16.044390 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.267928 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.267997 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.268010 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.268028 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.268041 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:16Z","lastTransitionTime":"2026-03-13T10:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.301037 4632 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.373762 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.373826 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.373840 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.373858 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.373872 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:16Z","lastTransitionTime":"2026-03-13T10:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.476209 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.476244 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.476252 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.476265 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.476277 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:16Z","lastTransitionTime":"2026-03-13T10:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.578907 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.578968 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.578984 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.579001 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.579010 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:16Z","lastTransitionTime":"2026-03-13T10:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.682301 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.682351 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.682360 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.682375 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.682386 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:16Z","lastTransitionTime":"2026-03-13T10:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.784422 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.784458 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.784467 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.784479 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.784488 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:16Z","lastTransitionTime":"2026-03-13T10:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.887345 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.887375 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.887383 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.887397 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.887407 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:16Z","lastTransitionTime":"2026-03-13T10:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.989687 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.989723 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.989732 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.989745 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:16 crc kubenswrapper[4632]: I0313 10:05:16.989754 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:16Z","lastTransitionTime":"2026-03-13T10:05:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.056695 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.093175 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.093249 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.093265 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.093282 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.093325 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:17Z","lastTransitionTime":"2026-03-13T10:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.196712 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.196757 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.196767 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.196786 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.196799 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:17Z","lastTransitionTime":"2026-03-13T10:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.299857 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.300283 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.300395 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.300528 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.300614 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:17Z","lastTransitionTime":"2026-03-13T10:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.403218 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.403273 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.403283 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.403300 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.403309 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:17Z","lastTransitionTime":"2026-03-13T10:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.505594 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.505634 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.505643 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.505657 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.505666 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:17Z","lastTransitionTime":"2026-03-13T10:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.608932 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.608994 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.609006 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.609025 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.609038 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:17Z","lastTransitionTime":"2026-03-13T10:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.712093 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.712140 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.712151 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.712168 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.712180 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:17Z","lastTransitionTime":"2026-03-13T10:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.814579 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.814631 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.814640 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.814659 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.814668 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:17Z","lastTransitionTime":"2026-03-13T10:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.880113 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.880224 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.880283 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.880365 4632 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.880428 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:25.880411291 +0000 UTC m=+99.902941424 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.880490 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:05:25.880482494 +0000 UTC m=+99.903012627 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.880568 4632 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.880607 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:25.880599637 +0000 UTC m=+99.903129770 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.917597 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.917928 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.918055 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.918125 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.918191 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:17Z","lastTransitionTime":"2026-03-13T10:05:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.981665 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.981711 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:17 crc kubenswrapper[4632]: I0313 10:05:17.981730 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.982023 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.982090 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.982107 4632 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.982184 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:25.98216139 +0000 UTC m=+100.004691593 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.982202 4632 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.982300 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs podName:ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad nodeName:}" failed. No retries permitted until 2026-03-13 10:05:25.982282533 +0000 UTC m=+100.004812666 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs") pod "network-metrics-daemon-z2vlz" (UID: "ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.982806 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.982834 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.982849 4632 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:17 crc kubenswrapper[4632]: E0313 10:05:17.982890 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:25.982879331 +0000 UTC m=+100.005409464 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.020566 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.020605 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.020618 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.020637 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.020649 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:18Z","lastTransitionTime":"2026-03-13T10:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.044045 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.044092 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.044068 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:18 crc kubenswrapper[4632]: E0313 10:05:18.044175 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.044045 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:18 crc kubenswrapper[4632]: E0313 10:05:18.044302 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:18 crc kubenswrapper[4632]: E0313 10:05:18.044370 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:18 crc kubenswrapper[4632]: E0313 10:05:18.044431 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.056477 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.065039 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.073550 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.081043 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.090584 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.099747 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.106726 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.112353 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.124318 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.124358 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.124369 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.124397 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.124410 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:18Z","lastTransitionTime":"2026-03-13T10:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.126817 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.134819 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.150792 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.161604 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.174050 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.188418 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.198968 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.209883 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.226783 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.226834 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.226854 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.226878 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.226895 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:18Z","lastTransitionTime":"2026-03-13T10:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.326815 4632 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.329829 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.329878 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.329893 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.329913 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.329925 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:18Z","lastTransitionTime":"2026-03-13T10:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.387653 4632 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.433265 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.433309 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.433321 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.433338 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.433358 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:18Z","lastTransitionTime":"2026-03-13T10:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.536099 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.536150 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.536167 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.536185 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.536200 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:18Z","lastTransitionTime":"2026-03-13T10:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.639814 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.639855 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.639865 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.639885 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.639895 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:18Z","lastTransitionTime":"2026-03-13T10:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.742623 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.742684 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.742703 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.742724 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.742735 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:18Z","lastTransitionTime":"2026-03-13T10:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.851127 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.851164 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.851176 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.851193 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.851205 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:18Z","lastTransitionTime":"2026-03-13T10:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.954314 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.954361 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.954372 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.954388 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:18 crc kubenswrapper[4632]: I0313 10:05:18.954400 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:18Z","lastTransitionTime":"2026-03-13T10:05:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.056907 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.056974 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.056987 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.057003 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.057018 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:19Z","lastTransitionTime":"2026-03-13T10:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.159117 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.159149 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.159161 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.159177 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.159188 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:19Z","lastTransitionTime":"2026-03-13T10:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.262350 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.262629 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.262759 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.262848 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.262966 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:19Z","lastTransitionTime":"2026-03-13T10:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.365610 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.365845 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.365964 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.366091 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.366160 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:19Z","lastTransitionTime":"2026-03-13T10:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.468309 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.468982 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.469062 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.469127 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.469264 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:19Z","lastTransitionTime":"2026-03-13T10:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.571576 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.571615 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.571626 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.571641 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.571653 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:19Z","lastTransitionTime":"2026-03-13T10:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.674524 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.674815 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.675063 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.675239 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.675312 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:19Z","lastTransitionTime":"2026-03-13T10:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.777656 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.777708 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.777722 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.777745 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.777763 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:19Z","lastTransitionTime":"2026-03-13T10:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.881075 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.881458 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.881553 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.881643 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.881719 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:19Z","lastTransitionTime":"2026-03-13T10:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.984971 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.985017 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.985028 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.985043 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:19 crc kubenswrapper[4632]: I0313 10:05:19.985054 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:19Z","lastTransitionTime":"2026-03-13T10:05:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.043995 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:20 crc kubenswrapper[4632]: E0313 10:05:20.044180 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.044200 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.044464 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:20 crc kubenswrapper[4632]: E0313 10:05:20.044493 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.044634 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:20 crc kubenswrapper[4632]: E0313 10:05:20.044695 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:20 crc kubenswrapper[4632]: E0313 10:05:20.044884 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.087647 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.087694 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.087715 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.087731 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.087741 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:20Z","lastTransitionTime":"2026-03-13T10:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.191521 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.191618 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.191637 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.191661 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.191678 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:20Z","lastTransitionTime":"2026-03-13T10:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.294328 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.294371 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.294381 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.294397 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.294407 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:20Z","lastTransitionTime":"2026-03-13T10:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.343026 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.343368 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.343461 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.343573 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.343667 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:20Z","lastTransitionTime":"2026-03-13T10:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:20 crc kubenswrapper[4632]: E0313 10:05:20.355358 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.360552 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.360816 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.360882 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.360963 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.361037 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:20Z","lastTransitionTime":"2026-03-13T10:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:20 crc kubenswrapper[4632]: E0313 10:05:20.374795 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.379321 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.379381 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.379586 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.379613 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.379631 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:20Z","lastTransitionTime":"2026-03-13T10:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:20 crc kubenswrapper[4632]: E0313 10:05:20.392073 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.396647 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.396709 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.396719 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.396741 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.396755 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:20Z","lastTransitionTime":"2026-03-13T10:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:20 crc kubenswrapper[4632]: E0313 10:05:20.409181 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.415479 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.415594 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.415613 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.415641 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.415678 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:20Z","lastTransitionTime":"2026-03-13T10:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:20 crc kubenswrapper[4632]: E0313 10:05:20.429728 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:20Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:20 crc kubenswrapper[4632]: E0313 10:05:20.429867 4632 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.432459 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.432520 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.432533 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.432550 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.432561 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:20Z","lastTransitionTime":"2026-03-13T10:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.536282 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.536331 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.536341 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.536360 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.536372 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:20Z","lastTransitionTime":"2026-03-13T10:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.639620 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.639996 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.640299 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.640462 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.640601 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:20Z","lastTransitionTime":"2026-03-13T10:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.742864 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.743167 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.743406 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.743604 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.743744 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:20Z","lastTransitionTime":"2026-03-13T10:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.846213 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.846540 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.846784 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.846990 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.847208 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:20Z","lastTransitionTime":"2026-03-13T10:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.950825 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.950887 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.950900 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.950921 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:20 crc kubenswrapper[4632]: I0313 10:05:20.950968 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:20Z","lastTransitionTime":"2026-03-13T10:05:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.054423 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.054489 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.054500 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.054518 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.054528 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:21Z","lastTransitionTime":"2026-03-13T10:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.157196 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.157511 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.157524 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.157543 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.157555 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:21Z","lastTransitionTime":"2026-03-13T10:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.260893 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.260972 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.261017 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.261037 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.261048 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:21Z","lastTransitionTime":"2026-03-13T10:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.363359 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.363397 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.363408 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.363432 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.363447 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:21Z","lastTransitionTime":"2026-03-13T10:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.466766 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.466837 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.466851 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.466871 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.466885 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:21Z","lastTransitionTime":"2026-03-13T10:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.569642 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.569704 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.569716 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.569740 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.569751 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:21Z","lastTransitionTime":"2026-03-13T10:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.571259 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9"} Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.571320 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7"} Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.589434 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.608010 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.622058 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.638106 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.659215 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.671547 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.673737 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.673810 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.673840 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.673861 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.673873 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:21Z","lastTransitionTime":"2026-03-13T10:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.744702 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.758056 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.771659 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.776510 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.776571 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.776586 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.776605 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.776617 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:21Z","lastTransitionTime":"2026-03-13T10:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.786222 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.796903 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.808166 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.832131 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.846920 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.865075 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.874100 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.879502 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.879566 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.879580 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.879600 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.879613 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:21Z","lastTransitionTime":"2026-03-13T10:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.982639 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.982687 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.982698 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.982715 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:21 crc kubenswrapper[4632]: I0313 10:05:21.982727 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:21Z","lastTransitionTime":"2026-03-13T10:05:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.043452 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.043592 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.043663 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:22 crc kubenswrapper[4632]: E0313 10:05:22.043760 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.043595 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:22 crc kubenswrapper[4632]: E0313 10:05:22.043623 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:22 crc kubenswrapper[4632]: E0313 10:05:22.043888 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:22 crc kubenswrapper[4632]: E0313 10:05:22.043950 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.044711 4632 scope.go:117] "RemoveContainer" containerID="9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae" Mar 13 10:05:22 crc kubenswrapper[4632]: E0313 10:05:22.044995 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.085474 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.085513 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.085525 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.085541 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.085553 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:22Z","lastTransitionTime":"2026-03-13T10:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.189262 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.189327 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.189341 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.189361 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.189374 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:22Z","lastTransitionTime":"2026-03-13T10:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.292046 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.292081 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.292090 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.292103 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.292123 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:22Z","lastTransitionTime":"2026-03-13T10:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.395277 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.395334 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.395352 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.395632 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.395668 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:22Z","lastTransitionTime":"2026-03-13T10:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.499461 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.499503 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.499512 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.499531 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.499542 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:22Z","lastTransitionTime":"2026-03-13T10:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.603809 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.603873 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.603885 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.603908 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.603922 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:22Z","lastTransitionTime":"2026-03-13T10:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.707293 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.707342 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.707361 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.707379 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.707390 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:22Z","lastTransitionTime":"2026-03-13T10:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.810540 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.810599 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.810610 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.810629 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.810640 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:22Z","lastTransitionTime":"2026-03-13T10:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.914819 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.914898 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.914913 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.914959 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:22 crc kubenswrapper[4632]: I0313 10:05:22.914979 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:22Z","lastTransitionTime":"2026-03-13T10:05:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.018370 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.018420 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.018436 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.018462 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.018525 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:23Z","lastTransitionTime":"2026-03-13T10:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.121988 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.122032 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.122044 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.122062 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.122073 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:23Z","lastTransitionTime":"2026-03-13T10:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.284331 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.284398 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.284424 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.284450 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.284465 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:23Z","lastTransitionTime":"2026-03-13T10:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.401283 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.401818 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.401835 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.401861 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.401878 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:23Z","lastTransitionTime":"2026-03-13T10:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.536135 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.536187 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.536199 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.536223 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.536237 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:23Z","lastTransitionTime":"2026-03-13T10:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.580804 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3"} Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.658484 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.658529 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.658541 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.658558 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.658570 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:23Z","lastTransitionTime":"2026-03-13T10:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.760761 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.761207 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.761314 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.761420 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.761519 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:23Z","lastTransitionTime":"2026-03-13T10:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.864761 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.865231 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.865349 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.865473 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.865563 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:23Z","lastTransitionTime":"2026-03-13T10:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.969584 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.969641 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.969654 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.969671 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:23 crc kubenswrapper[4632]: I0313 10:05:23.969684 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:23Z","lastTransitionTime":"2026-03-13T10:05:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.043407 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.043476 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.043505 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.043601 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:24 crc kubenswrapper[4632]: E0313 10:05:24.043928 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:24 crc kubenswrapper[4632]: E0313 10:05:24.045039 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:24 crc kubenswrapper[4632]: E0313 10:05:24.045137 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:24 crc kubenswrapper[4632]: E0313 10:05:24.045413 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.073859 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.073914 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.074126 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.074440 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.074455 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:24Z","lastTransitionTime":"2026-03-13T10:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.181581 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.181614 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.181624 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.181639 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.181649 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:24Z","lastTransitionTime":"2026-03-13T10:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.284242 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.284738 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.284749 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.284765 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.284780 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:24Z","lastTransitionTime":"2026-03-13T10:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.387516 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.387556 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.387567 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.387583 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.387593 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:24Z","lastTransitionTime":"2026-03-13T10:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.491649 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.491693 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.491704 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.491720 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.491731 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:24Z","lastTransitionTime":"2026-03-13T10:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.594843 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.594899 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.594912 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.594930 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.595101 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:24Z","lastTransitionTime":"2026-03-13T10:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.698533 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.698592 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.698602 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.698615 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.698627 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:24Z","lastTransitionTime":"2026-03-13T10:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.801790 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.801829 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.801838 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.801852 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.801865 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:24Z","lastTransitionTime":"2026-03-13T10:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.904438 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.904487 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.904499 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.904518 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:24 crc kubenswrapper[4632]: I0313 10:05:24.904530 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:24Z","lastTransitionTime":"2026-03-13T10:05:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.006886 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.006925 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.006949 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.006965 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.006976 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:25Z","lastTransitionTime":"2026-03-13T10:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.111104 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.111278 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.111372 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.111473 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.111620 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:25Z","lastTransitionTime":"2026-03-13T10:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.214731 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.214791 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.214810 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.214839 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.214856 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:25Z","lastTransitionTime":"2026-03-13T10:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.318276 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.318305 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.318316 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.318333 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.318346 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:25Z","lastTransitionTime":"2026-03-13T10:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.423919 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.424381 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.424394 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.424412 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.424424 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:25Z","lastTransitionTime":"2026-03-13T10:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.526639 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.526681 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.526693 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.526708 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.526716 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:25Z","lastTransitionTime":"2026-03-13T10:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.593639 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-zwlc8" event={"ID":"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7","Type":"ContainerStarted","Data":"a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.596249 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gqf22" event={"ID":"4ec8e301-3037-4de0-94d2-32c49709660e","Type":"ContainerStarted","Data":"0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.598955 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.600469 4632 generic.go:334] "Generic (PLEG): container finished" podID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerID="fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50" exitCode=0 Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.600522 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerDied","Data":"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.602931 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-n55jt" event={"ID":"b29b9ad7-8cc9-434f-8731-a86265c383fd","Type":"ContainerStarted","Data":"6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.607142 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" event={"ID":"b054ca08-1d09-4eca-a608-eb5b9323959a","Type":"ContainerStarted","Data":"a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.609438 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.611594 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" event={"ID":"b0c542d5-8c38-4243-8af7-cfc0d8e22773","Type":"ContainerStarted","Data":"a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.611627 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" event={"ID":"b0c542d5-8c38-4243-8af7-cfc0d8e22773","Type":"ContainerStarted","Data":"55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.618846 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.629690 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.629724 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.629733 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.629746 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.629755 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:25Z","lastTransitionTime":"2026-03-13T10:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.645482 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.667082 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.682396 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.695814 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.708330 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.722185 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.733317 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.733380 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.733394 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.733414 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.733426 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:25Z","lastTransitionTime":"2026-03-13T10:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.739563 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.773916 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.835743 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.835781 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.835791 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.835805 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.835814 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:25Z","lastTransitionTime":"2026-03-13T10:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.839198 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.866533 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.879564 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.894813 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.919231 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.943853 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.944969 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.945119 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.945150 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.945182 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.945195 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:25Z","lastTransitionTime":"2026-03-13T10:05:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.961349 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.978723 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:05:25 crc kubenswrapper[4632]: E0313 10:05:25.978864 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:05:41.978839785 +0000 UTC m=+116.001369918 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.979138 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.979292 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:25 crc kubenswrapper[4632]: E0313 10:05:25.979336 4632 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:05:25 crc kubenswrapper[4632]: E0313 10:05:25.979555 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:41.979542896 +0000 UTC m=+116.002073029 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:05:25 crc kubenswrapper[4632]: E0313 10:05:25.979373 4632 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:05:25 crc kubenswrapper[4632]: E0313 10:05:25.979779 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:41.979769024 +0000 UTC m=+116.002299157 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:05:25 crc kubenswrapper[4632]: I0313 10:05:25.981748 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:25Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.016821 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.041266 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.043310 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.043452 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.043535 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.043466 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:26 crc kubenswrapper[4632]: E0313 10:05:26.043458 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:26 crc kubenswrapper[4632]: E0313 10:05:26.043752 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:26 crc kubenswrapper[4632]: E0313 10:05:26.043627 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:26 crc kubenswrapper[4632]: E0313 10:05:26.044993 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.047428 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.047471 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.047481 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.047497 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.047507 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:26Z","lastTransitionTime":"2026-03-13T10:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.068578 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.080578 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.080713 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.080771 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:26 crc kubenswrapper[4632]: E0313 10:05:26.081186 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:05:26 crc kubenswrapper[4632]: E0313 10:05:26.081231 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:05:26 crc kubenswrapper[4632]: E0313 10:05:26.081246 4632 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:26 crc kubenswrapper[4632]: E0313 10:05:26.081244 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:05:26 crc kubenswrapper[4632]: E0313 10:05:26.081294 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:05:26 crc kubenswrapper[4632]: E0313 10:05:26.081310 4632 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:26 crc kubenswrapper[4632]: E0313 10:05:26.081331 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:42.081312065 +0000 UTC m=+116.103842198 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:26 crc kubenswrapper[4632]: E0313 10:05:26.081399 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 10:05:42.081372367 +0000 UTC m=+116.103902500 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:26 crc kubenswrapper[4632]: E0313 10:05:26.081481 4632 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:05:26 crc kubenswrapper[4632]: E0313 10:05:26.081521 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs podName:ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad nodeName:}" failed. No retries permitted until 2026-03-13 10:05:42.081510841 +0000 UTC m=+116.104041174 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs") pod "network-metrics-daemon-z2vlz" (UID: "ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.088806 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.107787 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.125753 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.137410 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.150322 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.150350 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.150358 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.150372 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.150380 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:26Z","lastTransitionTime":"2026-03-13T10:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.176100 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.194877 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.210300 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.220250 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.249413 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.252217 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.252239 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.252248 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.252262 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.252290 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:26Z","lastTransitionTime":"2026-03-13T10:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.271607 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.297323 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.343792 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.354843 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.354870 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.354878 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.354906 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.354915 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:26Z","lastTransitionTime":"2026-03-13T10:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.459399 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.459445 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.459461 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.459479 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.459491 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:26Z","lastTransitionTime":"2026-03-13T10:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.562492 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.562534 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.562545 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.562561 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.562574 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:26Z","lastTransitionTime":"2026-03-13T10:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.618848 4632 generic.go:334] "Generic (PLEG): container finished" podID="b054ca08-1d09-4eca-a608-eb5b9323959a" containerID="a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35" exitCode=0 Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.618956 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" event={"ID":"b054ca08-1d09-4eca-a608-eb5b9323959a","Type":"ContainerDied","Data":"a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.637683 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerStarted","Data":"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.637730 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerStarted","Data":"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.637742 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerStarted","Data":"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.637754 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerStarted","Data":"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.637765 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerStarted","Data":"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.662467 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.664409 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.664443 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.664458 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.664475 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.664487 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:26Z","lastTransitionTime":"2026-03-13T10:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.680696 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.700593 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.731420 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.752175 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.767991 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.768019 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.768027 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.768057 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.768067 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:26Z","lastTransitionTime":"2026-03-13T10:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.775015 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.802825 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.831997 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.851260 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.864817 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.971047 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.971273 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.971363 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.971447 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.971524 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:26Z","lastTransitionTime":"2026-03-13T10:05:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:26 crc kubenswrapper[4632]: I0313 10:05:26.984215 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.005965 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:27Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.027488 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:27Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.047337 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:27Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.069880 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:27Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.093503 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:27Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.106565 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.106604 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.106618 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.106649 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.106663 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:27Z","lastTransitionTime":"2026-03-13T10:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.210500 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.210536 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.210547 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.210563 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.210572 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:27Z","lastTransitionTime":"2026-03-13T10:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.313825 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.313882 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.313892 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.313907 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.313918 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:27Z","lastTransitionTime":"2026-03-13T10:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.417449 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.417500 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.417510 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.417529 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.417538 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:27Z","lastTransitionTime":"2026-03-13T10:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.519529 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.519578 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.519588 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.519605 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.519614 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:27Z","lastTransitionTime":"2026-03-13T10:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.621529 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.621557 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.621565 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.621578 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.621586 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:27Z","lastTransitionTime":"2026-03-13T10:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.687313 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerStarted","Data":"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f"} Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.689071 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" event={"ID":"b054ca08-1d09-4eca-a608-eb5b9323959a","Type":"ContainerStarted","Data":"c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d"} Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.691087 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c"} Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.782985 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.783021 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.783031 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.783045 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.783054 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:27Z","lastTransitionTime":"2026-03-13T10:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.787406 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:27Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.810859 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:27Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.842972 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:27Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.871363 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:27Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.885544 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.885596 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.885611 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.885631 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.885639 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:27Z","lastTransitionTime":"2026-03-13T10:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.897038 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:27Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.938849 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:27Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.961505 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:27Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.977298 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:27Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.988491 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.988541 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.988553 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.988570 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.988584 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:27Z","lastTransitionTime":"2026-03-13T10:05:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:27 crc kubenswrapper[4632]: I0313 10:05:27.996022 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"services have not yet been read at least once, cannot construct envvars\\\",\\\"reason\\\":\\\"CreateContainerConfigError\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:27Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.015688 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.039278 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.051799 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.051816 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.051909 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.051930 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:28 crc kubenswrapper[4632]: E0313 10:05:28.052055 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:28 crc kubenswrapper[4632]: E0313 10:05:28.052466 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:28 crc kubenswrapper[4632]: E0313 10:05:28.052527 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:28 crc kubenswrapper[4632]: E0313 10:05:28.052592 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.070107 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.085752 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.091568 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.091618 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.091634 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.091653 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.091665 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:28Z","lastTransitionTime":"2026-03-13T10:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.118630 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.138428 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.153147 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.181228 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.195221 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.195325 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.195340 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.195359 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.195402 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:28Z","lastTransitionTime":"2026-03-13T10:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.195440 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.227570 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.263850 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.295305 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.297567 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.297610 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.297622 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.297640 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.297653 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:28Z","lastTransitionTime":"2026-03-13T10:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.333874 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.352472 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.374063 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.383925 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.394930 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.401041 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.401103 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.401176 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.401207 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.401229 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:28Z","lastTransitionTime":"2026-03-13T10:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.412201 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.424310 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.436616 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.462695 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.476099 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.485026 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.494228 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.575185 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.575225 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.575237 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.575254 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.575265 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:28Z","lastTransitionTime":"2026-03-13T10:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.588158 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.600646 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.627140 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.638031 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.654966 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.666840 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.677902 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.677876 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.678007 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.678022 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.678038 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.678047 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:28Z","lastTransitionTime":"2026-03-13T10:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.687818 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.701221 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.715722 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.734817 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.749919 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.767702 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.780812 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.780849 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.780858 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.780876 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.780886 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:28Z","lastTransitionTime":"2026-03-13T10:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.785479 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.800246 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.888456 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.888504 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.888516 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.888554 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.888580 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:28Z","lastTransitionTime":"2026-03-13T10:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.991388 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.991410 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.991417 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.991431 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:28 crc kubenswrapper[4632]: I0313 10:05:28.991443 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:28Z","lastTransitionTime":"2026-03-13T10:05:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.096707 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.096841 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.097104 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.097373 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.097635 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:29Z","lastTransitionTime":"2026-03-13T10:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.200578 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.200613 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.200622 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.200636 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.200644 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:29Z","lastTransitionTime":"2026-03-13T10:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.303231 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.303270 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.303284 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.303300 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.303310 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:29Z","lastTransitionTime":"2026-03-13T10:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.406389 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.406442 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.406458 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.406479 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.406495 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:29Z","lastTransitionTime":"2026-03-13T10:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.511341 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.511414 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.511441 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.511472 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.511495 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:29Z","lastTransitionTime":"2026-03-13T10:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.615767 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.615816 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.615834 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.615854 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.615868 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:29Z","lastTransitionTime":"2026-03-13T10:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.704613 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerStarted","Data":"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d"} Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.720062 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.720123 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.720138 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.720161 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.720177 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:29Z","lastTransitionTime":"2026-03-13T10:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.823078 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.823117 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.823129 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.823147 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.823159 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:29Z","lastTransitionTime":"2026-03-13T10:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.926093 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.926134 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.926144 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.926182 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:29 crc kubenswrapper[4632]: I0313 10:05:29.926212 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:29Z","lastTransitionTime":"2026-03-13T10:05:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.029058 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.029126 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.029140 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.029163 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.029177 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.043501 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.043622 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.043642 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.043550 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:30 crc kubenswrapper[4632]: E0313 10:05:30.043763 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:30 crc kubenswrapper[4632]: E0313 10:05:30.044042 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:30 crc kubenswrapper[4632]: E0313 10:05:30.044179 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:30 crc kubenswrapper[4632]: E0313 10:05:30.044293 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.132476 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.132562 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.132578 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.132600 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.132615 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.235530 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.235569 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.235579 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.235594 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.235603 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.338422 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.338794 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.339111 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.339322 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.339499 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.442151 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.442185 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.442222 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.442238 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.442250 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.556929 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.556992 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.557001 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.557016 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.557026 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.659716 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.659800 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.659814 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.659833 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.659844 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.710185 4632 generic.go:334] "Generic (PLEG): container finished" podID="b054ca08-1d09-4eca-a608-eb5b9323959a" containerID="c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d" exitCode=0 Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.710218 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" event={"ID":"b054ca08-1d09-4eca-a608-eb5b9323959a","Type":"ContainerDied","Data":"c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.726237 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.740174 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.755191 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.762112 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.762343 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.762483 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.762596 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.762709 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.769555 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.781331 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.781367 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.781378 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.781393 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.781403 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.783438 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: E0313 10:05:30.793828 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.794372 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.798650 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.798871 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.799008 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.799122 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.799199 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.807620 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: E0313 10:05:30.812445 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.818761 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.819069 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.819113 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.819128 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.819144 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.819178 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.831858 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: E0313 10:05:30.834860 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.838562 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.838594 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.838605 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.838623 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.838633 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.848428 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: E0313 10:05:30.851891 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.856394 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.856548 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.856645 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.856735 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.856866 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.859831 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.869623 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: E0313 10:05:30.869695 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:30Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: E0313 10:05:30.869985 4632 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.875076 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.875109 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.875119 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.875135 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.875146 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.879099 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.897879 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.910395 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.922332 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:30Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.978064 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.978101 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.978146 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.978163 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:30 crc kubenswrapper[4632]: I0313 10:05:30.978173 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:30Z","lastTransitionTime":"2026-03-13T10:05:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.081120 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.081385 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.081466 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.081600 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.081719 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:31Z","lastTransitionTime":"2026-03-13T10:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.184345 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.184667 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.184850 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.184955 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.185045 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:31Z","lastTransitionTime":"2026-03-13T10:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.287076 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.287434 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.287505 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.287610 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.287683 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:31Z","lastTransitionTime":"2026-03-13T10:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.390210 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.390264 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.390280 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.390300 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.390313 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:31Z","lastTransitionTime":"2026-03-13T10:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.493185 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.493232 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.493241 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.493257 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.493266 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:31Z","lastTransitionTime":"2026-03-13T10:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.595654 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.595693 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.595701 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.595715 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.595724 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:31Z","lastTransitionTime":"2026-03-13T10:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.697829 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.697882 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.697893 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.697907 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.697917 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:31Z","lastTransitionTime":"2026-03-13T10:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.716984 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" event={"ID":"b054ca08-1d09-4eca-a608-eb5b9323959a","Type":"ContainerStarted","Data":"b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e"} Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.800224 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.800263 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.800279 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.800299 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.800309 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:31Z","lastTransitionTime":"2026-03-13T10:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.903042 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.903096 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.903109 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.903127 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:31 crc kubenswrapper[4632]: I0313 10:05:31.903141 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:31Z","lastTransitionTime":"2026-03-13T10:05:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.004789 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.004822 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.004830 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.004844 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.004852 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:32Z","lastTransitionTime":"2026-03-13T10:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.044185 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.044225 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.044232 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:32 crc kubenswrapper[4632]: E0313 10:05:32.044350 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.044370 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:32 crc kubenswrapper[4632]: E0313 10:05:32.044480 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:32 crc kubenswrapper[4632]: E0313 10:05:32.044546 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:32 crc kubenswrapper[4632]: E0313 10:05:32.044587 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.107049 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.107132 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.107149 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.107165 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.107176 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:32Z","lastTransitionTime":"2026-03-13T10:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.210073 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.210121 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.210132 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.210152 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.210164 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:32Z","lastTransitionTime":"2026-03-13T10:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.314825 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.314862 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.314873 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.314892 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.314904 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:32Z","lastTransitionTime":"2026-03-13T10:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.418340 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.418382 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.418394 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.418411 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.418422 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:32Z","lastTransitionTime":"2026-03-13T10:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.521685 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.521721 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.521730 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.521744 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.521753 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:32Z","lastTransitionTime":"2026-03-13T10:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.624597 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.624648 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.624657 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.624674 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.624689 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:32Z","lastTransitionTime":"2026-03-13T10:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.727882 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerStarted","Data":"0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9"} Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.730414 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.730521 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.730642 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.734369 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.734431 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.734457 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.734496 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.734511 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:32Z","lastTransitionTime":"2026-03-13T10:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.738721 4632 generic.go:334] "Generic (PLEG): container finished" podID="b054ca08-1d09-4eca-a608-eb5b9323959a" containerID="b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e" exitCode=0 Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.738774 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" event={"ID":"b054ca08-1d09-4eca-a608-eb5b9323959a","Type":"ContainerDied","Data":"b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e"} Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.749869 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.765522 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.766195 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.768458 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.782882 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.801136 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.816357 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.831302 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.838929 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.839228 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.839241 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.839258 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.839270 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:32Z","lastTransitionTime":"2026-03-13T10:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.855509 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.869048 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.911223 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.925411 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.939630 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.941791 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.941834 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.941845 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.941862 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.941872 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:32Z","lastTransitionTime":"2026-03-13T10:05:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.950541 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.974253 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:32 crc kubenswrapper[4632]: I0313 10:05:32.994048 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.030930 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.045310 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.045403 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.045418 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.045442 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.045458 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:33Z","lastTransitionTime":"2026-03-13T10:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.053605 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.072604 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.098670 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.121385 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.137032 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.148763 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.148798 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.148807 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.148822 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.148833 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:33Z","lastTransitionTime":"2026-03-13T10:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.151595 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.170642 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.186711 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.201261 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.221270 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.242668 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.250628 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.250660 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.250669 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.250683 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.250693 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:33Z","lastTransitionTime":"2026-03-13T10:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.258907 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.276545 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.299330 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.324552 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.353611 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.353649 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.353658 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.353679 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.353693 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:33Z","lastTransitionTime":"2026-03-13T10:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.355973 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.372154 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.456581 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.456615 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.456623 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.456639 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.456650 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:33Z","lastTransitionTime":"2026-03-13T10:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.559140 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.559184 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.559193 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.559210 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.559222 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:33Z","lastTransitionTime":"2026-03-13T10:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.662584 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.662658 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.662669 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.662688 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.662701 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:33Z","lastTransitionTime":"2026-03-13T10:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.746804 4632 generic.go:334] "Generic (PLEG): container finished" podID="b054ca08-1d09-4eca-a608-eb5b9323959a" containerID="4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9" exitCode=0 Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.748153 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" event={"ID":"b054ca08-1d09-4eca-a608-eb5b9323959a","Type":"ContainerDied","Data":"4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9"} Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.764333 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.765742 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.765765 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.765840 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.766107 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.766125 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:33Z","lastTransitionTime":"2026-03-13T10:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.806723 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.825421 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.841503 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.862956 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.870779 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.870850 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.870866 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.870887 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.870899 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:33Z","lastTransitionTime":"2026-03-13T10:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.877981 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.892048 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.906151 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.920385 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.938815 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.960657 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.973752 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.973810 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.973831 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.973847 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.973858 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:33Z","lastTransitionTime":"2026-03-13T10:05:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.976507 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:33 crc kubenswrapper[4632]: I0313 10:05:33.992881 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:33Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.009009 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:34Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.026287 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:34Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.040872 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:34Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.044182 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.044249 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.044306 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:34 crc kubenswrapper[4632]: E0313 10:05:34.044322 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.044343 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:34 crc kubenswrapper[4632]: E0313 10:05:34.044494 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:34 crc kubenswrapper[4632]: E0313 10:05:34.044561 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:34 crc kubenswrapper[4632]: E0313 10:05:34.044658 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.077219 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.077283 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.077306 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.077329 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.077343 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:34Z","lastTransitionTime":"2026-03-13T10:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.180700 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.180753 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.180766 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.180789 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.180805 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:34Z","lastTransitionTime":"2026-03-13T10:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.284506 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.284571 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.284588 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.284611 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.284628 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:34Z","lastTransitionTime":"2026-03-13T10:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.388180 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.388263 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.388275 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.388318 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.388331 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:34Z","lastTransitionTime":"2026-03-13T10:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.493462 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.493516 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.493525 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.493540 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.493550 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:34Z","lastTransitionTime":"2026-03-13T10:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.596547 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.596617 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.596629 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.596650 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.596665 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:34Z","lastTransitionTime":"2026-03-13T10:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.699532 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.699590 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.699602 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.699621 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.699634 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:34Z","lastTransitionTime":"2026-03-13T10:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.864432 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" event={"ID":"b054ca08-1d09-4eca-a608-eb5b9323959a","Type":"ContainerStarted","Data":"8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09"} Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.882158 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:34Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.896684 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:34Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.908767 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.908802 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.908810 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.908824 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.908833 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:34Z","lastTransitionTime":"2026-03-13T10:05:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.913997 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:34Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.934916 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:34Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.949109 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:34Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.966054 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:34Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:34 crc kubenswrapper[4632]: I0313 10:05:34.979080 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:34Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.010722 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.010801 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.010812 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.010825 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.010833 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:35Z","lastTransitionTime":"2026-03-13T10:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.045563 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:35Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.057707 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:35Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.071633 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:35Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.081361 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:35Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.109356 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:35Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.115283 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.115319 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.115330 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.115369 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.115383 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:35Z","lastTransitionTime":"2026-03-13T10:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.123838 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:35Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.138764 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:35Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.153204 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:35Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.165715 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:35Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.218715 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.218760 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.218769 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.218787 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.218800 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:35Z","lastTransitionTime":"2026-03-13T10:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.321737 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.321789 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.321803 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.321823 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.321836 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:35Z","lastTransitionTime":"2026-03-13T10:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.424069 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.424098 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.424107 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.424120 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.424128 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:35Z","lastTransitionTime":"2026-03-13T10:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.526295 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.526324 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.526332 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.526344 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.526352 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:35Z","lastTransitionTime":"2026-03-13T10:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.628201 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.628425 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.628516 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.628676 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.628776 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:35Z","lastTransitionTime":"2026-03-13T10:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.731393 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.731677 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.731772 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.731868 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.732019 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:35Z","lastTransitionTime":"2026-03-13T10:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.960429 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.961204 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.961325 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.961442 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:35 crc kubenswrapper[4632]: I0313 10:05:35.961558 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:35Z","lastTransitionTime":"2026-03-13T10:05:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.048195 4632 scope.go:117] "RemoveContainer" containerID="9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.049139 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:36 crc kubenswrapper[4632]: E0313 10:05:36.049336 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.049990 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.050080 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:36 crc kubenswrapper[4632]: E0313 10:05:36.051348 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.051460 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:36 crc kubenswrapper[4632]: E0313 10:05:36.051638 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:36 crc kubenswrapper[4632]: E0313 10:05:36.051777 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.068058 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.068590 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.068687 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.068772 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.068843 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:36Z","lastTransitionTime":"2026-03-13T10:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.172679 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.172721 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.172733 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.172752 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.172763 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:36Z","lastTransitionTime":"2026-03-13T10:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.306126 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.306178 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.306197 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.306216 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.306226 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:36Z","lastTransitionTime":"2026-03-13T10:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.409989 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.410030 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.410039 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.410052 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.410069 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:36Z","lastTransitionTime":"2026-03-13T10:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.512561 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.512615 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.512628 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.512648 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.512661 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:36Z","lastTransitionTime":"2026-03-13T10:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.616278 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.616335 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.616353 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.616371 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.616381 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:36Z","lastTransitionTime":"2026-03-13T10:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.719617 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.719959 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.720090 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.720200 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.720304 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:36Z","lastTransitionTime":"2026-03-13T10:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.922395 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.922731 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.922834 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.923022 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.923131 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:36Z","lastTransitionTime":"2026-03-13T10:05:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.967344 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 10:05:36 crc kubenswrapper[4632]: I0313 10:05:36.969717 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94"} Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.026771 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.026851 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.026868 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.026893 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.026920 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:37Z","lastTransitionTime":"2026-03-13T10:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.237195 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.237231 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.237243 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.237261 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.237273 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:37Z","lastTransitionTime":"2026-03-13T10:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.340312 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.340354 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.340366 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.340383 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.340394 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:37Z","lastTransitionTime":"2026-03-13T10:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.444768 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.444815 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.444829 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.444848 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.444862 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:37Z","lastTransitionTime":"2026-03-13T10:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.548398 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.548454 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.548464 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.548482 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.548494 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:37Z","lastTransitionTime":"2026-03-13T10:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.651647 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.651706 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.651719 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.651738 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.652147 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:37Z","lastTransitionTime":"2026-03-13T10:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.755346 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.755424 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.755438 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.755460 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.755478 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:37Z","lastTransitionTime":"2026-03-13T10:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.858373 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.858431 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.858440 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.858458 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.858466 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:37Z","lastTransitionTime":"2026-03-13T10:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.964903 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.964972 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.964983 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.965002 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.965017 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:37Z","lastTransitionTime":"2026-03-13T10:05:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.976749 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/0.log" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.980725 4632 generic.go:334] "Generic (PLEG): container finished" podID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerID="0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9" exitCode=1 Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.980824 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerDied","Data":"0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9"} Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.981695 4632 scope.go:117] "RemoveContainer" containerID="0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.985399 4632 generic.go:334] "Generic (PLEG): container finished" podID="b054ca08-1d09-4eca-a608-eb5b9323959a" containerID="8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09" exitCode=0 Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.985447 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" event={"ID":"b054ca08-1d09-4eca-a608-eb5b9323959a","Type":"ContainerDied","Data":"8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09"} Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.985804 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:05:37 crc kubenswrapper[4632]: I0313 10:05:37.999249 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:37Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.013993 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.033309 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.045018 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.045091 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.045118 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:38 crc kubenswrapper[4632]: E0313 10:05:38.045174 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.045016 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:38 crc kubenswrapper[4632]: E0313 10:05:38.045322 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:38 crc kubenswrapper[4632]: E0313 10:05:38.045462 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:38 crc kubenswrapper[4632]: E0313 10:05:38.045547 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.048674 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.067352 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.068927 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.069002 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.069015 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.069033 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.069045 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:38Z","lastTransitionTime":"2026-03-13T10:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.087276 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.104623 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.132997 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.155469 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"message\\\":\\\"s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577081 6495 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577192 6495 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.577561 6495 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.578699 6495 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 10:05:37.578717 6495 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 10:05:37.578749 6495 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 10:05:37.578851 6495 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 10:05:37.578884 6495 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 10:05:37.578917 6495 factory.go:656] Stopping watch factory\\\\nI0313 10:05:37.578961 6495 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.171519 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.171843 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.172232 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.172247 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.172265 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.172277 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:38Z","lastTransitionTime":"2026-03-13T10:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.189059 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.203288 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.215690 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.229192 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.243519 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.276985 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.277032 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.277044 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.277067 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.277081 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:38Z","lastTransitionTime":"2026-03-13T10:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.279640 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.304084 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.324275 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.346252 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.368610 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.383881 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.383932 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.383961 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.383984 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.384003 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:38Z","lastTransitionTime":"2026-03-13T10:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.385787 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.402915 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.436645 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.474040 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.496447 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.496506 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.496515 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.496547 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.496559 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:38Z","lastTransitionTime":"2026-03-13T10:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.515053 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.541458 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.561297 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.583899 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.599359 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.599423 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.599439 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.599481 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.599496 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:38Z","lastTransitionTime":"2026-03-13T10:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.601833 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.622290 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.654096 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"message\\\":\\\"s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577081 6495 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577192 6495 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.577561 6495 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.578699 6495 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 10:05:37.578717 6495 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 10:05:37.578749 6495 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 10:05:37.578851 6495 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 10:05:37.578884 6495 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 10:05:37.578917 6495 factory.go:656] Stopping watch factory\\\\nI0313 10:05:37.578961 6495 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.667214 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.679455 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.693804 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.702590 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.702646 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.702658 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.702691 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.702701 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:38Z","lastTransitionTime":"2026-03-13T10:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.707992 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.720128 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.731212 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.753219 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"message\\\":\\\"s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577081 6495 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577192 6495 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.577561 6495 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.578699 6495 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 10:05:37.578717 6495 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 10:05:37.578749 6495 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 10:05:37.578851 6495 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 10:05:37.578884 6495 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 10:05:37.578917 6495 factory.go:656] Stopping watch factory\\\\nI0313 10:05:37.578961 6495 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.766559 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.777534 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.798209 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.805698 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.805773 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.805792 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.805818 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.805836 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:38Z","lastTransitionTime":"2026-03-13T10:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.814679 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.831464 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.848269 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.866801 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.884156 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.909610 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.909665 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.909687 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.909711 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.909737 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:38Z","lastTransitionTime":"2026-03-13T10:05:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.912919 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.928030 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:38Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.996139 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/0.log" Mar 13 10:05:38 crc kubenswrapper[4632]: I0313 10:05:38.999504 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerStarted","Data":"934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e"} Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.000105 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.005147 4632 generic.go:334] "Generic (PLEG): container finished" podID="b054ca08-1d09-4eca-a608-eb5b9323959a" containerID="be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41" exitCode=0 Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.005213 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" event={"ID":"b054ca08-1d09-4eca-a608-eb5b9323959a","Type":"ContainerDied","Data":"be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41"} Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.012442 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.012479 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.012489 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.012508 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.012519 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:39Z","lastTransitionTime":"2026-03-13T10:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.022214 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.050070 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.068925 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.086959 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.105064 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.116615 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.116669 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.116681 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.116701 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.116715 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:39Z","lastTransitionTime":"2026-03-13T10:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.128131 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.189925 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.215594 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.221234 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.221275 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.221288 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.221307 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.221322 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:39Z","lastTransitionTime":"2026-03-13T10:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.251556 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"message\\\":\\\"s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577081 6495 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577192 6495 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.577561 6495 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.578699 6495 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 10:05:37.578717 6495 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 10:05:37.578749 6495 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 10:05:37.578851 6495 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 10:05:37.578884 6495 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 10:05:37.578917 6495 factory.go:656] Stopping watch factory\\\\nI0313 10:05:37.578961 6495 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.268766 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.289138 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.310587 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.324862 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.324918 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.324930 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.324967 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.324982 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:39Z","lastTransitionTime":"2026-03-13T10:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.330773 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.350321 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.373335 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.394448 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.424670 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"message\\\":\\\"s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577081 6495 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577192 6495 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.577561 6495 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.578699 6495 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 10:05:37.578717 6495 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 10:05:37.578749 6495 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 10:05:37.578851 6495 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 10:05:37.578884 6495 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 10:05:37.578917 6495 factory.go:656] Stopping watch factory\\\\nI0313 10:05:37.578961 6495 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.428056 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.428156 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.428167 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.428187 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.428199 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:39Z","lastTransitionTime":"2026-03-13T10:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.444014 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.462696 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.482370 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.501033 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.513470 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.529822 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.531872 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.532066 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.532168 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.532268 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.532399 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:39Z","lastTransitionTime":"2026-03-13T10:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.554041 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.567546 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.588238 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.605074 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.619850 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.635646 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.635707 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.635715 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.635734 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.635749 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:39Z","lastTransitionTime":"2026-03-13T10:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.636133 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.657712 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.679318 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.696019 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:39Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.739814 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.739864 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.739874 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.739893 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.739904 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:39Z","lastTransitionTime":"2026-03-13T10:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.842711 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.842752 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.842764 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.842785 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.842797 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:39Z","lastTransitionTime":"2026-03-13T10:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.946154 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.946311 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.946327 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.946346 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:39 crc kubenswrapper[4632]: I0313 10:05:39.946360 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:39Z","lastTransitionTime":"2026-03-13T10:05:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.035957 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" event={"ID":"b054ca08-1d09-4eca-a608-eb5b9323959a","Type":"ContainerStarted","Data":"1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3"} Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.044074 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:40 crc kubenswrapper[4632]: E0313 10:05:40.044175 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.044451 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:40 crc kubenswrapper[4632]: E0313 10:05:40.044520 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.044579 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:40 crc kubenswrapper[4632]: E0313 10:05:40.044628 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.044678 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:40 crc kubenswrapper[4632]: E0313 10:05:40.044738 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.052313 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.052362 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.052396 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.052419 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.052434 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:40Z","lastTransitionTime":"2026-03-13T10:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.053691 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.066443 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.082349 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.096043 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.111505 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.127836 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.145448 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.157004 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.157071 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.157088 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.157114 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.157129 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:40Z","lastTransitionTime":"2026-03-13T10:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.161573 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.176612 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.198816 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"message\\\":\\\"s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577081 6495 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577192 6495 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.577561 6495 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.578699 6495 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 10:05:37.578717 6495 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 10:05:37.578749 6495 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 10:05:37.578851 6495 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 10:05:37.578884 6495 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 10:05:37.578917 6495 factory.go:656] Stopping watch factory\\\\nI0313 10:05:37.578961 6495 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.216371 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.230722 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.244253 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.260530 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.260983 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.261016 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.261031 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.261051 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.261065 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:40Z","lastTransitionTime":"2026-03-13T10:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.278349 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.299867 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:40Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.365668 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.365746 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.365764 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.365788 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.365805 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:40Z","lastTransitionTime":"2026-03-13T10:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.469236 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.469307 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.469319 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.469347 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.469361 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:40Z","lastTransitionTime":"2026-03-13T10:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.573053 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.573131 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.573146 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.573172 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.573188 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:40Z","lastTransitionTime":"2026-03-13T10:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.677155 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.677213 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.677226 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.677253 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.677271 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:40Z","lastTransitionTime":"2026-03-13T10:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.780233 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.780296 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.780311 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.780330 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.780343 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:40Z","lastTransitionTime":"2026-03-13T10:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.883007 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.883054 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.883066 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.883086 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.883100 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:40Z","lastTransitionTime":"2026-03-13T10:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.986328 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.986401 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.986412 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.986432 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.986445 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:40Z","lastTransitionTime":"2026-03-13T10:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.988070 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.988142 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.988157 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.988181 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:40 crc kubenswrapper[4632]: I0313 10:05:40.988195 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:40Z","lastTransitionTime":"2026-03-13T10:05:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:41 crc kubenswrapper[4632]: E0313 10:05:41.003885 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.008418 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.008465 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.008480 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.008505 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.008521 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:41Z","lastTransitionTime":"2026-03-13T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:41 crc kubenswrapper[4632]: E0313 10:05:41.023643 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.028358 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.028403 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.028413 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.028435 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.028448 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:41Z","lastTransitionTime":"2026-03-13T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.044347 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/1.log" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.044982 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/0.log" Mar 13 10:05:41 crc kubenswrapper[4632]: E0313 10:05:41.048178 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.049154 4632 generic.go:334] "Generic (PLEG): container finished" podID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerID="934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e" exitCode=1 Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.049220 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerDied","Data":"934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e"} Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.049332 4632 scope.go:117] "RemoveContainer" containerID="0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.050341 4632 scope.go:117] "RemoveContainer" containerID="934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e" Mar 13 10:05:41 crc kubenswrapper[4632]: E0313 10:05:41.050629 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.054380 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.054433 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.054444 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.054462 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.054474 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:41Z","lastTransitionTime":"2026-03-13T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.073067 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: E0313 10:05:41.074799 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.080358 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.080402 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.080416 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.080439 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.080454 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:41Z","lastTransitionTime":"2026-03-13T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.090690 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: E0313 10:05:41.098442 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:41Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: E0313 10:05:41.098631 4632 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.100927 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.100989 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.101004 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.101026 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.101037 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:41Z","lastTransitionTime":"2026-03-13T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.105282 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.124062 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.138658 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.153177 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.178847 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"message\\\":\\\"s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577081 6495 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577192 6495 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.577561 6495 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.578699 6495 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 10:05:37.578717 6495 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 10:05:37.578749 6495 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 10:05:37.578851 6495 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 10:05:37.578884 6495 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 10:05:37.578917 6495 factory.go:656] Stopping watch factory\\\\nI0313 10:05:37.578961 6495 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"message\\\":\\\"{c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.681766 6736 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0313 10:05:39.682059 6736 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.682124 6736 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.682142 6736 ovnkube.go:599] Stopped ovnkube\\\\nI0313 10:05:39.682173 6736 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 10:05:39.682232 6736 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.193919 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.204903 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.204989 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.205003 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.205026 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.205039 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:41Z","lastTransitionTime":"2026-03-13T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.213511 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.232975 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.253170 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.271850 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.289064 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.307424 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.308082 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.308122 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.308137 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.308161 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.308174 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:41Z","lastTransitionTime":"2026-03-13T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.323595 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.341233 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:41Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.413502 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.413571 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.413588 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.413610 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.413629 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:41Z","lastTransitionTime":"2026-03-13T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.517510 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.517555 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.517567 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.517585 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.517596 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:41Z","lastTransitionTime":"2026-03-13T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.620809 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.620864 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.620878 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.620897 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.620911 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:41Z","lastTransitionTime":"2026-03-13T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.724025 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.724070 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.724082 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.724098 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.724108 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:41Z","lastTransitionTime":"2026-03-13T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.827981 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.828030 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.828045 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.828065 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.828078 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:41Z","lastTransitionTime":"2026-03-13T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.932410 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.932474 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.932488 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.932512 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:41 crc kubenswrapper[4632]: I0313 10:05:41.932525 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:41Z","lastTransitionTime":"2026-03-13T10:05:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.012609 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.012812 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.012900 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.013053 4632 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.013159 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:06:14.013127415 +0000 UTC m=+148.035657548 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.013466 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:06:14.013453475 +0000 UTC m=+148.035983608 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.013568 4632 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.013618 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:06:14.01360781 +0000 UTC m=+148.036137943 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.036289 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.036367 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.036381 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.036400 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.036416 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:42Z","lastTransitionTime":"2026-03-13T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.043795 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.043870 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.043929 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.044031 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.043795 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.044198 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.044332 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.044457 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.058705 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/1.log" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.113707 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.113788 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.113818 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.114029 4632 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.114114 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs podName:ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad nodeName:}" failed. No retries permitted until 2026-03-13 10:06:14.114088429 +0000 UTC m=+148.136618562 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs") pod "network-metrics-daemon-z2vlz" (UID: "ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.114400 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.114422 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.114437 4632 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.114480 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 10:06:14.11446572 +0000 UTC m=+148.136995853 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.114542 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.114555 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.114566 4632 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:42 crc kubenswrapper[4632]: E0313 10:05:42.114595 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 10:06:14.114584094 +0000 UTC m=+148.137114227 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.139418 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.139871 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.139986 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.140080 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.140144 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:42Z","lastTransitionTime":"2026-03-13T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.243830 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.243894 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.244208 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.244274 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.244299 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:42Z","lastTransitionTime":"2026-03-13T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.349366 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.349698 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.349778 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.349869 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.349984 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:42Z","lastTransitionTime":"2026-03-13T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.452908 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.452990 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.453004 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.453022 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.453035 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:42Z","lastTransitionTime":"2026-03-13T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.556544 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.556602 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.556640 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.556662 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.556676 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:42Z","lastTransitionTime":"2026-03-13T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.660189 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.660250 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.660263 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.660283 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.660298 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:42Z","lastTransitionTime":"2026-03-13T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.763691 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.763760 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.763769 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.763785 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.763795 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:42Z","lastTransitionTime":"2026-03-13T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.866210 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.866270 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.866283 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.866348 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.866405 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:42Z","lastTransitionTime":"2026-03-13T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.969533 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.969611 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.969628 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.969646 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:42 crc kubenswrapper[4632]: I0313 10:05:42.969706 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:42Z","lastTransitionTime":"2026-03-13T10:05:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.071637 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.071700 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.071712 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.071731 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.071745 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:43Z","lastTransitionTime":"2026-03-13T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.174571 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.174627 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.174641 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.174660 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.174678 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:43Z","lastTransitionTime":"2026-03-13T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.278108 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.278171 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.278186 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.278212 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.278229 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:43Z","lastTransitionTime":"2026-03-13T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.381507 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.381573 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.381585 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.381605 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.381615 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:43Z","lastTransitionTime":"2026-03-13T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.484400 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.484446 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.484456 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.484469 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.484478 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:43Z","lastTransitionTime":"2026-03-13T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.587488 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.587531 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.587540 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.587552 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.587561 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:43Z","lastTransitionTime":"2026-03-13T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.690602 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.690678 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.690694 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.690714 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.690726 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:43Z","lastTransitionTime":"2026-03-13T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.793596 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.793662 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.793683 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.793708 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.793726 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:43Z","lastTransitionTime":"2026-03-13T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.897041 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.897086 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.897095 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.897111 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.897125 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:43Z","lastTransitionTime":"2026-03-13T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.999830 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.999902 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.999911 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:43 crc kubenswrapper[4632]: I0313 10:05:43.999925 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:43.999963 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:43Z","lastTransitionTime":"2026-03-13T10:05:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.043747 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.043811 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.043811 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:44 crc kubenswrapper[4632]: E0313 10:05:44.043916 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:44 crc kubenswrapper[4632]: E0313 10:05:44.044046 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.044107 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:44 crc kubenswrapper[4632]: E0313 10:05:44.044178 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:44 crc kubenswrapper[4632]: E0313 10:05:44.044245 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.103469 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.104062 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.104146 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.104226 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.104301 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:44Z","lastTransitionTime":"2026-03-13T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.207285 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.207354 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.207368 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.207389 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.207406 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:44Z","lastTransitionTime":"2026-03-13T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.311122 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.311197 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.311219 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.311245 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.311260 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:44Z","lastTransitionTime":"2026-03-13T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.414126 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.414183 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.414195 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.414214 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.414226 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:44Z","lastTransitionTime":"2026-03-13T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.517232 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.517274 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.517283 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.517297 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.517307 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:44Z","lastTransitionTime":"2026-03-13T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.619670 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.619777 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.619786 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.619800 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.619808 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:44Z","lastTransitionTime":"2026-03-13T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.721706 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.721812 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.721831 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.723191 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.723250 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:44Z","lastTransitionTime":"2026-03-13T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.826359 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.826406 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.826419 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.826435 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.826449 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:44Z","lastTransitionTime":"2026-03-13T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.929322 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.929375 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.929387 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.929402 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:44 crc kubenswrapper[4632]: I0313 10:05:44.929412 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:44Z","lastTransitionTime":"2026-03-13T10:05:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.031695 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.031732 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.031742 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.031756 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.032012 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:45Z","lastTransitionTime":"2026-03-13T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.134699 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.134747 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.134755 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.134774 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.134784 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:45Z","lastTransitionTime":"2026-03-13T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.237678 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.237730 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.237738 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.237753 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.237763 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:45Z","lastTransitionTime":"2026-03-13T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.340606 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.340652 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.340664 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.340682 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.340694 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:45Z","lastTransitionTime":"2026-03-13T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.443287 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.443356 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.443372 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.443395 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.443414 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:45Z","lastTransitionTime":"2026-03-13T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.546435 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.546485 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.546498 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.546517 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.546531 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:45Z","lastTransitionTime":"2026-03-13T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.648810 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.648870 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.648885 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.648904 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.648921 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:45Z","lastTransitionTime":"2026-03-13T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.750961 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.751010 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.751022 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.751039 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.751050 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:45Z","lastTransitionTime":"2026-03-13T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.854450 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.854491 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.854499 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.854518 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.854527 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:45Z","lastTransitionTime":"2026-03-13T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.958014 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.958084 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.958109 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.958140 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:45 crc kubenswrapper[4632]: I0313 10:05:45.958166 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:45Z","lastTransitionTime":"2026-03-13T10:05:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.044208 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.044216 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.044380 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:46 crc kubenswrapper[4632]: E0313 10:05:46.044406 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.044430 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:46 crc kubenswrapper[4632]: E0313 10:05:46.044517 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:46 crc kubenswrapper[4632]: E0313 10:05:46.044670 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:46 crc kubenswrapper[4632]: E0313 10:05:46.044740 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.060679 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.060724 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.060756 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.060772 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.060782 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:46Z","lastTransitionTime":"2026-03-13T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.093597 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.105281 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.119829 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.135729 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.153491 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.164439 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.164486 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.164499 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.164518 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.164527 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:46Z","lastTransitionTime":"2026-03-13T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.177273 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"message\\\":\\\"s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577081 6495 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577192 6495 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.577561 6495 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.578699 6495 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 10:05:37.578717 6495 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 10:05:37.578749 6495 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 10:05:37.578851 6495 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 10:05:37.578884 6495 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 10:05:37.578917 6495 factory.go:656] Stopping watch factory\\\\nI0313 10:05:37.578961 6495 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"message\\\":\\\"{c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.681766 6736 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0313 10:05:39.682059 6736 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.682124 6736 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.682142 6736 ovnkube.go:599] Stopped ovnkube\\\\nI0313 10:05:39.682173 6736 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 10:05:39.682232 6736 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.189201 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.210474 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.224322 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.238513 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.251313 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.264840 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.267837 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.267908 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.267924 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.268308 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.268350 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:46Z","lastTransitionTime":"2026-03-13T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.279729 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.294027 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.308430 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.322529 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.334557 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:46Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.370715 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.370750 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.370760 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.370776 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.370787 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:46Z","lastTransitionTime":"2026-03-13T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.473376 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.473480 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.473496 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.473520 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.473535 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:46Z","lastTransitionTime":"2026-03-13T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.576247 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.576297 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.576306 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.576319 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.576329 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:46Z","lastTransitionTime":"2026-03-13T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.679386 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.679450 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.679471 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.679495 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.679513 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:46Z","lastTransitionTime":"2026-03-13T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.782487 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.782524 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.782533 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.782549 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.782560 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:46Z","lastTransitionTime":"2026-03-13T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.886153 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.886213 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.886226 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.886246 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.886260 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:46Z","lastTransitionTime":"2026-03-13T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.990172 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.990245 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.990259 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.990281 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:46 crc kubenswrapper[4632]: I0313 10:05:46.990295 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:46Z","lastTransitionTime":"2026-03-13T10:05:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.062437 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.093778 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.093827 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.093841 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.093866 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.093879 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:47Z","lastTransitionTime":"2026-03-13T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.196968 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.197018 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.197029 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.197047 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.197060 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:47Z","lastTransitionTime":"2026-03-13T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.300432 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.300493 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.300504 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.300523 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.300536 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:47Z","lastTransitionTime":"2026-03-13T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.404749 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.404802 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.404817 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.404840 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.404903 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:47Z","lastTransitionTime":"2026-03-13T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.508308 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.508347 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.508357 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.508373 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.508384 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:47Z","lastTransitionTime":"2026-03-13T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.612227 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.612691 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.612777 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.612911 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.613026 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:47Z","lastTransitionTime":"2026-03-13T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.715616 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.716020 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.716104 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.716183 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.716301 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:47Z","lastTransitionTime":"2026-03-13T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.820118 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.820260 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.820275 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.820300 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.820317 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:47Z","lastTransitionTime":"2026-03-13T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.924090 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.924151 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.924163 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.924181 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:47 crc kubenswrapper[4632]: I0313 10:05:47.924193 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:47Z","lastTransitionTime":"2026-03-13T10:05:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:48 crc kubenswrapper[4632]: E0313 10:05:48.025525 4632 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.044174 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.044266 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:48 crc kubenswrapper[4632]: E0313 10:05:48.044341 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.044441 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:48 crc kubenswrapper[4632]: E0313 10:05:48.044449 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.044490 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:48 crc kubenswrapper[4632]: E0313 10:05:48.044770 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:48 crc kubenswrapper[4632]: E0313 10:05:48.044647 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.064064 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.081932 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.094806 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.109121 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.134166 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0b158df3a1ade30707dc3ca7240a0945ad79d93f45888229313da4ded182c7d9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"message\\\":\\\"s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577081 6495 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0313 10:05:37.577192 6495 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.577561 6495 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0313 10:05:37.578699 6495 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0313 10:05:37.578717 6495 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0313 10:05:37.578749 6495 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0313 10:05:37.578783 6495 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0313 10:05:37.578851 6495 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0313 10:05:37.578884 6495 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0313 10:05:37.578917 6495 factory.go:656] Stopping watch factory\\\\nI0313 10:05:37.578961 6495 handler.go:208] Removed *v1.EgressFirewall ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"message\\\":\\\"{c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.681766 6736 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0313 10:05:39.682059 6736 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.682124 6736 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.682142 6736 ovnkube.go:599] Stopped ovnkube\\\\nI0313 10:05:39.682173 6736 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 10:05:39.682232 6736 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.149172 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: E0313 10:05:48.156347 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.166002 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.179730 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.197082 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"caab849a-f2dd-453b-85cc-768f57800789\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://951a71f1358e164fefc6fbe9404cf1d0e387d58b0bc1d060eeca64a85e0fc08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:18Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 10:03:50.495764 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 10:03:50.504436 1 observer_polling.go:159] Starting file observer\\\\nI0313 10:03:50.582400 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 10:03:50.592133 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 10:04:18.170496 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 10:04:18.170641 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:04:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc37f87081e3682692ee20f12c80aa65fbcb8604b381f313b8073f8019b96dbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2666e43569f4f3fb7d9deec4b70dc873d86a3a7c4ac2fac7eea45198e35ecf3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.211666 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.226294 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.240925 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.252991 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.265766 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.277305 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.290179 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:48 crc kubenswrapper[4632]: I0313 10:05:48.299647 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:48Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:50 crc kubenswrapper[4632]: I0313 10:05:50.044277 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:50 crc kubenswrapper[4632]: I0313 10:05:50.044362 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:50 crc kubenswrapper[4632]: E0313 10:05:50.044819 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:50 crc kubenswrapper[4632]: I0313 10:05:50.044421 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:50 crc kubenswrapper[4632]: E0313 10:05:50.044932 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:50 crc kubenswrapper[4632]: I0313 10:05:50.044370 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:50 crc kubenswrapper[4632]: E0313 10:05:50.045086 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:50 crc kubenswrapper[4632]: E0313 10:05:50.045153 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.274472 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.274535 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.274555 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.274577 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.274593 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:51Z","lastTransitionTime":"2026-03-13T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:51 crc kubenswrapper[4632]: E0313 10:05:51.288079 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:51Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.292430 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.292465 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.292476 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.292492 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.292502 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:51Z","lastTransitionTime":"2026-03-13T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:51 crc kubenswrapper[4632]: E0313 10:05:51.303563 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:51Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.308025 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.308072 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.308082 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.308099 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.308108 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:51Z","lastTransitionTime":"2026-03-13T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:51 crc kubenswrapper[4632]: E0313 10:05:51.321830 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:51Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.325311 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.325359 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.325369 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.325386 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.325397 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:51Z","lastTransitionTime":"2026-03-13T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:51 crc kubenswrapper[4632]: E0313 10:05:51.338048 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:51Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.341869 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.341916 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.341927 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.341972 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:05:51 crc kubenswrapper[4632]: I0313 10:05:51.341985 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:05:51Z","lastTransitionTime":"2026-03-13T10:05:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:05:51 crc kubenswrapper[4632]: E0313 10:05:51.357083 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:51Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:51 crc kubenswrapper[4632]: E0313 10:05:51.357248 4632 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 10:05:52 crc kubenswrapper[4632]: I0313 10:05:52.043996 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:52 crc kubenswrapper[4632]: E0313 10:05:52.044157 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:52 crc kubenswrapper[4632]: I0313 10:05:52.044209 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:52 crc kubenswrapper[4632]: I0313 10:05:52.044253 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:52 crc kubenswrapper[4632]: E0313 10:05:52.044315 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:52 crc kubenswrapper[4632]: I0313 10:05:52.044351 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:52 crc kubenswrapper[4632]: E0313 10:05:52.044395 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:52 crc kubenswrapper[4632]: E0313 10:05:52.044434 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:53 crc kubenswrapper[4632]: I0313 10:05:53.054692 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 13 10:05:53 crc kubenswrapper[4632]: E0313 10:05:53.157697 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:05:54 crc kubenswrapper[4632]: I0313 10:05:54.043279 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:54 crc kubenswrapper[4632]: I0313 10:05:54.043342 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:54 crc kubenswrapper[4632]: E0313 10:05:54.043460 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:54 crc kubenswrapper[4632]: I0313 10:05:54.043480 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:54 crc kubenswrapper[4632]: I0313 10:05:54.043289 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:54 crc kubenswrapper[4632]: E0313 10:05:54.043577 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:54 crc kubenswrapper[4632]: E0313 10:05:54.043675 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:54 crc kubenswrapper[4632]: E0313 10:05:54.043721 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.044171 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:56 crc kubenswrapper[4632]: E0313 10:05:56.044349 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.044345 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:56 crc kubenswrapper[4632]: E0313 10:05:56.044427 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.044516 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.044524 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:56 crc kubenswrapper[4632]: E0313 10:05:56.044806 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:56 crc kubenswrapper[4632]: E0313 10:05:56.045468 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.045530 4632 scope.go:117] "RemoveContainer" containerID="934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.065877 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.086020 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.101872 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.127512 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.140896 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3570521a-40ff-48d4-a6c2-ef53f64eca38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bd8f71a5e4bfa40758e7d51545f1b2eff43f11071060201770a574f89d391bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc4f274e9e301fdd8acbd720c998a2aa1c5e00df4fc254bcacab4acb539b8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.153910 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.165756 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.185468 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"message\\\":\\\"{c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.681766 6736 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0313 10:05:39.682059 6736 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.682124 6736 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.682142 6736 ovnkube.go:599] Stopped ovnkube\\\\nI0313 10:05:39.682173 6736 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 10:05:39.682232 6736 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.204374 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.220274 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.235195 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.253299 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"caab849a-f2dd-453b-85cc-768f57800789\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://951a71f1358e164fefc6fbe9404cf1d0e387d58b0bc1d060eeca64a85e0fc08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:18Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 10:03:50.495764 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 10:03:50.504436 1 observer_polling.go:159] Starting file observer\\\\nI0313 10:03:50.582400 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 10:03:50.592133 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 10:04:18.170496 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 10:04:18.170641 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:04:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc37f87081e3682692ee20f12c80aa65fbcb8604b381f313b8073f8019b96dbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2666e43569f4f3fb7d9deec4b70dc873d86a3a7c4ac2fac7eea45198e35ecf3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.277007 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.296288 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.319377 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.356711 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.376230 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:56 crc kubenswrapper[4632]: I0313 10:05:56.395273 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:56Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.122790 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/2.log" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.123485 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/1.log" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.125978 4632 generic.go:334] "Generic (PLEG): container finished" podID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerID="85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe" exitCode=1 Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.126014 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerDied","Data":"85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe"} Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.126047 4632 scope.go:117] "RemoveContainer" containerID="934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.126724 4632 scope.go:117] "RemoveContainer" containerID="85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe" Mar 13 10:05:57 crc kubenswrapper[4632]: E0313 10:05:57.126881 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.142529 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.156176 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.171090 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.186576 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.199594 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3570521a-40ff-48d4-a6c2-ef53f64eca38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bd8f71a5e4bfa40758e7d51545f1b2eff43f11071060201770a574f89d391bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc4f274e9e301fdd8acbd720c998a2aa1c5e00df4fc254bcacab4acb539b8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.212036 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.225009 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.237028 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.252734 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.267446 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.279690 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.291671 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.317489 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"message\\\":\\\"{c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.681766 6736 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0313 10:05:39.682059 6736 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.682124 6736 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.682142 6736 ovnkube.go:599] Stopped ovnkube\\\\nI0313 10:05:39.682173 6736 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 10:05:39.682232 6736 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:57Z\\\",\\\"message\\\":\\\":(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0313 10:05:57.066140 6990 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0313 10:05:57.066148 6990 services_controller.go:452] Built service openshift-marketplace/marketplace-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066155 6990 services_controller.go:453] Built service openshift-marketplace/marketplace-operator-metrics template LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066161 6990 services_controller.go:454] Service openshift-marketplace/marketplace-operator-metrics for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0313 10:05:57.066185 6990 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/api\\\\\\\"}\\\\nI0313 10:05:57.066204 6990 services_controller.go:360] Finished syncing service api on namespace openshift-apiserver for network=default : 1.29706ms\\\\nF0313 10:05:57.066212 6990 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.329142 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.343194 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.354854 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.367442 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"caab849a-f2dd-453b-85cc-768f57800789\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://951a71f1358e164fefc6fbe9404cf1d0e387d58b0bc1d060eeca64a85e0fc08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:18Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 10:03:50.495764 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 10:03:50.504436 1 observer_polling.go:159] Starting file observer\\\\nI0313 10:03:50.582400 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 10:03:50.592133 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 10:04:18.170496 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 10:04:18.170641 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:04:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc37f87081e3682692ee20f12c80aa65fbcb8604b381f313b8073f8019b96dbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2666e43569f4f3fb7d9deec4b70dc873d86a3a7c4ac2fac7eea45198e35ecf3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:57 crc kubenswrapper[4632]: I0313 10:05:57.383421 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:57Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.043580 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.043641 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.043695 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.043768 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:05:58 crc kubenswrapper[4632]: E0313 10:05:58.043764 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:05:58 crc kubenswrapper[4632]: E0313 10:05:58.043856 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:05:58 crc kubenswrapper[4632]: E0313 10:05:58.044112 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:05:58 crc kubenswrapper[4632]: E0313 10:05:58.044165 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.060553 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.071570 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.088424 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"caab849a-f2dd-453b-85cc-768f57800789\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://951a71f1358e164fefc6fbe9404cf1d0e387d58b0bc1d060eeca64a85e0fc08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:18Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 10:03:50.495764 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 10:03:50.504436 1 observer_polling.go:159] Starting file observer\\\\nI0313 10:03:50.582400 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 10:03:50.592133 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 10:04:18.170496 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 10:04:18.170641 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:04:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc37f87081e3682692ee20f12c80aa65fbcb8604b381f313b8073f8019b96dbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2666e43569f4f3fb7d9deec4b70dc873d86a3a7c4ac2fac7eea45198e35ecf3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.103601 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.119695 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.130346 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/2.log" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.136764 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.149703 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: E0313 10:05:58.158482 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.163268 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.174711 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3570521a-40ff-48d4-a6c2-ef53f64eca38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bd8f71a5e4bfa40758e7d51545f1b2eff43f11071060201770a574f89d391bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc4f274e9e301fdd8acbd720c998a2aa1c5e00df4fc254bcacab4acb539b8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.197555 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.209484 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.219047 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.231209 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.243928 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.253336 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.263025 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.279442 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"message\\\":\\\"{c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.681766 6736 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0313 10:05:39.682059 6736 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.682124 6736 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.682142 6736 ovnkube.go:599] Stopped ovnkube\\\\nI0313 10:05:39.682173 6736 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 10:05:39.682232 6736 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:57Z\\\",\\\"message\\\":\\\":(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0313 10:05:57.066140 6990 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0313 10:05:57.066148 6990 services_controller.go:452] Built service openshift-marketplace/marketplace-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066155 6990 services_controller.go:453] Built service openshift-marketplace/marketplace-operator-metrics template LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066161 6990 services_controller.go:454] Service openshift-marketplace/marketplace-operator-metrics for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0313 10:05:57.066185 6990 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/api\\\\\\\"}\\\\nI0313 10:05:57.066204 6990 services_controller.go:360] Finished syncing service api on namespace openshift-apiserver for network=default : 1.29706ms\\\\nF0313 10:05:57.066212 6990 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:05:58 crc kubenswrapper[4632]: I0313 10:05:58.289231 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:05:58Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:00 crc kubenswrapper[4632]: I0313 10:06:00.043661 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:00 crc kubenswrapper[4632]: I0313 10:06:00.043782 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:00 crc kubenswrapper[4632]: I0313 10:06:00.043692 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:00 crc kubenswrapper[4632]: E0313 10:06:00.043870 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:00 crc kubenswrapper[4632]: E0313 10:06:00.043976 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:00 crc kubenswrapper[4632]: E0313 10:06:00.044069 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:00 crc kubenswrapper[4632]: I0313 10:06:00.044305 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:00 crc kubenswrapper[4632]: E0313 10:06:00.044399 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.373268 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.373313 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.373324 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.373340 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.373350 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:01Z","lastTransitionTime":"2026-03-13T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:01 crc kubenswrapper[4632]: E0313 10:06:01.387202 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:01Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.390925 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.390966 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.390977 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.390990 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.390999 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:01Z","lastTransitionTime":"2026-03-13T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:01 crc kubenswrapper[4632]: E0313 10:06:01.404552 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:01Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.408502 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.408528 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.408536 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.408548 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.408556 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:01Z","lastTransitionTime":"2026-03-13T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:01 crc kubenswrapper[4632]: E0313 10:06:01.422692 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:01Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.427602 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.427630 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.427639 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.427651 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.427660 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:01Z","lastTransitionTime":"2026-03-13T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:01 crc kubenswrapper[4632]: E0313 10:06:01.439692 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:01Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.444427 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.444469 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.444481 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.444498 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:01 crc kubenswrapper[4632]: I0313 10:06:01.444512 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:01Z","lastTransitionTime":"2026-03-13T10:06:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:01 crc kubenswrapper[4632]: E0313 10:06:01.463523 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:01Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:01 crc kubenswrapper[4632]: E0313 10:06:01.463744 4632 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 10:06:02 crc kubenswrapper[4632]: I0313 10:06:02.043461 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:02 crc kubenswrapper[4632]: I0313 10:06:02.043461 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:02 crc kubenswrapper[4632]: I0313 10:06:02.043567 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:02 crc kubenswrapper[4632]: E0313 10:06:02.043636 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:02 crc kubenswrapper[4632]: I0313 10:06:02.043714 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:02 crc kubenswrapper[4632]: E0313 10:06:02.043829 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:02 crc kubenswrapper[4632]: E0313 10:06:02.043896 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:02 crc kubenswrapper[4632]: E0313 10:06:02.044024 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:03 crc kubenswrapper[4632]: E0313 10:06:03.159982 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:06:04 crc kubenswrapper[4632]: I0313 10:06:04.043130 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:04 crc kubenswrapper[4632]: I0313 10:06:04.043252 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:04 crc kubenswrapper[4632]: E0313 10:06:04.043603 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:04 crc kubenswrapper[4632]: I0313 10:06:04.043380 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:04 crc kubenswrapper[4632]: E0313 10:06:04.043930 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:04 crc kubenswrapper[4632]: I0313 10:06:04.043314 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:04 crc kubenswrapper[4632]: E0313 10:06:04.044254 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:04 crc kubenswrapper[4632]: E0313 10:06:04.043790 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:06 crc kubenswrapper[4632]: I0313 10:06:06.043851 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:06 crc kubenswrapper[4632]: I0313 10:06:06.043932 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:06 crc kubenswrapper[4632]: I0313 10:06:06.043893 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:06 crc kubenswrapper[4632]: E0313 10:06:06.044060 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:06 crc kubenswrapper[4632]: I0313 10:06:06.043856 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:06 crc kubenswrapper[4632]: E0313 10:06:06.044181 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:06 crc kubenswrapper[4632]: E0313 10:06:06.044261 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:06 crc kubenswrapper[4632]: E0313 10:06:06.044320 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.043639 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.043681 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.043733 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.043827 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:08 crc kubenswrapper[4632]: E0313 10:06:08.043928 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:08 crc kubenswrapper[4632]: E0313 10:06:08.044062 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:08 crc kubenswrapper[4632]: E0313 10:06:08.044160 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:08 crc kubenswrapper[4632]: E0313 10:06:08.044228 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.063904 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.076059 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.091362 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.104879 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3570521a-40ff-48d4-a6c2-ef53f64eca38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bd8f71a5e4bfa40758e7d51545f1b2eff43f11071060201770a574f89d391bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc4f274e9e301fdd8acbd720c998a2aa1c5e00df4fc254bcacab4acb539b8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.119359 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.131004 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.151022 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://934a538bc834077a7381421f37a69e4d28792692b3d6686b4c56e39c6561d79e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"message\\\":\\\"{c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.681766 6736 default_network_controller.go:776] Recording success event on pod openshift-network-operator/iptables-alerter-4ln5h\\\\nI0313 10:05:39.682059 6736 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Switch Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.682124 6736 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:c94130be-172c-477c-88c4-40cc7eba30fe}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {eb8eef51-1a8d-43f9-ae2e-3b2cc00ded60}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0313 10:05:39.682142 6736 ovnkube.go:599] Stopped ovnkube\\\\nI0313 10:05:39.682173 6736 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF0313 10:05:39.682232 6736 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:57Z\\\",\\\"message\\\":\\\":(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0313 10:05:57.066140 6990 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0313 10:05:57.066148 6990 services_controller.go:452] Built service openshift-marketplace/marketplace-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066155 6990 services_controller.go:453] Built service openshift-marketplace/marketplace-operator-metrics template LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066161 6990 services_controller.go:454] Service openshift-marketplace/marketplace-operator-metrics for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0313 10:05:57.066185 6990 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/api\\\\\\\"}\\\\nI0313 10:05:57.066204 6990 services_controller.go:360] Finished syncing service api on namespace openshift-apiserver for network=default : 1.29706ms\\\\nF0313 10:05:57.066212 6990 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: E0313 10:06:08.161181 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.168346 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.193575 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.210264 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.226477 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.246288 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"caab849a-f2dd-453b-85cc-768f57800789\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://951a71f1358e164fefc6fbe9404cf1d0e387d58b0bc1d060eeca64a85e0fc08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:18Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 10:03:50.495764 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 10:03:50.504436 1 observer_polling.go:159] Starting file observer\\\\nI0313 10:03:50.582400 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 10:03:50.592133 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 10:04:18.170496 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 10:04:18.170641 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:04:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc37f87081e3682692ee20f12c80aa65fbcb8604b381f313b8073f8019b96dbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2666e43569f4f3fb7d9deec4b70dc873d86a3a7c4ac2fac7eea45198e35ecf3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.262417 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.275898 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.290430 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.303849 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.318334 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:08 crc kubenswrapper[4632]: I0313 10:06:08.330839 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:08Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.043929 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:10 crc kubenswrapper[4632]: E0313 10:06:10.044145 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.043980 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:10 crc kubenswrapper[4632]: E0313 10:06:10.044236 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.043981 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.043835 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:10 crc kubenswrapper[4632]: E0313 10:06:10.045206 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:10 crc kubenswrapper[4632]: E0313 10:06:10.045303 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.045748 4632 scope.go:117] "RemoveContainer" containerID="85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe" Mar 13 10:06:10 crc kubenswrapper[4632]: E0313 10:06:10.046103 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.063833 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.084784 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.100632 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.117403 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.136285 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.152288 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3570521a-40ff-48d4-a6c2-ef53f64eca38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bd8f71a5e4bfa40758e7d51545f1b2eff43f11071060201770a574f89d391bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc4f274e9e301fdd8acbd720c998a2aa1c5e00df4fc254bcacab4acb539b8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.168346 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.182288 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.194973 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.207806 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.221545 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.232644 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.242462 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.259918 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:57Z\\\",\\\"message\\\":\\\":(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0313 10:05:57.066140 6990 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0313 10:05:57.066148 6990 services_controller.go:452] Built service openshift-marketplace/marketplace-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066155 6990 services_controller.go:453] Built service openshift-marketplace/marketplace-operator-metrics template LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066161 6990 services_controller.go:454] Service openshift-marketplace/marketplace-operator-metrics for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0313 10:05:57.066185 6990 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/api\\\\\\\"}\\\\nI0313 10:05:57.066204 6990 services_controller.go:360] Finished syncing service api on namespace openshift-apiserver for network=default : 1.29706ms\\\\nF0313 10:05:57.066212 6990 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.272692 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.289397 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.301245 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.316984 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"caab849a-f2dd-453b-85cc-768f57800789\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://951a71f1358e164fefc6fbe9404cf1d0e387d58b0bc1d060eeca64a85e0fc08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:18Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 10:03:50.495764 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 10:03:50.504436 1 observer_polling.go:159] Starting file observer\\\\nI0313 10:03:50.582400 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 10:03:50.592133 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 10:04:18.170496 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 10:04:18.170641 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:04:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc37f87081e3682692ee20f12c80aa65fbcb8604b381f313b8073f8019b96dbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2666e43569f4f3fb7d9deec4b70dc873d86a3a7c4ac2fac7eea45198e35ecf3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:10Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.490596 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:06:10 crc kubenswrapper[4632]: I0313 10:06:10.498047 4632 scope.go:117] "RemoveContainer" containerID="85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe" Mar 13 10:06:10 crc kubenswrapper[4632]: E0313 10:06:10.498300 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.525918 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.526352 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.526367 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.526386 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.526399 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:11Z","lastTransitionTime":"2026-03-13T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:11 crc kubenswrapper[4632]: E0313 10:06:11.541020 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:11Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.551500 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.551557 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.551869 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.551893 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.551914 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:11Z","lastTransitionTime":"2026-03-13T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:11 crc kubenswrapper[4632]: E0313 10:06:11.567353 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:11Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.571281 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.571343 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.571356 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.571371 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.571381 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:11Z","lastTransitionTime":"2026-03-13T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:11 crc kubenswrapper[4632]: E0313 10:06:11.582417 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:11Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.586478 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.586557 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.586573 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.586591 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.586621 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:11Z","lastTransitionTime":"2026-03-13T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:11 crc kubenswrapper[4632]: E0313 10:06:11.598206 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:11Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.602917 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.603003 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.603016 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.603035 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:11 crc kubenswrapper[4632]: I0313 10:06:11.603049 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:11Z","lastTransitionTime":"2026-03-13T10:06:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:11 crc kubenswrapper[4632]: E0313 10:06:11.616405 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:11Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:11 crc kubenswrapper[4632]: E0313 10:06:11.616526 4632 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 10:06:12 crc kubenswrapper[4632]: I0313 10:06:12.044306 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:12 crc kubenswrapper[4632]: I0313 10:06:12.044416 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:12 crc kubenswrapper[4632]: I0313 10:06:12.044451 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:12 crc kubenswrapper[4632]: E0313 10:06:12.044529 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:12 crc kubenswrapper[4632]: I0313 10:06:12.044663 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:12 crc kubenswrapper[4632]: E0313 10:06:12.044751 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:12 crc kubenswrapper[4632]: E0313 10:06:12.044890 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:12 crc kubenswrapper[4632]: E0313 10:06:12.045029 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:13 crc kubenswrapper[4632]: E0313 10:06:13.163488 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:06:14 crc kubenswrapper[4632]: I0313 10:06:14.043772 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.043955 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:14 crc kubenswrapper[4632]: I0313 10:06:14.043804 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:14 crc kubenswrapper[4632]: I0313 10:06:14.044009 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.044026 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:14 crc kubenswrapper[4632]: I0313 10:06:14.043776 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.044086 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.044304 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:14 crc kubenswrapper[4632]: I0313 10:06:14.050150 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:06:14 crc kubenswrapper[4632]: I0313 10:06:14.050379 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:14 crc kubenswrapper[4632]: I0313 10:06:14.050434 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.050500 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:18.050442728 +0000 UTC m=+212.072972901 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.050524 4632 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.050589 4632 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.050675 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:07:18.050654944 +0000 UTC m=+212.073185107 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.050906 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:07:18.050884781 +0000 UTC m=+212.073415024 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:06:14 crc kubenswrapper[4632]: I0313 10:06:14.151926 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:14 crc kubenswrapper[4632]: I0313 10:06:14.152008 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:14 crc kubenswrapper[4632]: I0313 10:06:14.152033 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.152225 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.152246 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.152259 4632 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.152295 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.152348 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.152366 4632 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.152319 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 10:07:18.152298344 +0000 UTC m=+212.174828477 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.152476 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 10:07:18.152448099 +0000 UTC m=+212.174978232 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.152550 4632 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:06:14 crc kubenswrapper[4632]: E0313 10:06:14.152588 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs podName:ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad nodeName:}" failed. No retries permitted until 2026-03-13 10:07:18.152575682 +0000 UTC m=+212.175106025 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs") pod "network-metrics-daemon-z2vlz" (UID: "ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.044189 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:16 crc kubenswrapper[4632]: E0313 10:06:16.044390 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.044513 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.044599 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:16 crc kubenswrapper[4632]: E0313 10:06:16.044728 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:16 crc kubenswrapper[4632]: E0313 10:06:16.044806 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.044406 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:16 crc kubenswrapper[4632]: E0313 10:06:16.044933 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.195606 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gqf22_4ec8e301-3037-4de0-94d2-32c49709660e/kube-multus/0.log" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.195658 4632 generic.go:334] "Generic (PLEG): container finished" podID="4ec8e301-3037-4de0-94d2-32c49709660e" containerID="0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d" exitCode=1 Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.195696 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gqf22" event={"ID":"4ec8e301-3037-4de0-94d2-32c49709660e","Type":"ContainerDied","Data":"0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d"} Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.199277 4632 scope.go:117] "RemoveContainer" containerID="0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.216533 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.237981 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.253598 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.288071 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:57Z\\\",\\\"message\\\":\\\":(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0313 10:05:57.066140 6990 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0313 10:05:57.066148 6990 services_controller.go:452] Built service openshift-marketplace/marketplace-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066155 6990 services_controller.go:453] Built service openshift-marketplace/marketplace-operator-metrics template LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066161 6990 services_controller.go:454] Service openshift-marketplace/marketplace-operator-metrics for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0313 10:05:57.066185 6990 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/api\\\\\\\"}\\\\nI0313 10:05:57.066204 6990 services_controller.go:360] Finished syncing service api on namespace openshift-apiserver for network=default : 1.29706ms\\\\nF0313 10:05:57.066212 6990 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.305529 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.324104 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.339203 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.353712 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"caab849a-f2dd-453b-85cc-768f57800789\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://951a71f1358e164fefc6fbe9404cf1d0e387d58b0bc1d060eeca64a85e0fc08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:18Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 10:03:50.495764 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 10:03:50.504436 1 observer_polling.go:159] Starting file observer\\\\nI0313 10:03:50.582400 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 10:03:50.592133 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 10:04:18.170496 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 10:04:18.170641 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:04:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc37f87081e3682692ee20f12c80aa65fbcb8604b381f313b8073f8019b96dbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2666e43569f4f3fb7d9deec4b70dc873d86a3a7c4ac2fac7eea45198e35ecf3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.372173 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.389415 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.406166 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.419473 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.433571 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:16Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:06:15Z\\\",\\\"message\\\":\\\"2026-03-13T10:05:29+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68d567cc-883d-42b1-94b2-7210d6f53151\\\\n2026-03-13T10:05:29+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68d567cc-883d-42b1-94b2-7210d6f53151 to /host/opt/cni/bin/\\\\n2026-03-13T10:05:30Z [verbose] multus-daemon started\\\\n2026-03-13T10:05:30Z [verbose] Readiness Indicator file check\\\\n2026-03-13T10:06:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.447536 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3570521a-40ff-48d4-a6c2-ef53f64eca38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bd8f71a5e4bfa40758e7d51545f1b2eff43f11071060201770a574f89d391bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc4f274e9e301fdd8acbd720c998a2aa1c5e00df4fc254bcacab4acb539b8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.460316 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.474889 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.486334 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:16 crc kubenswrapper[4632]: I0313 10:06:16.499989 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:16Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.200185 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gqf22_4ec8e301-3037-4de0-94d2-32c49709660e/kube-multus/0.log" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.200260 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gqf22" event={"ID":"4ec8e301-3037-4de0-94d2-32c49709660e","Type":"ContainerStarted","Data":"e48bcc5861bda7a15e45c892fa67ba73299d99e896f36f2cb68274a659ec5d34"} Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.216655 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.230038 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.241780 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.252033 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.269507 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:57Z\\\",\\\"message\\\":\\\":(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0313 10:05:57.066140 6990 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0313 10:05:57.066148 6990 services_controller.go:452] Built service openshift-marketplace/marketplace-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066155 6990 services_controller.go:453] Built service openshift-marketplace/marketplace-operator-metrics template LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066161 6990 services_controller.go:454] Service openshift-marketplace/marketplace-operator-metrics for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0313 10:05:57.066185 6990 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/api\\\\\\\"}\\\\nI0313 10:05:57.066204 6990 services_controller.go:360] Finished syncing service api on namespace openshift-apiserver for network=default : 1.29706ms\\\\nF0313 10:05:57.066212 6990 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.279607 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.292336 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.303858 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.315524 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"caab849a-f2dd-453b-85cc-768f57800789\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://951a71f1358e164fefc6fbe9404cf1d0e387d58b0bc1d060eeca64a85e0fc08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:18Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 10:03:50.495764 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 10:03:50.504436 1 observer_polling.go:159] Starting file observer\\\\nI0313 10:03:50.582400 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 10:03:50.592133 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 10:04:18.170496 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 10:04:18.170641 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:04:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc37f87081e3682692ee20f12c80aa65fbcb8604b381f313b8073f8019b96dbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2666e43569f4f3fb7d9deec4b70dc873d86a3a7c4ac2fac7eea45198e35ecf3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.328924 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e48bcc5861bda7a15e45c892fa67ba73299d99e896f36f2cb68274a659ec5d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:06:15Z\\\",\\\"message\\\":\\\"2026-03-13T10:05:29+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68d567cc-883d-42b1-94b2-7210d6f53151\\\\n2026-03-13T10:05:29+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68d567cc-883d-42b1-94b2-7210d6f53151 to /host/opt/cni/bin/\\\\n2026-03-13T10:05:30Z [verbose] multus-daemon started\\\\n2026-03-13T10:05:30Z [verbose] Readiness Indicator file check\\\\n2026-03-13T10:06:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.343186 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.355017 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.365999 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.377840 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.388163 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3570521a-40ff-48d4-a6c2-ef53f64eca38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bd8f71a5e4bfa40758e7d51545f1b2eff43f11071060201770a574f89d391bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc4f274e9e301fdd8acbd720c998a2aa1c5e00df4fc254bcacab4acb539b8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.400421 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.413492 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:17 crc kubenswrapper[4632]: I0313 10:06:17.422833 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:17Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.044187 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.044263 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.044225 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.044187 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:18 crc kubenswrapper[4632]: E0313 10:06:18.044365 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:18 crc kubenswrapper[4632]: E0313 10:06:18.044428 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:18 crc kubenswrapper[4632]: E0313 10:06:18.044529 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:18 crc kubenswrapper[4632]: E0313 10:06:18.044642 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.059868 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e48bcc5861bda7a15e45c892fa67ba73299d99e896f36f2cb68274a659ec5d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:06:15Z\\\",\\\"message\\\":\\\"2026-03-13T10:05:29+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68d567cc-883d-42b1-94b2-7210d6f53151\\\\n2026-03-13T10:05:29+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68d567cc-883d-42b1-94b2-7210d6f53151 to /host/opt/cni/bin/\\\\n2026-03-13T10:05:30Z [verbose] multus-daemon started\\\\n2026-03-13T10:05:30Z [verbose] Readiness Indicator file check\\\\n2026-03-13T10:06:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.080003 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.096021 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.109855 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.126038 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.138219 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3570521a-40ff-48d4-a6c2-ef53f64eca38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bd8f71a5e4bfa40758e7d51545f1b2eff43f11071060201770a574f89d391bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc4f274e9e301fdd8acbd720c998a2aa1c5e00df4fc254bcacab4acb539b8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.150604 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.163772 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: E0313 10:06:18.164256 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.175653 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.196584 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.209437 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.220215 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.232069 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.249393 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:57Z\\\",\\\"message\\\":\\\":(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0313 10:05:57.066140 6990 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0313 10:05:57.066148 6990 services_controller.go:452] Built service openshift-marketplace/marketplace-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066155 6990 services_controller.go:453] Built service openshift-marketplace/marketplace-operator-metrics template LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066161 6990 services_controller.go:454] Service openshift-marketplace/marketplace-operator-metrics for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0313 10:05:57.066185 6990 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/api\\\\\\\"}\\\\nI0313 10:05:57.066204 6990 services_controller.go:360] Finished syncing service api on namespace openshift-apiserver for network=default : 1.29706ms\\\\nF0313 10:05:57.066212 6990 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.260846 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.274638 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.286976 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:18 crc kubenswrapper[4632]: I0313 10:06:18.299892 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"caab849a-f2dd-453b-85cc-768f57800789\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://951a71f1358e164fefc6fbe9404cf1d0e387d58b0bc1d060eeca64a85e0fc08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:18Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 10:03:50.495764 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 10:03:50.504436 1 observer_polling.go:159] Starting file observer\\\\nI0313 10:03:50.582400 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 10:03:50.592133 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 10:04:18.170496 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 10:04:18.170641 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:04:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc37f87081e3682692ee20f12c80aa65fbcb8604b381f313b8073f8019b96dbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2666e43569f4f3fb7d9deec4b70dc873d86a3a7c4ac2fac7eea45198e35ecf3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:18Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:20 crc kubenswrapper[4632]: I0313 10:06:20.043829 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:20 crc kubenswrapper[4632]: I0313 10:06:20.043913 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:20 crc kubenswrapper[4632]: I0313 10:06:20.043865 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:20 crc kubenswrapper[4632]: I0313 10:06:20.044020 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:20 crc kubenswrapper[4632]: E0313 10:06:20.044094 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:20 crc kubenswrapper[4632]: E0313 10:06:20.044287 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:20 crc kubenswrapper[4632]: E0313 10:06:20.044385 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:20 crc kubenswrapper[4632]: E0313 10:06:20.044484 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.702598 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.702656 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.702674 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.702698 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.702715 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:21Z","lastTransitionTime":"2026-03-13T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:21 crc kubenswrapper[4632]: E0313 10:06:21.724177 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:21Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.730251 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.730312 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.730329 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.730368 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.730385 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:21Z","lastTransitionTime":"2026-03-13T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:21 crc kubenswrapper[4632]: E0313 10:06:21.746039 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:21Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.750641 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.750675 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.750683 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.750697 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.750715 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:21Z","lastTransitionTime":"2026-03-13T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:21 crc kubenswrapper[4632]: E0313 10:06:21.768146 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:21Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.773442 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.773498 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.773515 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.773538 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.773555 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:21Z","lastTransitionTime":"2026-03-13T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:21 crc kubenswrapper[4632]: E0313 10:06:21.788096 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:21Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.792637 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.792681 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.792693 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.792708 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:21 crc kubenswrapper[4632]: I0313 10:06:21.792717 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:21Z","lastTransitionTime":"2026-03-13T10:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:21 crc kubenswrapper[4632]: E0313 10:06:21.804894 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:21Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:21 crc kubenswrapper[4632]: E0313 10:06:21.805045 4632 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 10:06:22 crc kubenswrapper[4632]: I0313 10:06:22.043959 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:22 crc kubenswrapper[4632]: I0313 10:06:22.044075 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:22 crc kubenswrapper[4632]: I0313 10:06:22.044128 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:22 crc kubenswrapper[4632]: I0313 10:06:22.044034 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:22 crc kubenswrapper[4632]: E0313 10:06:22.044327 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:22 crc kubenswrapper[4632]: E0313 10:06:22.044493 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:22 crc kubenswrapper[4632]: E0313 10:06:22.044586 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:22 crc kubenswrapper[4632]: E0313 10:06:22.044674 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:23 crc kubenswrapper[4632]: E0313 10:06:23.165701 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:06:24 crc kubenswrapper[4632]: I0313 10:06:24.044232 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:24 crc kubenswrapper[4632]: I0313 10:06:24.044353 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:24 crc kubenswrapper[4632]: I0313 10:06:24.044298 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:24 crc kubenswrapper[4632]: E0313 10:06:24.044567 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:24 crc kubenswrapper[4632]: E0313 10:06:24.044491 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:24 crc kubenswrapper[4632]: E0313 10:06:24.044730 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:24 crc kubenswrapper[4632]: I0313 10:06:24.044345 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:24 crc kubenswrapper[4632]: E0313 10:06:24.045264 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.043812 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.043917 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.044008 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.044077 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:26 crc kubenswrapper[4632]: E0313 10:06:26.044615 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:26 crc kubenswrapper[4632]: E0313 10:06:26.044917 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:26 crc kubenswrapper[4632]: E0313 10:06:26.045058 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:26 crc kubenswrapper[4632]: E0313 10:06:26.045168 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.045744 4632 scope.go:117] "RemoveContainer" containerID="85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.231697 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/2.log" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.233789 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerStarted","Data":"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b"} Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.234230 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.251392 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"caab849a-f2dd-453b-85cc-768f57800789\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://951a71f1358e164fefc6fbe9404cf1d0e387d58b0bc1d060eeca64a85e0fc08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:18Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 10:03:50.495764 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 10:03:50.504436 1 observer_polling.go:159] Starting file observer\\\\nI0313 10:03:50.582400 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 10:03:50.592133 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 10:04:18.170496 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 10:04:18.170641 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:04:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc37f87081e3682692ee20f12c80aa65fbcb8604b381f313b8073f8019b96dbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2666e43569f4f3fb7d9deec4b70dc873d86a3a7c4ac2fac7eea45198e35ecf3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.265337 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.276166 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.287268 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.300599 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e48bcc5861bda7a15e45c892fa67ba73299d99e896f36f2cb68274a659ec5d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:06:15Z\\\",\\\"message\\\":\\\"2026-03-13T10:05:29+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68d567cc-883d-42b1-94b2-7210d6f53151\\\\n2026-03-13T10:05:29+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68d567cc-883d-42b1-94b2-7210d6f53151 to /host/opt/cni/bin/\\\\n2026-03-13T10:05:30Z [verbose] multus-daemon started\\\\n2026-03-13T10:05:30Z [verbose] Readiness Indicator file check\\\\n2026-03-13T10:06:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.320831 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.335588 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.346114 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.358254 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.372700 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3570521a-40ff-48d4-a6c2-ef53f64eca38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bd8f71a5e4bfa40758e7d51545f1b2eff43f11071060201770a574f89d391bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc4f274e9e301fdd8acbd720c998a2aa1c5e00df4fc254bcacab4acb539b8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.384761 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.396272 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.414407 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:57Z\\\",\\\"message\\\":\\\":(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0313 10:05:57.066140 6990 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0313 10:05:57.066148 6990 services_controller.go:452] Built service openshift-marketplace/marketplace-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066155 6990 services_controller.go:453] Built service openshift-marketplace/marketplace-operator-metrics template LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066161 6990 services_controller.go:454] Service openshift-marketplace/marketplace-operator-metrics for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0313 10:05:57.066185 6990 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/api\\\\\\\"}\\\\nI0313 10:05:57.066204 6990 services_controller.go:360] Finished syncing service api on namespace openshift-apiserver for network=default : 1.29706ms\\\\nF0313 10:05:57.066212 6990 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.425771 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.440018 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.454303 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.469507 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:26 crc kubenswrapper[4632]: I0313 10:06:26.482008 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:26Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.043596 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:28 crc kubenswrapper[4632]: E0313 10:06:28.044070 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.044087 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:28 crc kubenswrapper[4632]: E0313 10:06:28.044209 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.043626 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.043893 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:28 crc kubenswrapper[4632]: E0313 10:06:28.044277 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:28 crc kubenswrapper[4632]: E0313 10:06:28.044320 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.057726 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-q7bcc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-z2vlz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.071988 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"894cdc70-0747-4975-a22f-0dbd657e91a3\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:51Z\\\",\\\"message\\\":\\\"le observer\\\\nW0313 10:04:51.224345 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0313 10:04:51.224743 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0313 10:04:51.231279 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3511042304/tls.crt::/tmp/serving-cert-3511042304/tls.key\\\\\\\"\\\\nI0313 10:04:51.512346 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0313 10:04:51.516641 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0313 10:04:51.516667 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0313 10:04:51.516694 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0313 10:04:51.516704 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0313 10:04:51.524417 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0313 10:04:51.524471 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524478 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0313 10:04:51.524490 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0313 10:04:51.524495 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0313 10:04:51.524500 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0313 10:04:51.524505 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0313 10:04:51.524633 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0313 10:04:51.529203 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:04:50Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.084587 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3570521a-40ff-48d4-a6c2-ef53f64eca38\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://9bd8f71a5e4bfa40758e7d51545f1b2eff43f11071060201770a574f89d391bc\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ecc4f274e9e301fdd8acbd720c998a2aa1c5e00df4fc254bcacab4acb539b8ce\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://452fd4724dbce515315ce137b8477166e605db76fec46d7b0ab23756c3cf1c52\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.100452 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.114440 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dcdb8c440e2f83530ee3fd4be3d39b21575674047ecba0e26719eb06bde38dbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://39037d4d6f8345706dd36a2591ead2b263a8f0bcc4558684d2c06a5eccfbfbf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.141508 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b40c6b3-0061-4224-82d5-3ccf67998722\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:05:57Z\\\",\\\"message\\\":\\\":(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0313 10:05:57.066140 6990 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0313 10:05:57.066148 6990 services_controller.go:452] Built service openshift-marketplace/marketplace-operator-metrics per-node LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066155 6990 services_controller.go:453] Built service openshift-marketplace/marketplace-operator-metrics template LB for network=default: []services.LB{}\\\\nI0313 10:05:57.066161 6990 services_controller.go:454] Service openshift-marketplace/marketplace-operator-metrics for network=default has 2 cluster-wide, 0 per-node configs, 0 template configs, making 1 (cluster) 0 (per node) and 0 (template) load balancers\\\\nI0313 10:05:57.066185 6990 loadbalancer.go:304] Deleted 0 stale LBs for map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-apiserver/api\\\\\\\"}\\\\nI0313 10:05:57.066204 6990 services_controller.go:360] Finished syncing service api on namespace openshift-apiserver for network=default : 1.29706ms\\\\nF0313 10:05:57.066212 6990 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:56Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:06:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dj6cl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-qb725\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.151896 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ceeb823b-9a8c-403a-9a60-1d74ba0fbffe\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://da62efee0b0e5a420abd2f18aae7b1ad532b0dd5dcda4d36d4efa9b039bd8811\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3129c4f8116cd2f517b7c37b8d2af3f30c4e028d18bacf2f99042ec954b60611\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:03:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.166853 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: E0313 10:06:28.167072 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.182073 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.198263 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-n55jt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b29b9ad7-8cc9-434f-8731-a86265c383fd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f8c3d39bd1ac3290c35d993771da5d7915b468b7376aadccf7b459f65dc7138\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-pgn9f\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-n55jt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.212072 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-zwlc8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9a50974e-f938-40f7-ace5-2a3b4cb1f3e7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a0cdf9ec894d29357b7b171dd1552444b97348f86d71e58afd9ab6bae1a05654\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mq5zl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-zwlc8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.226678 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"caab849a-f2dd-453b-85cc-768f57800789\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:04:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:03:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://951a71f1358e164fefc6fbe9404cf1d0e387d58b0bc1d060eeca64a85e0fc08e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://59e800b40d3f99e6d38fa7e28f06be6260e6ebe4bb5e7c73de6734d0617092ac\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-13T10:04:18Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0313 10:03:50.495764 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0313 10:03:50.504436 1 observer_polling.go:159] Starting file observer\\\\nI0313 10:03:50.582400 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0313 10:03:50.592133 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0313 10:04:18.170496 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0313 10:04:18.170641 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:04:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc37f87081e3682692ee20f12c80aa65fbcb8604b381f313b8073f8019b96dbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2666e43569f4f3fb7d9deec4b70dc873d86a3a7c4ac2fac7eea45198e35ecf3a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:03:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:03:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.243229 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1b3ab26f24c338001da561ca80dccbab8a99da54054f6b7ddd2b0d5ac02f7dfa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.257612 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d77b18a7-7ad9-4bf5-bff5-da45878af7f4\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e74ac702432406b147f9db010a8945c3e54f6f15a64346e895833c39bcc8f6d9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vnh6w\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-zkscb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.272002 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:27Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be8dc5225d33306dd28d6043b50b88e39a677356681187c0f881966e12e9494c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.288502 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-gqf22" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4ec8e301-3037-4de0-94d2-32c49709660e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e48bcc5861bda7a15e45c892fa67ba73299d99e896f36f2cb68274a659ec5d34\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-13T10:06:15Z\\\",\\\"message\\\":\\\"2026-03-13T10:05:29+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_68d567cc-883d-42b1-94b2-7210d6f53151\\\\n2026-03-13T10:05:29+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_68d567cc-883d-42b1-94b2-7210d6f53151 to /host/opt/cni/bin/\\\\n2026-03-13T10:05:30Z [verbose] multus-daemon started\\\\n2026-03-13T10:05:30Z [verbose] Readiness Indicator file check\\\\n2026-03-13T10:06:15Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:06:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-8d5c4\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-gqf22\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.305394 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b054ca08-1d09-4eca-a608-eb5b9323959a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1443078dbf73aa095c42c9554ca6cfd14a0b7b3b58fdf9f9ca610ea8ae4faaf3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a5751cc23a251bd0c422190506273ba218a1268ad61dde76e71ca7fca7db0e35\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2aec1c6df1b2f1a979a083b85e27dedfee5df54edcd6072534e3e6336f0791d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b11c0fe86b3039760a8f2421d22ecdbf9a435ae7afbec2af53c04d7b7326655e\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:31Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4dc94caceb393a1194da38ccf2d8cb2847902c15e49246643d4c2fbd259f71a9\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:33Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8c5c8753ebd5c621b714f9ad16bfcbeb42cb5b41a65a0bf8262e9d9a861e0a09\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://be5cd3ab03fb352008bc7646fa1f6f5f4825c107aef63eb0697721666d30cd41\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T10:05:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T10:05:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l9s5s\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-qlc8m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:28 crc kubenswrapper[4632]: I0313 10:06:28.319204 4632 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b0c542d5-8c38-4243-8af7-cfc0d8e22773\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T10:05:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://55b54038392eeeed9eefaa615e0f9a4a34341efebcd6f6e628d7a911545ca154\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a71f678e4bdb921e73ed7f26a9d4565fc5668c94082003bb79357a2bf793b06f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T10:05:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-ffbwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T10:05:10Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-kbtt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:28Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:30 crc kubenswrapper[4632]: I0313 10:06:30.043512 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:30 crc kubenswrapper[4632]: I0313 10:06:30.043665 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:30 crc kubenswrapper[4632]: E0313 10:06:30.043783 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:30 crc kubenswrapper[4632]: I0313 10:06:30.043540 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:30 crc kubenswrapper[4632]: E0313 10:06:30.043868 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:30 crc kubenswrapper[4632]: I0313 10:06:30.043510 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:30 crc kubenswrapper[4632]: E0313 10:06:30.044263 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:30 crc kubenswrapper[4632]: E0313 10:06:30.044341 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.005534 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.005601 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.005615 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.005635 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.005650 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:32Z","lastTransitionTime":"2026-03-13T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:32 crc kubenswrapper[4632]: E0313 10:06:32.022001 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.026093 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.026153 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.026164 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.026184 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.026196 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:32Z","lastTransitionTime":"2026-03-13T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:32 crc kubenswrapper[4632]: E0313 10:06:32.039869 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.043512 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.043665 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.043723 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.043810 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:32 crc kubenswrapper[4632]: E0313 10:06:32.043853 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.044013 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.044039 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.044050 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.044063 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.044075 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:32Z","lastTransitionTime":"2026-03-13T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:32 crc kubenswrapper[4632]: E0313 10:06:32.044128 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:32 crc kubenswrapper[4632]: E0313 10:06:32.043968 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:32 crc kubenswrapper[4632]: E0313 10:06:32.044216 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:32 crc kubenswrapper[4632]: E0313 10:06:32.058892 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.061994 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.062026 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.062038 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.062055 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.062065 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:32Z","lastTransitionTime":"2026-03-13T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:32 crc kubenswrapper[4632]: E0313 10:06:32.074338 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.077990 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.078021 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.078031 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.078044 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:32 crc kubenswrapper[4632]: I0313 10:06:32.078053 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:32Z","lastTransitionTime":"2026-03-13T10:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:32 crc kubenswrapper[4632]: E0313 10:06:32.088530 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"7800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"24148068Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"8\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"24608868Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T10:06:32Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"b5d63e17-4c81-494f-81b9-40163ac26c6b\\\",\\\"systemUUID\\\":\\\"e8be0c8f-16ef-4a1d-b190-772a9f649bc5\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T10:06:32Z is after 2025-08-24T17:21:41Z" Mar 13 10:06:32 crc kubenswrapper[4632]: E0313 10:06:32.088637 4632 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 10:06:33 crc kubenswrapper[4632]: I0313 10:06:33.060722 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 13 10:06:33 crc kubenswrapper[4632]: E0313 10:06:33.168803 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:06:34 crc kubenswrapper[4632]: I0313 10:06:34.044140 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:34 crc kubenswrapper[4632]: E0313 10:06:34.044282 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:34 crc kubenswrapper[4632]: I0313 10:06:34.044299 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:34 crc kubenswrapper[4632]: E0313 10:06:34.044461 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:34 crc kubenswrapper[4632]: I0313 10:06:34.044165 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:34 crc kubenswrapper[4632]: E0313 10:06:34.044675 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:34 crc kubenswrapper[4632]: I0313 10:06:34.044807 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:34 crc kubenswrapper[4632]: E0313 10:06:34.044932 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:36 crc kubenswrapper[4632]: I0313 10:06:36.044106 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:36 crc kubenswrapper[4632]: I0313 10:06:36.045058 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:36 crc kubenswrapper[4632]: I0313 10:06:36.044126 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:36 crc kubenswrapper[4632]: I0313 10:06:36.044106 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:36 crc kubenswrapper[4632]: E0313 10:06:36.045120 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:36 crc kubenswrapper[4632]: E0313 10:06:36.045199 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:36 crc kubenswrapper[4632]: E0313 10:06:36.045291 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:36 crc kubenswrapper[4632]: E0313 10:06:36.045362 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.044314 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.044412 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.044374 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.044317 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:38 crc kubenswrapper[4632]: E0313 10:06:38.044623 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:38 crc kubenswrapper[4632]: E0313 10:06:38.044697 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:38 crc kubenswrapper[4632]: E0313 10:06:38.044869 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:38 crc kubenswrapper[4632]: E0313 10:06:38.044977 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.111477 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-qlc8m" podStartSLOduration=119.111456792 podStartE2EDuration="1m59.111456792s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:06:38.087376923 +0000 UTC m=+172.109907076" watchObservedRunningTime="2026-03-13 10:06:38.111456792 +0000 UTC m=+172.133986925" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.126323 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-kbtt2" podStartSLOduration=117.126300993 podStartE2EDuration="1m57.126300993s" podCreationTimestamp="2026-03-13 10:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:06:38.111874815 +0000 UTC m=+172.134404948" watchObservedRunningTime="2026-03-13 10:06:38.126300993 +0000 UTC m=+172.148831126" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.163168 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-gqf22" podStartSLOduration=119.163149259 podStartE2EDuration="1m59.163149259s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:06:38.162526729 +0000 UTC m=+172.185056862" watchObservedRunningTime="2026-03-13 10:06:38.163149259 +0000 UTC m=+172.185679402" Mar 13 10:06:38 crc kubenswrapper[4632]: E0313 10:06:38.169787 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.210247 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=45.210223652 podStartE2EDuration="45.210223652s" podCreationTimestamp="2026-03-13 10:05:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:06:38.18958499 +0000 UTC m=+172.212115123" watchObservedRunningTime="2026-03-13 10:06:38.210223652 +0000 UTC m=+172.232753785" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.275338 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=5.275319785 podStartE2EDuration="5.275319785s" podCreationTimestamp="2026-03-13 10:06:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:06:38.273825799 +0000 UTC m=+172.296355952" watchObservedRunningTime="2026-03-13 10:06:38.275319785 +0000 UTC m=+172.297849928" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.294542 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=87.294525392 podStartE2EDuration="1m27.294525392s" podCreationTimestamp="2026-03-13 10:05:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:06:38.294318185 +0000 UTC m=+172.316848318" watchObservedRunningTime="2026-03-13 10:06:38.294525392 +0000 UTC m=+172.317055545" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.319551 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-n55jt" podStartSLOduration=119.319533369 podStartE2EDuration="1m59.319533369s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:06:38.318502738 +0000 UTC m=+172.341032881" watchObservedRunningTime="2026-03-13 10:06:38.319533369 +0000 UTC m=+172.342063512" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.356779 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podStartSLOduration=118.356762647 podStartE2EDuration="1m58.356762647s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:06:38.356296591 +0000 UTC m=+172.378826744" watchObservedRunningTime="2026-03-13 10:06:38.356762647 +0000 UTC m=+172.379292780" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.357023 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-zwlc8" podStartSLOduration=119.357018584 podStartE2EDuration="1m59.357018584s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:06:38.328674864 +0000 UTC m=+172.351205007" watchObservedRunningTime="2026-03-13 10:06:38.357018584 +0000 UTC m=+172.379548717" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.367646 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=81.367636045 podStartE2EDuration="1m21.367636045s" podCreationTimestamp="2026-03-13 10:05:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:06:38.367552052 +0000 UTC m=+172.390082185" watchObservedRunningTime="2026-03-13 10:06:38.367636045 +0000 UTC m=+172.390166188" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.389485 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podStartSLOduration=119.389461452 podStartE2EDuration="1m59.389461452s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:06:38.389312558 +0000 UTC m=+172.411842701" watchObservedRunningTime="2026-03-13 10:06:38.389461452 +0000 UTC m=+172.411991595" Mar 13 10:06:38 crc kubenswrapper[4632]: I0313 10:06:38.420570 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=51.420549709 podStartE2EDuration="51.420549709s" podCreationTimestamp="2026-03-13 10:05:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:06:38.407854784 +0000 UTC m=+172.430384927" watchObservedRunningTime="2026-03-13 10:06:38.420549709 +0000 UTC m=+172.443079842" Mar 13 10:06:40 crc kubenswrapper[4632]: I0313 10:06:40.043225 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:40 crc kubenswrapper[4632]: I0313 10:06:40.043262 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:40 crc kubenswrapper[4632]: I0313 10:06:40.043262 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:40 crc kubenswrapper[4632]: E0313 10:06:40.043386 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:40 crc kubenswrapper[4632]: I0313 10:06:40.043421 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:40 crc kubenswrapper[4632]: E0313 10:06:40.043540 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:40 crc kubenswrapper[4632]: E0313 10:06:40.043613 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:40 crc kubenswrapper[4632]: E0313 10:06:40.043672 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:40 crc kubenswrapper[4632]: I0313 10:06:40.514540 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" probeResult="failure" output="" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.044072 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.044131 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.044134 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.044132 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:42 crc kubenswrapper[4632]: E0313 10:06:42.044343 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:42 crc kubenswrapper[4632]: E0313 10:06:42.044397 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:42 crc kubenswrapper[4632]: E0313 10:06:42.044569 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:42 crc kubenswrapper[4632]: E0313 10:06:42.044728 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.105178 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.105223 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.105232 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.105246 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.105254 4632 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T10:06:42Z","lastTransitionTime":"2026-03-13T10:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.154109 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz"] Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.154541 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.159744 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.159971 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.159977 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.160242 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.281155 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.281309 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.281394 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.281451 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.281555 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.363501 4632 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.373868 4632 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.382221 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.382267 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.382303 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.382329 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.382355 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.382405 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.382487 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.383242 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-service-ca\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.387893 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.406034 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9d64d959-3f9f-43eb-b37f-79c8ec6c38bd-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-5lrdz\" (UID: \"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: I0313 10:06:42.471381 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" Mar 13 10:06:42 crc kubenswrapper[4632]: W0313 10:06:42.484905 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d64d959_3f9f_43eb_b37f_79c8ec6c38bd.slice/crio-71750ded659d29caf6b277a47cbf881699484dc3bfd621c59f2723f533ebae09 WatchSource:0}: Error finding container 71750ded659d29caf6b277a47cbf881699484dc3bfd621c59f2723f533ebae09: Status 404 returned error can't find the container with id 71750ded659d29caf6b277a47cbf881699484dc3bfd621c59f2723f533ebae09 Mar 13 10:06:43 crc kubenswrapper[4632]: E0313 10:06:43.171188 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:06:43 crc kubenswrapper[4632]: I0313 10:06:43.388972 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" event={"ID":"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd","Type":"ContainerStarted","Data":"5a8ef2bbbb0db2a9f2a7bfaaba16dd25db7a7a570f20b3976fb4a807b682cbde"} Mar 13 10:06:43 crc kubenswrapper[4632]: I0313 10:06:43.389325 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" event={"ID":"9d64d959-3f9f-43eb-b37f-79c8ec6c38bd","Type":"ContainerStarted","Data":"71750ded659d29caf6b277a47cbf881699484dc3bfd621c59f2723f533ebae09"} Mar 13 10:06:43 crc kubenswrapper[4632]: I0313 10:06:43.409810 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-5lrdz" podStartSLOduration=124.409777729 podStartE2EDuration="2m4.409777729s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:06:43.408747407 +0000 UTC m=+177.431277560" watchObservedRunningTime="2026-03-13 10:06:43.409777729 +0000 UTC m=+177.432307862" Mar 13 10:06:44 crc kubenswrapper[4632]: I0313 10:06:44.043817 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:44 crc kubenswrapper[4632]: I0313 10:06:44.043914 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:44 crc kubenswrapper[4632]: E0313 10:06:44.044046 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:44 crc kubenswrapper[4632]: I0313 10:06:44.044126 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:44 crc kubenswrapper[4632]: E0313 10:06:44.044149 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:44 crc kubenswrapper[4632]: I0313 10:06:44.043914 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:44 crc kubenswrapper[4632]: E0313 10:06:44.044404 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:44 crc kubenswrapper[4632]: E0313 10:06:44.044479 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:46 crc kubenswrapper[4632]: I0313 10:06:46.044274 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:46 crc kubenswrapper[4632]: I0313 10:06:46.044342 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:46 crc kubenswrapper[4632]: I0313 10:06:46.044418 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:46 crc kubenswrapper[4632]: I0313 10:06:46.045138 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:46 crc kubenswrapper[4632]: E0313 10:06:46.045361 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:46 crc kubenswrapper[4632]: E0313 10:06:46.045824 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:46 crc kubenswrapper[4632]: E0313 10:06:46.046164 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:46 crc kubenswrapper[4632]: E0313 10:06:46.046451 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:47 crc kubenswrapper[4632]: I0313 10:06:47.407340 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/3.log" Mar 13 10:06:47 crc kubenswrapper[4632]: I0313 10:06:47.408028 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/2.log" Mar 13 10:06:47 crc kubenswrapper[4632]: I0313 10:06:47.410617 4632 generic.go:334] "Generic (PLEG): container finished" podID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerID="8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b" exitCode=1 Mar 13 10:06:47 crc kubenswrapper[4632]: I0313 10:06:47.410663 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerDied","Data":"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b"} Mar 13 10:06:47 crc kubenswrapper[4632]: I0313 10:06:47.410697 4632 scope.go:117] "RemoveContainer" containerID="85cd256cfb44a7c0f0961b308f26a8fdd3f76e9371a32d157163576e9eec7dfe" Mar 13 10:06:47 crc kubenswrapper[4632]: I0313 10:06:47.412197 4632 scope.go:117] "RemoveContainer" containerID="8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b" Mar 13 10:06:47 crc kubenswrapper[4632]: E0313 10:06:47.412370 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" Mar 13 10:06:48 crc kubenswrapper[4632]: I0313 10:06:48.043256 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:48 crc kubenswrapper[4632]: I0313 10:06:48.043276 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:48 crc kubenswrapper[4632]: I0313 10:06:48.043335 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:48 crc kubenswrapper[4632]: I0313 10:06:48.043352 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:48 crc kubenswrapper[4632]: E0313 10:06:48.045462 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:48 crc kubenswrapper[4632]: E0313 10:06:48.045646 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:48 crc kubenswrapper[4632]: E0313 10:06:48.045779 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:48 crc kubenswrapper[4632]: E0313 10:06:48.045983 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:48 crc kubenswrapper[4632]: E0313 10:06:48.172659 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:06:48 crc kubenswrapper[4632]: I0313 10:06:48.415233 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/3.log" Mar 13 10:06:50 crc kubenswrapper[4632]: I0313 10:06:50.043413 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:50 crc kubenswrapper[4632]: I0313 10:06:50.043482 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:50 crc kubenswrapper[4632]: E0313 10:06:50.043589 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:50 crc kubenswrapper[4632]: I0313 10:06:50.043663 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:50 crc kubenswrapper[4632]: E0313 10:06:50.043790 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:50 crc kubenswrapper[4632]: I0313 10:06:50.043426 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:50 crc kubenswrapper[4632]: E0313 10:06:50.043903 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:50 crc kubenswrapper[4632]: E0313 10:06:50.044020 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:52 crc kubenswrapper[4632]: I0313 10:06:52.043615 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:52 crc kubenswrapper[4632]: I0313 10:06:52.043773 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:52 crc kubenswrapper[4632]: E0313 10:06:52.043788 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:52 crc kubenswrapper[4632]: I0313 10:06:52.043926 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:52 crc kubenswrapper[4632]: I0313 10:06:52.044021 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:52 crc kubenswrapper[4632]: E0313 10:06:52.044045 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:52 crc kubenswrapper[4632]: E0313 10:06:52.044174 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:52 crc kubenswrapper[4632]: E0313 10:06:52.044302 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:53 crc kubenswrapper[4632]: E0313 10:06:53.174466 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:06:54 crc kubenswrapper[4632]: I0313 10:06:54.043806 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:54 crc kubenswrapper[4632]: I0313 10:06:54.044021 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:54 crc kubenswrapper[4632]: I0313 10:06:54.043894 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:54 crc kubenswrapper[4632]: I0313 10:06:54.044015 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:54 crc kubenswrapper[4632]: E0313 10:06:54.044218 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:54 crc kubenswrapper[4632]: E0313 10:06:54.044261 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:54 crc kubenswrapper[4632]: E0313 10:06:54.044405 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:54 crc kubenswrapper[4632]: E0313 10:06:54.044531 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:56 crc kubenswrapper[4632]: I0313 10:06:56.043995 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:56 crc kubenswrapper[4632]: I0313 10:06:56.044032 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:56 crc kubenswrapper[4632]: E0313 10:06:56.045235 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:56 crc kubenswrapper[4632]: I0313 10:06:56.044101 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:56 crc kubenswrapper[4632]: E0313 10:06:56.045569 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:56 crc kubenswrapper[4632]: E0313 10:06:56.045389 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:56 crc kubenswrapper[4632]: I0313 10:06:56.044065 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:56 crc kubenswrapper[4632]: E0313 10:06:56.045848 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:58 crc kubenswrapper[4632]: I0313 10:06:58.044158 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:06:58 crc kubenswrapper[4632]: I0313 10:06:58.044164 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:06:58 crc kubenswrapper[4632]: I0313 10:06:58.044178 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:06:58 crc kubenswrapper[4632]: I0313 10:06:58.044195 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:06:58 crc kubenswrapper[4632]: E0313 10:06:58.045280 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:06:58 crc kubenswrapper[4632]: E0313 10:06:58.045465 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:06:58 crc kubenswrapper[4632]: E0313 10:06:58.045564 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:06:58 crc kubenswrapper[4632]: E0313 10:06:58.045520 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:06:58 crc kubenswrapper[4632]: E0313 10:06:58.176234 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:07:00 crc kubenswrapper[4632]: I0313 10:07:00.043739 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:00 crc kubenswrapper[4632]: I0313 10:07:00.043795 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:00 crc kubenswrapper[4632]: I0313 10:07:00.043754 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:00 crc kubenswrapper[4632]: E0313 10:07:00.043911 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:00 crc kubenswrapper[4632]: E0313 10:07:00.043994 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:00 crc kubenswrapper[4632]: E0313 10:07:00.044074 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:00 crc kubenswrapper[4632]: I0313 10:07:00.045017 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:00 crc kubenswrapper[4632]: E0313 10:07:00.045169 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:02 crc kubenswrapper[4632]: I0313 10:07:02.043915 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:02 crc kubenswrapper[4632]: I0313 10:07:02.044016 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:02 crc kubenswrapper[4632]: I0313 10:07:02.044054 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:02 crc kubenswrapper[4632]: I0313 10:07:02.043964 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:02 crc kubenswrapper[4632]: E0313 10:07:02.044141 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:02 crc kubenswrapper[4632]: E0313 10:07:02.044234 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:02 crc kubenswrapper[4632]: E0313 10:07:02.044296 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:02 crc kubenswrapper[4632]: E0313 10:07:02.044452 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:02 crc kubenswrapper[4632]: I0313 10:07:02.466023 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gqf22_4ec8e301-3037-4de0-94d2-32c49709660e/kube-multus/1.log" Mar 13 10:07:02 crc kubenswrapper[4632]: I0313 10:07:02.466654 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gqf22_4ec8e301-3037-4de0-94d2-32c49709660e/kube-multus/0.log" Mar 13 10:07:02 crc kubenswrapper[4632]: I0313 10:07:02.466719 4632 generic.go:334] "Generic (PLEG): container finished" podID="4ec8e301-3037-4de0-94d2-32c49709660e" containerID="e48bcc5861bda7a15e45c892fa67ba73299d99e896f36f2cb68274a659ec5d34" exitCode=1 Mar 13 10:07:02 crc kubenswrapper[4632]: I0313 10:07:02.466760 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gqf22" event={"ID":"4ec8e301-3037-4de0-94d2-32c49709660e","Type":"ContainerDied","Data":"e48bcc5861bda7a15e45c892fa67ba73299d99e896f36f2cb68274a659ec5d34"} Mar 13 10:07:02 crc kubenswrapper[4632]: I0313 10:07:02.466807 4632 scope.go:117] "RemoveContainer" containerID="0ae022ab87d0aedd5bbc6440acda6466cde0d3712108042da1225ea73ca35d6d" Mar 13 10:07:02 crc kubenswrapper[4632]: I0313 10:07:02.467217 4632 scope.go:117] "RemoveContainer" containerID="e48bcc5861bda7a15e45c892fa67ba73299d99e896f36f2cb68274a659ec5d34" Mar 13 10:07:02 crc kubenswrapper[4632]: E0313 10:07:02.467386 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-gqf22_openshift-multus(4ec8e301-3037-4de0-94d2-32c49709660e)\"" pod="openshift-multus/multus-gqf22" podUID="4ec8e301-3037-4de0-94d2-32c49709660e" Mar 13 10:07:03 crc kubenswrapper[4632]: I0313 10:07:03.045121 4632 scope.go:117] "RemoveContainer" containerID="8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b" Mar 13 10:07:03 crc kubenswrapper[4632]: E0313 10:07:03.045520 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" Mar 13 10:07:03 crc kubenswrapper[4632]: E0313 10:07:03.177669 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:07:03 crc kubenswrapper[4632]: I0313 10:07:03.471130 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gqf22_4ec8e301-3037-4de0-94d2-32c49709660e/kube-multus/1.log" Mar 13 10:07:04 crc kubenswrapper[4632]: I0313 10:07:04.043718 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:04 crc kubenswrapper[4632]: I0313 10:07:04.043802 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:04 crc kubenswrapper[4632]: E0313 10:07:04.044194 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:04 crc kubenswrapper[4632]: I0313 10:07:04.043914 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:04 crc kubenswrapper[4632]: I0313 10:07:04.043813 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:04 crc kubenswrapper[4632]: E0313 10:07:04.044361 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:04 crc kubenswrapper[4632]: E0313 10:07:04.044282 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:04 crc kubenswrapper[4632]: E0313 10:07:04.044461 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:06 crc kubenswrapper[4632]: I0313 10:07:06.044281 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:06 crc kubenswrapper[4632]: E0313 10:07:06.044455 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:06 crc kubenswrapper[4632]: I0313 10:07:06.044542 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:06 crc kubenswrapper[4632]: E0313 10:07:06.044599 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:06 crc kubenswrapper[4632]: I0313 10:07:06.044648 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:06 crc kubenswrapper[4632]: E0313 10:07:06.044701 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:06 crc kubenswrapper[4632]: I0313 10:07:06.045362 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:06 crc kubenswrapper[4632]: E0313 10:07:06.046017 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:08 crc kubenswrapper[4632]: I0313 10:07:08.044076 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:08 crc kubenswrapper[4632]: I0313 10:07:08.044090 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:08 crc kubenswrapper[4632]: I0313 10:07:08.044113 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:08 crc kubenswrapper[4632]: I0313 10:07:08.044183 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:08 crc kubenswrapper[4632]: E0313 10:07:08.045317 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:08 crc kubenswrapper[4632]: E0313 10:07:08.045420 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:08 crc kubenswrapper[4632]: E0313 10:07:08.045483 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:08 crc kubenswrapper[4632]: E0313 10:07:08.045536 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:08 crc kubenswrapper[4632]: E0313 10:07:08.179337 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:07:10 crc kubenswrapper[4632]: I0313 10:07:10.043495 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:10 crc kubenswrapper[4632]: E0313 10:07:10.043665 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:10 crc kubenswrapper[4632]: I0313 10:07:10.043913 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:10 crc kubenswrapper[4632]: E0313 10:07:10.043998 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:10 crc kubenswrapper[4632]: I0313 10:07:10.044121 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:10 crc kubenswrapper[4632]: E0313 10:07:10.044167 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:10 crc kubenswrapper[4632]: I0313 10:07:10.044275 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:10 crc kubenswrapper[4632]: E0313 10:07:10.044316 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:12 crc kubenswrapper[4632]: I0313 10:07:12.043959 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:12 crc kubenswrapper[4632]: I0313 10:07:12.044013 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:12 crc kubenswrapper[4632]: I0313 10:07:12.043971 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:12 crc kubenswrapper[4632]: I0313 10:07:12.044186 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:12 crc kubenswrapper[4632]: E0313 10:07:12.044200 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:12 crc kubenswrapper[4632]: E0313 10:07:12.044265 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:12 crc kubenswrapper[4632]: E0313 10:07:12.044305 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:12 crc kubenswrapper[4632]: E0313 10:07:12.044344 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:13 crc kubenswrapper[4632]: I0313 10:07:13.044277 4632 scope.go:117] "RemoveContainer" containerID="e48bcc5861bda7a15e45c892fa67ba73299d99e896f36f2cb68274a659ec5d34" Mar 13 10:07:13 crc kubenswrapper[4632]: E0313 10:07:13.180854 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:07:13 crc kubenswrapper[4632]: I0313 10:07:13.509277 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gqf22_4ec8e301-3037-4de0-94d2-32c49709660e/kube-multus/1.log" Mar 13 10:07:13 crc kubenswrapper[4632]: I0313 10:07:13.509334 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gqf22" event={"ID":"4ec8e301-3037-4de0-94d2-32c49709660e","Type":"ContainerStarted","Data":"5fd2699ddbdedbd54069c44af8e38bc058b347d99af772939ae6ec1d10220723"} Mar 13 10:07:14 crc kubenswrapper[4632]: I0313 10:07:14.043440 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:14 crc kubenswrapper[4632]: I0313 10:07:14.043484 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:14 crc kubenswrapper[4632]: I0313 10:07:14.043445 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:14 crc kubenswrapper[4632]: E0313 10:07:14.043715 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:14 crc kubenswrapper[4632]: I0313 10:07:14.043788 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:14 crc kubenswrapper[4632]: E0313 10:07:14.043828 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:14 crc kubenswrapper[4632]: E0313 10:07:14.043924 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:14 crc kubenswrapper[4632]: E0313 10:07:14.044021 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:14 crc kubenswrapper[4632]: I0313 10:07:14.044580 4632 scope.go:117] "RemoveContainer" containerID="8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b" Mar 13 10:07:14 crc kubenswrapper[4632]: E0313 10:07:14.044737 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" Mar 13 10:07:16 crc kubenswrapper[4632]: I0313 10:07:16.043244 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:16 crc kubenswrapper[4632]: I0313 10:07:16.043320 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:16 crc kubenswrapper[4632]: E0313 10:07:16.043395 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:16 crc kubenswrapper[4632]: I0313 10:07:16.043339 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:16 crc kubenswrapper[4632]: E0313 10:07:16.043491 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:16 crc kubenswrapper[4632]: E0313 10:07:16.043697 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:16 crc kubenswrapper[4632]: I0313 10:07:16.044124 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:16 crc kubenswrapper[4632]: E0313 10:07:16.044221 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:18 crc kubenswrapper[4632]: I0313 10:07:18.043184 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:18 crc kubenswrapper[4632]: I0313 10:07:18.043184 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:18 crc kubenswrapper[4632]: I0313 10:07:18.043224 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.044374 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:18 crc kubenswrapper[4632]: I0313 10:07:18.044393 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.044456 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.044577 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.044634 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:18 crc kubenswrapper[4632]: I0313 10:07:18.071966 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.072089 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:09:20.072062916 +0000 UTC m=+334.094593049 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:18 crc kubenswrapper[4632]: I0313 10:07:18.072123 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:18 crc kubenswrapper[4632]: I0313 10:07:18.072170 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.072250 4632 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.072312 4632 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.072336 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:09:20.072323431 +0000 UTC m=+334.094853564 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.072355 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 10:09:20.072345381 +0000 UTC m=+334.094875514 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 10:07:18 crc kubenswrapper[4632]: I0313 10:07:18.173594 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:18 crc kubenswrapper[4632]: I0313 10:07:18.173677 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:18 crc kubenswrapper[4632]: I0313 10:07:18.173710 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.173838 4632 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.173837 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.173884 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.173897 4632 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.173861 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.173899 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs podName:ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad nodeName:}" failed. No retries permitted until 2026-03-13 10:09:20.173881433 +0000 UTC m=+334.196411576 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs") pod "network-metrics-daemon-z2vlz" (UID: "ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.174020 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 10:09:20.174003045 +0000 UTC m=+334.196533178 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.174070 4632 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.174164 4632 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.174220 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 10:09:20.174208439 +0000 UTC m=+334.196738682 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 10:07:18 crc kubenswrapper[4632]: E0313 10:07:18.182376 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:07:20 crc kubenswrapper[4632]: I0313 10:07:20.043796 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:20 crc kubenswrapper[4632]: I0313 10:07:20.043861 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:20 crc kubenswrapper[4632]: I0313 10:07:20.043805 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:20 crc kubenswrapper[4632]: E0313 10:07:20.043994 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:20 crc kubenswrapper[4632]: E0313 10:07:20.044064 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:20 crc kubenswrapper[4632]: I0313 10:07:20.043820 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:20 crc kubenswrapper[4632]: E0313 10:07:20.044205 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:20 crc kubenswrapper[4632]: E0313 10:07:20.044241 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:22 crc kubenswrapper[4632]: I0313 10:07:22.043866 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:22 crc kubenswrapper[4632]: I0313 10:07:22.043934 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:22 crc kubenswrapper[4632]: I0313 10:07:22.044082 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:22 crc kubenswrapper[4632]: E0313 10:07:22.044086 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:22 crc kubenswrapper[4632]: E0313 10:07:22.044246 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:22 crc kubenswrapper[4632]: E0313 10:07:22.044358 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:22 crc kubenswrapper[4632]: I0313 10:07:22.044417 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:22 crc kubenswrapper[4632]: E0313 10:07:22.044688 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:23 crc kubenswrapper[4632]: E0313 10:07:23.184088 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:07:24 crc kubenswrapper[4632]: I0313 10:07:24.044121 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:24 crc kubenswrapper[4632]: I0313 10:07:24.044121 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:24 crc kubenswrapper[4632]: E0313 10:07:24.044317 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:24 crc kubenswrapper[4632]: E0313 10:07:24.044372 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:24 crc kubenswrapper[4632]: I0313 10:07:24.044165 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:24 crc kubenswrapper[4632]: E0313 10:07:24.044446 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:24 crc kubenswrapper[4632]: I0313 10:07:24.044703 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:24 crc kubenswrapper[4632]: E0313 10:07:24.044789 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:26 crc kubenswrapper[4632]: I0313 10:07:26.043532 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:26 crc kubenswrapper[4632]: E0313 10:07:26.043666 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:26 crc kubenswrapper[4632]: I0313 10:07:26.043669 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:26 crc kubenswrapper[4632]: I0313 10:07:26.043532 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:26 crc kubenswrapper[4632]: E0313 10:07:26.043818 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:26 crc kubenswrapper[4632]: E0313 10:07:26.043732 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:26 crc kubenswrapper[4632]: I0313 10:07:26.044507 4632 scope.go:117] "RemoveContainer" containerID="8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b" Mar 13 10:07:26 crc kubenswrapper[4632]: E0313 10:07:26.044834 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-qb725_openshift-ovn-kubernetes(3b40c6b3-0061-4224-82d5-3ccf67998722)\"" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" Mar 13 10:07:26 crc kubenswrapper[4632]: I0313 10:07:26.044932 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:26 crc kubenswrapper[4632]: E0313 10:07:26.045210 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:28 crc kubenswrapper[4632]: I0313 10:07:28.044356 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:28 crc kubenswrapper[4632]: I0313 10:07:28.044415 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:28 crc kubenswrapper[4632]: I0313 10:07:28.044460 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:28 crc kubenswrapper[4632]: E0313 10:07:28.045655 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:28 crc kubenswrapper[4632]: E0313 10:07:28.045870 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:28 crc kubenswrapper[4632]: E0313 10:07:28.046017 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:28 crc kubenswrapper[4632]: I0313 10:07:28.046159 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:28 crc kubenswrapper[4632]: E0313 10:07:28.046267 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:28 crc kubenswrapper[4632]: E0313 10:07:28.185867 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:07:30 crc kubenswrapper[4632]: I0313 10:07:30.044000 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:30 crc kubenswrapper[4632]: I0313 10:07:30.044064 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:30 crc kubenswrapper[4632]: I0313 10:07:30.044075 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:30 crc kubenswrapper[4632]: E0313 10:07:30.044827 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:30 crc kubenswrapper[4632]: E0313 10:07:30.044512 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:30 crc kubenswrapper[4632]: E0313 10:07:30.044728 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:30 crc kubenswrapper[4632]: I0313 10:07:30.044101 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:30 crc kubenswrapper[4632]: E0313 10:07:30.044930 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:32 crc kubenswrapper[4632]: I0313 10:07:32.043981 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:32 crc kubenswrapper[4632]: I0313 10:07:32.044036 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:32 crc kubenswrapper[4632]: I0313 10:07:32.044049 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:32 crc kubenswrapper[4632]: I0313 10:07:32.044004 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:32 crc kubenswrapper[4632]: E0313 10:07:32.044152 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:32 crc kubenswrapper[4632]: E0313 10:07:32.044690 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:32 crc kubenswrapper[4632]: E0313 10:07:32.051518 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:32 crc kubenswrapper[4632]: E0313 10:07:32.052160 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:33 crc kubenswrapper[4632]: E0313 10:07:33.187094 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:07:34 crc kubenswrapper[4632]: I0313 10:07:34.044432 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:34 crc kubenswrapper[4632]: I0313 10:07:34.044494 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:34 crc kubenswrapper[4632]: I0313 10:07:34.044476 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:34 crc kubenswrapper[4632]: I0313 10:07:34.044447 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:34 crc kubenswrapper[4632]: E0313 10:07:34.044603 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:34 crc kubenswrapper[4632]: E0313 10:07:34.044673 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:34 crc kubenswrapper[4632]: E0313 10:07:34.044732 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:34 crc kubenswrapper[4632]: E0313 10:07:34.044786 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:36 crc kubenswrapper[4632]: I0313 10:07:36.044133 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:36 crc kubenswrapper[4632]: E0313 10:07:36.044275 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:36 crc kubenswrapper[4632]: I0313 10:07:36.044449 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:36 crc kubenswrapper[4632]: E0313 10:07:36.044518 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:36 crc kubenswrapper[4632]: I0313 10:07:36.044631 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:36 crc kubenswrapper[4632]: E0313 10:07:36.044675 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:36 crc kubenswrapper[4632]: I0313 10:07:36.044858 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:36 crc kubenswrapper[4632]: E0313 10:07:36.044963 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:38 crc kubenswrapper[4632]: I0313 10:07:38.043385 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:38 crc kubenswrapper[4632]: I0313 10:07:38.043396 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:38 crc kubenswrapper[4632]: I0313 10:07:38.043504 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:38 crc kubenswrapper[4632]: I0313 10:07:38.043509 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:38 crc kubenswrapper[4632]: E0313 10:07:38.044682 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:38 crc kubenswrapper[4632]: E0313 10:07:38.044568 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:38 crc kubenswrapper[4632]: E0313 10:07:38.045552 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:38 crc kubenswrapper[4632]: E0313 10:07:38.045667 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:38 crc kubenswrapper[4632]: I0313 10:07:38.046133 4632 scope.go:117] "RemoveContainer" containerID="8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b" Mar 13 10:07:38 crc kubenswrapper[4632]: E0313 10:07:38.188500 4632 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 10:07:38 crc kubenswrapper[4632]: I0313 10:07:38.588772 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/3.log" Mar 13 10:07:38 crc kubenswrapper[4632]: I0313 10:07:38.590929 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerStarted","Data":"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776"} Mar 13 10:07:38 crc kubenswrapper[4632]: I0313 10:07:38.591426 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:07:38 crc kubenswrapper[4632]: I0313 10:07:38.925693 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-z2vlz"] Mar 13 10:07:38 crc kubenswrapper[4632]: I0313 10:07:38.925822 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:38 crc kubenswrapper[4632]: E0313 10:07:38.925919 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:40 crc kubenswrapper[4632]: I0313 10:07:40.044207 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:40 crc kubenswrapper[4632]: I0313 10:07:40.044200 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:40 crc kubenswrapper[4632]: I0313 10:07:40.044219 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:40 crc kubenswrapper[4632]: E0313 10:07:40.044652 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:40 crc kubenswrapper[4632]: E0313 10:07:40.044932 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:40 crc kubenswrapper[4632]: E0313 10:07:40.045072 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:40 crc kubenswrapper[4632]: I0313 10:07:40.461442 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:07:40 crc kubenswrapper[4632]: I0313 10:07:40.461515 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:07:41 crc kubenswrapper[4632]: I0313 10:07:41.043920 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:41 crc kubenswrapper[4632]: E0313 10:07:41.044151 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:42 crc kubenswrapper[4632]: I0313 10:07:42.043176 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:42 crc kubenswrapper[4632]: I0313 10:07:42.043374 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:42 crc kubenswrapper[4632]: E0313 10:07:42.043560 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 10:07:42 crc kubenswrapper[4632]: I0313 10:07:42.043815 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:42 crc kubenswrapper[4632]: E0313 10:07:42.043890 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 10:07:42 crc kubenswrapper[4632]: E0313 10:07:42.044121 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.043762 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:43 crc kubenswrapper[4632]: E0313 10:07:43.043939 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-z2vlz" podUID="ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.283152 4632 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.321046 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c6jnc"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.321562 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.321927 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9v5nn"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.322313 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.322616 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.323224 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.323547 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-p9gp2"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.324091 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.324182 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.324738 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.328836 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.329255 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.329691 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.329937 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.330127 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.330626 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.340686 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.340766 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.340686 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.340697 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.340937 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.341139 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.341533 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.341693 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.345043 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.345537 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.345915 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.346526 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.346548 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.346589 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.346606 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.346694 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.346700 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.346859 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.347008 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.347227 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.352213 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.353829 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-w2hhj"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.354381 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-w2hhj" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.357901 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.358242 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.358460 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.358802 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.358944 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.359336 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.359450 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.359554 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.359654 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.359757 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.360053 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.360171 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.377032 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.377477 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.377514 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.377980 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.378101 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.378153 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.377984 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.378374 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.380049 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.380180 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.380309 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.380383 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.380415 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.380508 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.380784 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.381024 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.381467 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.381887 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8sl88"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.383566 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.384181 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.385306 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.390475 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.409523 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.409757 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.412549 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.412562 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.412693 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.412723 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.414704 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-svhr5"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.415441 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.417268 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sbtn5"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.417676 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-zn7mn"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.417788 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.412983 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.413164 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.415615 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.418271 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.418979 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.413205 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.421111 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.413277 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.415677 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.415757 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.415827 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.416043 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.416184 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.422182 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.423108 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.424016 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.424388 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6fqf5"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.424710 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.424761 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-6fqf5" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.425213 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.427286 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxs5z"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.427815 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.428012 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-k955n"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.428648 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.429596 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.430092 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.430182 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.430273 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.433350 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-t9vht"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.434142 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.435132 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.435756 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c6jnc"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.435925 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.436342 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-99hff"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.436509 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.437027 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.437781 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.438364 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9wxcs"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.438651 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.438895 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9wxcs" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.439707 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.439922 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.439936 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.440096 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.440238 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.440397 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.441284 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.442102 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.442739 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.449208 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.449466 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.449553 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.449723 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.449828 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.449944 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.450063 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.450229 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.450323 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.450411 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.450512 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.454608 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.454723 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.454816 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.454883 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.454970 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.455236 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.474647 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.476547 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.477330 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.478735 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.482466 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d19fca6e-5095-42b6-8590-32c5b2c73308-serving-cert\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.482573 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/37df1143-69fc-4d13-a5d3-790a9d14814a-etcd-client\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.482637 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/560e6c43-4285-4ca8-98b9-874e9dcb5810-config\") pod \"route-controller-manager-6576b87f9c-xthqz\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.482718 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/37df1143-69fc-4d13-a5d3-790a9d14814a-node-pullsecrets\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.482754 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-audit-policies\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.482845 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54b67d35-da46-4e38-9b9a-e91855d6d88d-auth-proxy-config\") pod \"machine-approver-56656f9798-hmljp\" (UID: \"54b67d35-da46-4e38-9b9a-e91855d6d88d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.482916 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/275c3112-6912-49f8-9d3f-8147662fb99f-config\") pod \"machine-api-operator-5694c8668f-c6jnc\" (UID: \"275c3112-6912-49f8-9d3f-8147662fb99f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.482999 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8966c5f5-d0a8-4533-842c-0930c1a97bd7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qtrc2\" (UID: \"8966c5f5-d0a8-4533-842c-0930c1a97bd7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.483074 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/37df1143-69fc-4d13-a5d3-790a9d14814a-audit-dir\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.483099 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.483386 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vvbt\" (UniqueName: \"kubernetes.io/projected/70f440bb-5dd8-4863-9749-bc5f7c547750-kube-api-access-6vvbt\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.483462 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.483493 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/54b67d35-da46-4e38-9b9a-e91855d6d88d-machine-approver-tls\") pod \"machine-approver-56656f9798-hmljp\" (UID: \"54b67d35-da46-4e38-9b9a-e91855d6d88d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.483552 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d19fca6e-5095-42b6-8590-32c5b2c73308-encryption-config\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.483643 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc56m\" (UniqueName: \"kubernetes.io/projected/d19fca6e-5095-42b6-8590-32c5b2c73308-kube-api-access-tc56m\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.483794 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-audit-dir\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.483866 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.484273 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/560e6c43-4285-4ca8-98b9-874e9dcb5810-serving-cert\") pod \"route-controller-manager-6576b87f9c-xthqz\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.484316 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37df1143-69fc-4d13-a5d3-790a9d14814a-serving-cert\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.484350 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-image-import-ca\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.484377 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c90710f-4595-425c-8be1-1436f43b5069-config\") pod \"openshift-apiserver-operator-796bbdcf4f-9sqbn\" (UID: \"2c90710f-4595-425c-8be1-1436f43b5069\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.484483 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r8xf\" (UniqueName: \"kubernetes.io/projected/54b67d35-da46-4e38-9b9a-e91855d6d88d-kube-api-access-4r8xf\") pod \"machine-approver-56656f9798-hmljp\" (UID: \"54b67d35-da46-4e38-9b9a-e91855d6d88d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.484526 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d19fca6e-5095-42b6-8590-32c5b2c73308-audit-policies\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.484597 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54b67d35-da46-4e38-9b9a-e91855d6d88d-config\") pod \"machine-approver-56656f9798-hmljp\" (UID: \"54b67d35-da46-4e38-9b9a-e91855d6d88d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.484696 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.484857 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/560e6c43-4285-4ca8-98b9-874e9dcb5810-client-ca\") pod \"route-controller-manager-6576b87f9c-xthqz\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.484968 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nsdx\" (UniqueName: \"kubernetes.io/projected/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-kube-api-access-9nsdx\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.485019 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d19fca6e-5095-42b6-8590-32c5b2c73308-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.485064 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d19fca6e-5095-42b6-8590-32c5b2c73308-audit-dir\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.485124 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx8gt\" (UniqueName: \"kubernetes.io/projected/8966c5f5-d0a8-4533-842c-0930c1a97bd7-kube-api-access-hx8gt\") pod \"cluster-samples-operator-665b6dd947-qtrc2\" (UID: \"8966c5f5-d0a8-4533-842c-0930c1a97bd7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.485336 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt2mw\" (UniqueName: \"kubernetes.io/projected/2c90710f-4595-425c-8be1-1436f43b5069-kube-api-access-rt2mw\") pod \"openshift-apiserver-operator-796bbdcf4f-9sqbn\" (UID: \"2c90710f-4595-425c-8be1-1436f43b5069\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.485412 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70f440bb-5dd8-4863-9749-bc5f7c547750-serving-cert\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.485582 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.485648 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.485741 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.485845 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-config\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.485875 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d19fca6e-5095-42b6-8590-32c5b2c73308-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.485951 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-etcd-serving-ca\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.485980 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/275c3112-6912-49f8-9d3f-8147662fb99f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c6jnc\" (UID: \"275c3112-6912-49f8-9d3f-8147662fb99f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486007 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdwcs\" (UniqueName: \"kubernetes.io/projected/560e6c43-4285-4ca8-98b9-874e9dcb5810-kube-api-access-sdwcs\") pod \"route-controller-manager-6576b87f9c-xthqz\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486121 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-config\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486159 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-client-ca\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486180 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d19fca6e-5095-42b6-8590-32c5b2c73308-etcd-client\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486218 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/37df1143-69fc-4d13-a5d3-790a9d14814a-encryption-config\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486280 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486318 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486360 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-audit\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486383 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486411 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c90710f-4595-425c-8be1-1436f43b5069-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-9sqbn\" (UID: \"2c90710f-4595-425c-8be1-1436f43b5069\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486509 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486567 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdwt2\" (UniqueName: \"kubernetes.io/projected/37df1143-69fc-4d13-a5d3-790a9d14814a-kube-api-access-hdwt2\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486681 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzjdx\" (UniqueName: \"kubernetes.io/projected/275c3112-6912-49f8-9d3f-8147662fb99f-kube-api-access-zzjdx\") pod \"machine-api-operator-5694c8668f-c6jnc\" (UID: \"275c3112-6912-49f8-9d3f-8147662fb99f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486814 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486860 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f869\" (UniqueName: \"kubernetes.io/projected/7d155f24-9bfc-4039-9981-10e7f724fa51-kube-api-access-8f869\") pod \"downloads-7954f5f757-w2hhj\" (UID: \"7d155f24-9bfc-4039-9981-10e7f724fa51\") " pod="openshift-console/downloads-7954f5f757-w2hhj" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.489058 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.486930 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/275c3112-6912-49f8-9d3f-8147662fb99f-images\") pod \"machine-api-operator-5694c8668f-c6jnc\" (UID: \"275c3112-6912-49f8-9d3f-8147662fb99f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.494028 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.496622 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.497870 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.502103 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.503155 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.523645 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.523724 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pvwll"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.524273 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.524304 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pvwll" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.526030 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.526285 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.529840 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2n99d"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.530431 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.530467 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.530820 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.531034 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.531220 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.531306 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.532074 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.532923 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.533484 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4hmjh"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.533882 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.534548 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.535135 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.535351 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zh465"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.535553 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.535860 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zh465" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.539439 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-68mjx"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.540273 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556606-mkrp2"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.540927 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.541495 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.540464 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-4hmjh" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.541081 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556606-mkrp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.542012 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.540510 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-68mjx" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.544747 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.548250 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-w2hhj"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.548290 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-p9gp2"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.548306 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.548578 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.552327 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hvrrc"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.554161 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9v5nn"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.554259 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.554726 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.556227 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6fqf5"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.557013 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.557932 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.559346 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-qcb4l"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.560101 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qcb4l" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.560525 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-svhr5"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.566382 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.571560 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8sl88"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.573412 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxs5z"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.575149 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.575929 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.591219 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.596439 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/275c3112-6912-49f8-9d3f-8147662fb99f-config\") pod \"machine-api-operator-5694c8668f-c6jnc\" (UID: \"275c3112-6912-49f8-9d3f-8147662fb99f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.596688 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8966c5f5-d0a8-4533-842c-0930c1a97bd7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qtrc2\" (UID: \"8966c5f5-d0a8-4533-842c-0930c1a97bd7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.596803 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/37df1143-69fc-4d13-a5d3-790a9d14814a-audit-dir\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.596888 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.596991 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vvbt\" (UniqueName: \"kubernetes.io/projected/70f440bb-5dd8-4863-9749-bc5f7c547750-kube-api-access-6vvbt\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.597071 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.597208 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/37df1143-69fc-4d13-a5d3-790a9d14814a-audit-dir\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.591552 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4hmjh"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.597643 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.597735 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.597799 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9wxcs"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.597871 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pvwll"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.597973 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-68mjx"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.598499 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/275c3112-6912-49f8-9d3f-8147662fb99f-config\") pod \"machine-api-operator-5694c8668f-c6jnc\" (UID: \"275c3112-6912-49f8-9d3f-8147662fb99f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.603216 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.603514 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.607853 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.611993 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/8966c5f5-d0a8-4533-842c-0930c1a97bd7-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-qtrc2\" (UID: \"8966c5f5-d0a8-4533-842c-0930c1a97bd7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.613391 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zn7mn"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.617895 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2n99d"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618088 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/54b67d35-da46-4e38-9b9a-e91855d6d88d-machine-approver-tls\") pod \"machine-approver-56656f9798-hmljp\" (UID: \"54b67d35-da46-4e38-9b9a-e91855d6d88d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618310 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d19fca6e-5095-42b6-8590-32c5b2c73308-encryption-config\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618355 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc56m\" (UniqueName: \"kubernetes.io/projected/d19fca6e-5095-42b6-8590-32c5b2c73308-kube-api-access-tc56m\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618389 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-audit-dir\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618425 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618465 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f5a50074-5531-442f-a0e9-0578f15634c1-console-serving-cert\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618511 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/560e6c43-4285-4ca8-98b9-874e9dcb5810-serving-cert\") pod \"route-controller-manager-6576b87f9c-xthqz\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618548 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37df1143-69fc-4d13-a5d3-790a9d14814a-serving-cert\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618585 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-image-import-ca\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618648 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c90710f-4595-425c-8be1-1436f43b5069-config\") pod \"openshift-apiserver-operator-796bbdcf4f-9sqbn\" (UID: \"2c90710f-4595-425c-8be1-1436f43b5069\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618700 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r8xf\" (UniqueName: \"kubernetes.io/projected/54b67d35-da46-4e38-9b9a-e91855d6d88d-kube-api-access-4r8xf\") pod \"machine-approver-56656f9798-hmljp\" (UID: \"54b67d35-da46-4e38-9b9a-e91855d6d88d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618727 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d19fca6e-5095-42b6-8590-32c5b2c73308-audit-policies\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618758 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f5a50074-5531-442f-a0e9-0578f15634c1-console-oauth-config\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618823 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54b67d35-da46-4e38-9b9a-e91855d6d88d-config\") pod \"machine-approver-56656f9798-hmljp\" (UID: \"54b67d35-da46-4e38-9b9a-e91855d6d88d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618877 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.618967 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/560e6c43-4285-4ca8-98b9-874e9dcb5810-client-ca\") pod \"route-controller-manager-6576b87f9c-xthqz\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619006 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nsdx\" (UniqueName: \"kubernetes.io/projected/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-kube-api-access-9nsdx\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619034 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d19fca6e-5095-42b6-8590-32c5b2c73308-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619058 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d19fca6e-5095-42b6-8590-32c5b2c73308-audit-dir\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619089 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-trusted-ca-bundle\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619121 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx8gt\" (UniqueName: \"kubernetes.io/projected/8966c5f5-d0a8-4533-842c-0930c1a97bd7-kube-api-access-hx8gt\") pod \"cluster-samples-operator-665b6dd947-qtrc2\" (UID: \"8966c5f5-d0a8-4533-842c-0930c1a97bd7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619166 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt2mw\" (UniqueName: \"kubernetes.io/projected/2c90710f-4595-425c-8be1-1436f43b5069-kube-api-access-rt2mw\") pod \"openshift-apiserver-operator-796bbdcf4f-9sqbn\" (UID: \"2c90710f-4595-425c-8be1-1436f43b5069\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619204 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70f440bb-5dd8-4863-9749-bc5f7c547750-serving-cert\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619232 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619291 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619333 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619364 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-etcd-serving-ca\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619393 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-config\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619420 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d19fca6e-5095-42b6-8590-32c5b2c73308-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619445 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/275c3112-6912-49f8-9d3f-8147662fb99f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c6jnc\" (UID: \"275c3112-6912-49f8-9d3f-8147662fb99f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619478 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdwcs\" (UniqueName: \"kubernetes.io/projected/560e6c43-4285-4ca8-98b9-874e9dcb5810-kube-api-access-sdwcs\") pod \"route-controller-manager-6576b87f9c-xthqz\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619510 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-config\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619538 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-client-ca\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619561 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d19fca6e-5095-42b6-8590-32c5b2c73308-etcd-client\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619591 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-console-config\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619622 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-audit\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619652 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/37df1143-69fc-4d13-a5d3-790a9d14814a-encryption-config\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619679 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619704 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619732 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-oauth-serving-cert\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619760 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpqvj\" (UniqueName: \"kubernetes.io/projected/f5a50074-5531-442f-a0e9-0578f15634c1-kube-api-access-gpqvj\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619803 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619830 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c90710f-4595-425c-8be1-1436f43b5069-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-9sqbn\" (UID: \"2c90710f-4595-425c-8be1-1436f43b5069\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619859 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619896 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdwt2\" (UniqueName: \"kubernetes.io/projected/37df1143-69fc-4d13-a5d3-790a9d14814a-kube-api-access-hdwt2\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619928 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzjdx\" (UniqueName: \"kubernetes.io/projected/275c3112-6912-49f8-9d3f-8147662fb99f-kube-api-access-zzjdx\") pod \"machine-api-operator-5694c8668f-c6jnc\" (UID: \"275c3112-6912-49f8-9d3f-8147662fb99f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.620034 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.620064 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f869\" (UniqueName: \"kubernetes.io/projected/7d155f24-9bfc-4039-9981-10e7f724fa51-kube-api-access-8f869\") pod \"downloads-7954f5f757-w2hhj\" (UID: \"7d155f24-9bfc-4039-9981-10e7f724fa51\") " pod="openshift-console/downloads-7954f5f757-w2hhj" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.620095 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/275c3112-6912-49f8-9d3f-8147662fb99f-images\") pod \"machine-api-operator-5694c8668f-c6jnc\" (UID: \"275c3112-6912-49f8-9d3f-8147662fb99f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.620123 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.620174 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-service-ca\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.620202 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/560e6c43-4285-4ca8-98b9-874e9dcb5810-config\") pod \"route-controller-manager-6576b87f9c-xthqz\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.620231 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d19fca6e-5095-42b6-8590-32c5b2c73308-serving-cert\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.620254 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-audit-dir\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.620259 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/37df1143-69fc-4d13-a5d3-790a9d14814a-etcd-client\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.620362 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/37df1143-69fc-4d13-a5d3-790a9d14814a-node-pullsecrets\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.620399 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-audit-policies\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.620443 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54b67d35-da46-4e38-9b9a-e91855d6d88d-auth-proxy-config\") pod \"machine-approver-56656f9798-hmljp\" (UID: \"54b67d35-da46-4e38-9b9a-e91855d6d88d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.621602 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/54b67d35-da46-4e38-9b9a-e91855d6d88d-auth-proxy-config\") pod \"machine-approver-56656f9798-hmljp\" (UID: \"54b67d35-da46-4e38-9b9a-e91855d6d88d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.622310 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.622316 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/d19fca6e-5095-42b6-8590-32c5b2c73308-encryption-config\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.623432 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-config\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.624383 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/d19fca6e-5095-42b6-8590-32c5b2c73308-audit-policies\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.625054 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/54b67d35-da46-4e38-9b9a-e91855d6d88d-config\") pod \"machine-approver-56656f9798-hmljp\" (UID: \"54b67d35-da46-4e38-9b9a-e91855d6d88d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.625873 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/37df1143-69fc-4d13-a5d3-790a9d14814a-etcd-client\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.627223 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-client-ca\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.628181 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/560e6c43-4285-4ca8-98b9-874e9dcb5810-client-ca\") pod \"route-controller-manager-6576b87f9c-xthqz\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.628534 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.629041 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d19fca6e-5095-42b6-8590-32c5b2c73308-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.629094 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/d19fca6e-5095-42b6-8590-32c5b2c73308-audit-dir\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.632233 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-image-import-ca\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.632279 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/d19fca6e-5095-42b6-8590-32c5b2c73308-etcd-client\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.632345 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.632583 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/37df1143-69fc-4d13-a5d3-790a9d14814a-node-pullsecrets\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.632625 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/560e6c43-4285-4ca8-98b9-874e9dcb5810-serving-cert\") pod \"route-controller-manager-6576b87f9c-xthqz\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.633060 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-audit\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.633165 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-audit-policies\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.633822 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/275c3112-6912-49f8-9d3f-8147662fb99f-images\") pod \"machine-api-operator-5694c8668f-c6jnc\" (UID: \"275c3112-6912-49f8-9d3f-8147662fb99f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.619975 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c90710f-4595-425c-8be1-1436f43b5069-config\") pod \"openshift-apiserver-operator-796bbdcf4f-9sqbn\" (UID: \"2c90710f-4595-425c-8be1-1436f43b5069\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.634726 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/37df1143-69fc-4d13-a5d3-790a9d14814a-serving-cert\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.635034 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.637749 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sbtn5"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.638277 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/560e6c43-4285-4ca8-98b9-874e9dcb5810-config\") pod \"route-controller-manager-6576b87f9c-xthqz\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.646670 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.647139 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70f440bb-5dd8-4863-9749-bc5f7c547750-serving-cert\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.647168 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/54b67d35-da46-4e38-9b9a-e91855d6d88d-machine-approver-tls\") pod \"machine-approver-56656f9798-hmljp\" (UID: \"54b67d35-da46-4e38-9b9a-e91855d6d88d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.653042 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.654679 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-etcd-serving-ca\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.654784 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.656189 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.656307 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.656421 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/37df1143-69fc-4d13-a5d3-790a9d14814a-encryption-config\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.656575 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37df1143-69fc-4d13-a5d3-790a9d14814a-trusted-ca-bundle\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.656878 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.658839 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c90710f-4595-425c-8be1-1436f43b5069-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-9sqbn\" (UID: \"2c90710f-4595-425c-8be1-1436f43b5069\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.660154 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.660757 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.663394 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-config\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.663476 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556606-mkrp2"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.663840 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.664935 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.665159 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.665256 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/d19fca6e-5095-42b6-8590-32c5b2c73308-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.666237 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.667148 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.668305 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.669334 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/275c3112-6912-49f8-9d3f-8147662fb99f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-c6jnc\" (UID: \"275c3112-6912-49f8-9d3f-8147662fb99f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.669466 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-99hff"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.669756 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.672155 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.675078 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zh465"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.676590 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-hlf9t"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.678050 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-k955n"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.678200 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hlf9t" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.678627 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qcb4l"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.679579 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d19fca6e-5095-42b6-8590-32c5b2c73308-serving-cert\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.680886 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hvrrc"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.683480 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.686596 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-g2wxc"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.687486 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-g2wxc" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.689968 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-g2wxc"] Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.703270 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.720979 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f5a50074-5531-442f-a0e9-0578f15634c1-console-serving-cert\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.721033 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f5a50074-5531-442f-a0e9-0578f15634c1-console-oauth-config\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.721071 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-trusted-ca-bundle\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.721116 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-console-config\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.721134 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-oauth-serving-cert\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.721148 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpqvj\" (UniqueName: \"kubernetes.io/projected/f5a50074-5531-442f-a0e9-0578f15634c1-kube-api-access-gpqvj\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.721184 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-service-ca\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.722250 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-console-config\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.722301 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-oauth-serving-cert\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.723472 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-trusted-ca-bundle\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.724234 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.725195 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f5a50074-5531-442f-a0e9-0578f15634c1-console-serving-cert\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.725355 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f5a50074-5531-442f-a0e9-0578f15634c1-console-oauth-config\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.725519 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-service-ca\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.743579 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.763582 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.783538 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.803201 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.823589 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.843883 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.864071 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.883734 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.904298 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.923768 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.944414 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.963400 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 13 10:07:43 crc kubenswrapper[4632]: I0313 10:07:43.983706 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.004376 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.023603 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.043176 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.043193 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.043201 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.043911 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.063509 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.083773 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.103898 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.124873 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.144065 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.163795 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.184606 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.204349 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.223722 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.243917 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.264335 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.282863 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.324137 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.344230 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.364454 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.384463 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.404375 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.424155 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.444505 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.463416 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.483481 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.502014 4632 request.go:700] Waited for 1.003734169s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-operator/configmaps?fieldSelector=metadata.name%3Dtrusted-ca&limit=500&resourceVersion=0 Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.517662 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.524204 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.545309 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.564912 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.584704 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.604105 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.644528 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.664298 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.684687 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.711412 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.723569 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.744697 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.763777 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.784353 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.803651 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.823129 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.844877 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.863923 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.884445 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.904350 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.923901 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.944273 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.964087 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 13 10:07:44 crc kubenswrapper[4632]: I0313 10:07:44.984127 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.005395 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.023719 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.044200 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.044807 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.063837 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.084657 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.105262 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.124797 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.145179 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.164799 4632 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.183937 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.203671 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.224265 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.245197 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.264313 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.284852 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.320695 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vvbt\" (UniqueName: \"kubernetes.io/projected/70f440bb-5dd8-4863-9749-bc5f7c547750-kube-api-access-6vvbt\") pod \"controller-manager-879f6c89f-9v5nn\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.339367 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc56m\" (UniqueName: \"kubernetes.io/projected/d19fca6e-5095-42b6-8590-32c5b2c73308-kube-api-access-tc56m\") pod \"apiserver-7bbb656c7d-7vrbc\" (UID: \"d19fca6e-5095-42b6-8590-32c5b2c73308\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.372785 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r8xf\" (UniqueName: \"kubernetes.io/projected/54b67d35-da46-4e38-9b9a-e91855d6d88d-kube-api-access-4r8xf\") pod \"machine-approver-56656f9798-hmljp\" (UID: \"54b67d35-da46-4e38-9b9a-e91855d6d88d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.379524 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt2mw\" (UniqueName: \"kubernetes.io/projected/2c90710f-4595-425c-8be1-1436f43b5069-kube-api-access-rt2mw\") pod \"openshift-apiserver-operator-796bbdcf4f-9sqbn\" (UID: \"2c90710f-4595-425c-8be1-1436f43b5069\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.398085 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nsdx\" (UniqueName: \"kubernetes.io/projected/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-kube-api-access-9nsdx\") pod \"oauth-openshift-558db77b4-8sl88\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.417904 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx8gt\" (UniqueName: \"kubernetes.io/projected/8966c5f5-d0a8-4533-842c-0930c1a97bd7-kube-api-access-hx8gt\") pod \"cluster-samples-operator-665b6dd947-qtrc2\" (UID: \"8966c5f5-d0a8-4533-842c-0930c1a97bd7\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.438108 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdwt2\" (UniqueName: \"kubernetes.io/projected/37df1143-69fc-4d13-a5d3-790a9d14814a-kube-api-access-hdwt2\") pod \"apiserver-76f77b778f-p9gp2\" (UID: \"37df1143-69fc-4d13-a5d3-790a9d14814a\") " pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.456370 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.458658 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzjdx\" (UniqueName: \"kubernetes.io/projected/275c3112-6912-49f8-9d3f-8147662fb99f-kube-api-access-zzjdx\") pod \"machine-api-operator-5694c8668f-c6jnc\" (UID: \"275c3112-6912-49f8-9d3f-8147662fb99f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.478233 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f869\" (UniqueName: \"kubernetes.io/projected/7d155f24-9bfc-4039-9981-10e7f724fa51-kube-api-access-8f869\") pod \"downloads-7954f5f757-w2hhj\" (UID: \"7d155f24-9bfc-4039-9981-10e7f724fa51\") " pod="openshift-console/downloads-7954f5f757-w2hhj" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.480882 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.501745 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdwcs\" (UniqueName: \"kubernetes.io/projected/560e6c43-4285-4ca8-98b9-874e9dcb5810-kube-api-access-sdwcs\") pod \"route-controller-manager-6576b87f9c-xthqz\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.503812 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.515419 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.521635 4632 request.go:700] Waited for 1.843009811s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-config-operator/secrets?fieldSelector=metadata.name%3Dnode-bootstrapper-token&limit=500&resourceVersion=0 Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.523850 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.525972 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.543834 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.547298 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.563652 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.573105 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.584294 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.597835 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.603991 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.631667 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-w2hhj" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.654915 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.655929 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.676528 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.676751 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpqvj\" (UniqueName: \"kubernetes.io/projected/f5a50074-5531-442f-a0e9-0578f15634c1-kube-api-access-gpqvj\") pod \"console-f9d7485db-zn7mn\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.684494 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.687579 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" event={"ID":"54b67d35-da46-4e38-9b9a-e91855d6d88d","Type":"ContainerStarted","Data":"80698629bef138bcda96418314faf78a7b20ad4ae2335c25b157e45f3f92fa55"} Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.708818 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.747253 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.756906 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22993daf-2b32-4be5-8eb7-f9194e903d62-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-946gp\" (UID: \"22993daf-2b32-4be5-8eb7-f9194e903d62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757013 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvs6x\" (UniqueName: \"kubernetes.io/projected/bbb27a61-7407-4cd7-84df-4b66fbdcf82d-kube-api-access-pvs6x\") pod \"machine-config-operator-74547568cd-99hff\" (UID: \"bbb27a61-7407-4cd7-84df-4b66fbdcf82d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757042 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/32f62e32-732b-4646-85f0-45b8ea6544a6-profile-collector-cert\") pod \"catalog-operator-68c6474976-r5v5p\" (UID: \"32f62e32-732b-4646-85f0-45b8ea6544a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757063 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7b8ca1c-c3de-4829-ab9f-860f76033c63-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hd8rx\" (UID: \"b7b8ca1c-c3de-4829-ab9f-860f76033c63\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757099 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cclg\" (UniqueName: \"kubernetes.io/projected/ef269b18-ea84-43c2-971c-e772149acbf6-kube-api-access-2cclg\") pod \"console-operator-58897d9998-sbtn5\" (UID: \"ef269b18-ea84-43c2-971c-e772149acbf6\") " pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757120 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb4c6\" (UID: \"e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757142 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f781cb50-1e1b-4586-ba59-b204b1a6beec-config\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757162 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f781cb50-1e1b-4586-ba59-b204b1a6beec-etcd-service-ca\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757183 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f781cb50-1e1b-4586-ba59-b204b1a6beec-etcd-ca\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757224 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bbb27a61-7407-4cd7-84df-4b66fbdcf82d-proxy-tls\") pod \"machine-config-operator-74547568cd-99hff\" (UID: \"bbb27a61-7407-4cd7-84df-4b66fbdcf82d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757244 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7mb4q\" (UniqueName: \"kubernetes.io/projected/c94773d8-a922-4778-b2ba-8937e9d6c19b-kube-api-access-7mb4q\") pod \"dns-operator-744455d44c-6fqf5\" (UID: \"c94773d8-a922-4778-b2ba-8937e9d6c19b\") " pod="openshift-dns-operator/dns-operator-744455d44c-6fqf5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757268 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9rrcn\" (UID: \"a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757291 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb4c6\" (UID: \"e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757314 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7b959a85-56a5-4296-9cf3-87741e1f9c39-stats-auth\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757335 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73-trusted-ca\") pod \"ingress-operator-5b745b69d9-jrkwc\" (UID: \"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757376 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7b959a85-56a5-4296-9cf3-87741e1f9c39-metrics-certs\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757396 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62kgq\" (UniqueName: \"kubernetes.io/projected/7b959a85-56a5-4296-9cf3-87741e1f9c39-kube-api-access-62kgq\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757429 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef269b18-ea84-43c2-971c-e772149acbf6-trusted-ca\") pod \"console-operator-58897d9998-sbtn5\" (UID: \"ef269b18-ea84-43c2-971c-e772149acbf6\") " pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757460 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f56fc09a-e2b7-46db-b938-f276df3f033e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757635 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22993daf-2b32-4be5-8eb7-f9194e903d62-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-946gp\" (UID: \"22993daf-2b32-4be5-8eb7-f9194e903d62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757673 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f781cb50-1e1b-4586-ba59-b204b1a6beec-serving-cert\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757700 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c94773d8-a922-4778-b2ba-8937e9d6c19b-metrics-tls\") pod \"dns-operator-744455d44c-6fqf5\" (UID: \"c94773d8-a922-4778-b2ba-8937e9d6c19b\") " pod="openshift-dns-operator/dns-operator-744455d44c-6fqf5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757727 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52zmx\" (UniqueName: \"kubernetes.io/projected/fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73-kube-api-access-52zmx\") pod \"ingress-operator-5b745b69d9-jrkwc\" (UID: \"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757880 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bbb27a61-7407-4cd7-84df-4b66fbdcf82d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-99hff\" (UID: \"bbb27a61-7407-4cd7-84df-4b66fbdcf82d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.757926 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73-bound-sa-token\") pod \"ingress-operator-5b745b69d9-jrkwc\" (UID: \"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.759510 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f781cb50-1e1b-4586-ba59-b204b1a6beec-etcd-client\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.759544 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cd4c3b3-6825-4bd2-97a5-330f91782d4b-config\") pod \"kube-controller-manager-operator-78b949d7b-rvkzz\" (UID: \"9cd4c3b3-6825-4bd2-97a5-330f91782d4b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.759580 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bbb27a61-7407-4cd7-84df-4b66fbdcf82d-images\") pod \"machine-config-operator-74547568cd-99hff\" (UID: \"bbb27a61-7407-4cd7-84df-4b66fbdcf82d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.759606 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7x7q\" (UniqueName: \"kubernetes.io/projected/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-kube-api-access-w7x7q\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.759627 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zq92\" (UniqueName: \"kubernetes.io/projected/a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59-kube-api-access-8zq92\") pod \"cluster-image-registry-operator-dc59b4c8b-9rrcn\" (UID: \"a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.759652 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f56fc09a-e2b7-46db-b938-f276df3f033e-registry-certificates\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.759683 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73-metrics-tls\") pod \"ingress-operator-5b745b69d9-jrkwc\" (UID: \"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.759708 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9rrcn\" (UID: \"a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.759751 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9cd4c3b3-6825-4bd2-97a5-330f91782d4b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rvkzz\" (UID: \"9cd4c3b3-6825-4bd2-97a5-330f91782d4b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.759781 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f56fc09a-e2b7-46db-b938-f276df3f033e-trusted-ca\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.759852 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9fc6g\" (UniqueName: \"kubernetes.io/projected/f660255f-8f78-4876-973d-db58f2ee7020-kube-api-access-9fc6g\") pod \"openshift-config-operator-7777fb866f-sk2l6\" (UID: \"f660255f-8f78-4876-973d-db58f2ee7020\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.759883 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f660255f-8f78-4876-973d-db58f2ee7020-serving-cert\") pod \"openshift-config-operator-7777fb866f-sk2l6\" (UID: \"f660255f-8f78-4876-973d-db58f2ee7020\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.759932 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-serving-cert\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.759997 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k7tt\" (UniqueName: \"kubernetes.io/projected/32f62e32-732b-4646-85f0-45b8ea6544a6-kube-api-access-4k7tt\") pod \"catalog-operator-68c6474976-r5v5p\" (UID: \"32f62e32-732b-4646-85f0-45b8ea6544a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.760016 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8be807d4-9bc2-41a1-b69f-1b0af031b5ab-proxy-tls\") pod \"machine-config-controller-84d6567774-rntsr\" (UID: \"8be807d4-9bc2-41a1-b69f-1b0af031b5ab\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.760063 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9cd4c3b3-6825-4bd2-97a5-330f91782d4b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rvkzz\" (UID: \"9cd4c3b3-6825-4bd2-97a5-330f91782d4b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.760122 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-service-ca-bundle\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.761981 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9rrcn\" (UID: \"a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.762067 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef269b18-ea84-43c2-971c-e772149acbf6-serving-cert\") pod \"console-operator-58897d9998-sbtn5\" (UID: \"ef269b18-ea84-43c2-971c-e772149acbf6\") " pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.762126 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b959a85-56a5-4296-9cf3-87741e1f9c39-service-ca-bundle\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.762165 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef269b18-ea84-43c2-971c-e772149acbf6-config\") pod \"console-operator-58897d9998-sbtn5\" (UID: \"ef269b18-ea84-43c2-971c-e772149acbf6\") " pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.762197 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-registry-tls\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.762217 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2493565c-3af9-4edf-a2f3-8a7a501e9305-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l6ndt\" (UID: \"2493565c-3af9-4edf-a2f3-8a7a501e9305\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.762247 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxh5f\" (UniqueName: \"kubernetes.io/projected/f781cb50-1e1b-4586-ba59-b204b1a6beec-kube-api-access-hxh5f\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.762337 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ghtc\" (UniqueName: \"kubernetes.io/projected/b7b8ca1c-c3de-4829-ab9f-860f76033c63-kube-api-access-2ghtc\") pod \"openshift-controller-manager-operator-756b6f6bc6-hd8rx\" (UID: \"b7b8ca1c-c3de-4829-ab9f-860f76033c63\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.763636 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.763726 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l6kft\" (UniqueName: \"kubernetes.io/projected/8be807d4-9bc2-41a1-b69f-1b0af031b5ab-kube-api-access-l6kft\") pod \"machine-config-controller-84d6567774-rntsr\" (UID: \"8be807d4-9bc2-41a1-b69f-1b0af031b5ab\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.763780 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f56fc09a-e2b7-46db-b938-f276df3f033e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.763804 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2493565c-3af9-4edf-a2f3-8a7a501e9305-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l6ndt\" (UID: \"2493565c-3af9-4edf-a2f3-8a7a501e9305\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.763824 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rq5z\" (UniqueName: \"kubernetes.io/projected/96067558-b20b-411c-b1af-b8fbb61df8f7-kube-api-access-9rq5z\") pod \"migrator-59844c95c7-9wxcs\" (UID: \"96067558-b20b-411c-b1af-b8fbb61df8f7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9wxcs" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.763881 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mprmj\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-kube-api-access-mprmj\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.763906 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8be807d4-9bc2-41a1-b69f-1b0af031b5ab-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rntsr\" (UID: \"8be807d4-9bc2-41a1-b69f-1b0af031b5ab\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.763995 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7b959a85-56a5-4296-9cf3-87741e1f9c39-default-certificate\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.764017 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f660255f-8f78-4876-973d-db58f2ee7020-available-featuregates\") pod \"openshift-config-operator-7777fb866f-sk2l6\" (UID: \"f660255f-8f78-4876-973d-db58f2ee7020\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.764051 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.764101 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-bound-sa-token\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.764121 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/32f62e32-732b-4646-85f0-45b8ea6544a6-srv-cert\") pod \"catalog-operator-68c6474976-r5v5p\" (UID: \"32f62e32-732b-4646-85f0-45b8ea6544a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.764141 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txsm4\" (UniqueName: \"kubernetes.io/projected/e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e-kube-api-access-txsm4\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb4c6\" (UID: \"e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.764211 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-config\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.764238 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.764260 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2493565c-3af9-4edf-a2f3-8a7a501e9305-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l6ndt\" (UID: \"2493565c-3af9-4edf-a2f3-8a7a501e9305\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.764276 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22993daf-2b32-4be5-8eb7-f9194e903d62-config\") pod \"kube-apiserver-operator-766d6c64bb-946gp\" (UID: \"22993daf-2b32-4be5-8eb7-f9194e903d62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.764294 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7b8ca1c-c3de-4829-ab9f-860f76033c63-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hd8rx\" (UID: \"b7b8ca1c-c3de-4829-ab9f-860f76033c63\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx" Mar 13 10:07:45 crc kubenswrapper[4632]: E0313 10:07:45.765800 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:46.265771738 +0000 UTC m=+240.288301871 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.790208 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.824316 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc"] Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865048 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865255 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f660255f-8f78-4876-973d-db58f2ee7020-serving-cert\") pod \"openshift-config-operator-7777fb866f-sk2l6\" (UID: \"f660255f-8f78-4876-973d-db58f2ee7020\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865286 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/946f5fcb-dde4-4784-965d-75a47187e703-signing-key\") pod \"service-ca-9c57cc56f-68mjx\" (UID: \"946f5fcb-dde4-4784-965d-75a47187e703\") " pod="openshift-service-ca/service-ca-9c57cc56f-68mjx" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865308 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/946f5fcb-dde4-4784-965d-75a47187e703-signing-cabundle\") pod \"service-ca-9c57cc56f-68mjx\" (UID: \"946f5fcb-dde4-4784-965d-75a47187e703\") " pod="openshift-service-ca/service-ca-9c57cc56f-68mjx" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865332 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l2w2\" (UniqueName: \"kubernetes.io/projected/e0e1f142-2930-4f9b-b851-f7f7df22676b-kube-api-access-8l2w2\") pod \"multus-admission-controller-857f4d67dd-4hmjh\" (UID: \"e0e1f142-2930-4f9b-b851-f7f7df22676b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4hmjh" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865355 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pst8h\" (UniqueName: \"kubernetes.io/projected/ebf1040d-57dd-47ef-b839-6f78a7c5c75f-kube-api-access-pst8h\") pod \"olm-operator-6b444d44fb-tqbl9\" (UID: \"ebf1040d-57dd-47ef-b839-6f78a7c5c75f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865375 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f0f88609-cbfe-4ccc-b5db-e5c1be771855-certs\") pod \"machine-config-server-hlf9t\" (UID: \"f0f88609-cbfe-4ccc-b5db-e5c1be771855\") " pod="openshift-machine-config-operator/machine-config-server-hlf9t" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865400 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-serving-cert\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865435 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k7tt\" (UniqueName: \"kubernetes.io/projected/32f62e32-732b-4646-85f0-45b8ea6544a6-kube-api-access-4k7tt\") pod \"catalog-operator-68c6474976-r5v5p\" (UID: \"32f62e32-732b-4646-85f0-45b8ea6544a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865456 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8be807d4-9bc2-41a1-b69f-1b0af031b5ab-proxy-tls\") pod \"machine-config-controller-84d6567774-rntsr\" (UID: \"8be807d4-9bc2-41a1-b69f-1b0af031b5ab\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865478 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9cd4c3b3-6825-4bd2-97a5-330f91782d4b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rvkzz\" (UID: \"9cd4c3b3-6825-4bd2-97a5-330f91782d4b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865500 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/528d3aa9-10bf-4029-a4d2-85768264fde8-secret-volume\") pod \"collect-profiles-29556600-r9flg\" (UID: \"528d3aa9-10bf-4029-a4d2-85768264fde8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865523 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-service-ca-bundle\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865544 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9rrcn\" (UID: \"a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865565 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4db028f0-524e-46fc-aa33-da38ed7b8fa6-cert\") pod \"ingress-canary-qcb4l\" (UID: \"4db028f0-524e-46fc-aa33-da38ed7b8fa6\") " pod="openshift-ingress-canary/ingress-canary-qcb4l" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865585 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef269b18-ea84-43c2-971c-e772149acbf6-serving-cert\") pod \"console-operator-58897d9998-sbtn5\" (UID: \"ef269b18-ea84-43c2-971c-e772149acbf6\") " pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865628 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-registration-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865651 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b959a85-56a5-4296-9cf3-87741e1f9c39-service-ca-bundle\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865673 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hcvp\" (UniqueName: \"kubernetes.io/projected/49c520f1-fb05-48ca-8435-1985ce668451-kube-api-access-2hcvp\") pod \"packageserver-d55dfcdfc-zgxcd\" (UID: \"49c520f1-fb05-48ca-8435-1985ce668451\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865693 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef269b18-ea84-43c2-971c-e772149acbf6-config\") pod \"console-operator-58897d9998-sbtn5\" (UID: \"ef269b18-ea84-43c2-971c-e772149acbf6\") " pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865716 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2332524f-f990-4ef2-90b3-8b90c389d873-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pvwll\" (UID: \"2332524f-f990-4ef2-90b3-8b90c389d873\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pvwll" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865756 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-registry-tls\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865775 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2493565c-3af9-4edf-a2f3-8a7a501e9305-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l6ndt\" (UID: \"2493565c-3af9-4edf-a2f3-8a7a501e9305\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865796 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxh5f\" (UniqueName: \"kubernetes.io/projected/f781cb50-1e1b-4586-ba59-b204b1a6beec-kube-api-access-hxh5f\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865820 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ghtc\" (UniqueName: \"kubernetes.io/projected/b7b8ca1c-c3de-4829-ab9f-860f76033c63-kube-api-access-2ghtc\") pod \"openshift-controller-manager-operator-756b6f6bc6-hd8rx\" (UID: \"b7b8ca1c-c3de-4829-ab9f-860f76033c63\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865845 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm8lp\" (UniqueName: \"kubernetes.io/projected/528d3aa9-10bf-4029-a4d2-85768264fde8-kube-api-access-vm8lp\") pod \"collect-profiles-29556600-r9flg\" (UID: \"528d3aa9-10bf-4029-a4d2-85768264fde8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865868 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l6kft\" (UniqueName: \"kubernetes.io/projected/8be807d4-9bc2-41a1-b69f-1b0af031b5ab-kube-api-access-l6kft\") pod \"machine-config-controller-84d6567774-rntsr\" (UID: \"8be807d4-9bc2-41a1-b69f-1b0af031b5ab\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865891 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f56fc09a-e2b7-46db-b938-f276df3f033e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865912 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2493565c-3af9-4edf-a2f3-8a7a501e9305-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l6ndt\" (UID: \"2493565c-3af9-4edf-a2f3-8a7a501e9305\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865932 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rq5z\" (UniqueName: \"kubernetes.io/projected/96067558-b20b-411c-b1af-b8fbb61df8f7-kube-api-access-9rq5z\") pod \"migrator-59844c95c7-9wxcs\" (UID: \"96067558-b20b-411c-b1af-b8fbb61df8f7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9wxcs" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.865975 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5h9s\" (UniqueName: \"kubernetes.io/projected/4db028f0-524e-46fc-aa33-da38ed7b8fa6-kube-api-access-p5h9s\") pod \"ingress-canary-qcb4l\" (UID: \"4db028f0-524e-46fc-aa33-da38ed7b8fa6\") " pod="openshift-ingress-canary/ingress-canary-qcb4l" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866007 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mprmj\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-kube-api-access-mprmj\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866031 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8be807d4-9bc2-41a1-b69f-1b0af031b5ab-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rntsr\" (UID: \"8be807d4-9bc2-41a1-b69f-1b0af031b5ab\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866053 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7b959a85-56a5-4296-9cf3-87741e1f9c39-default-certificate\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866077 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ebf1040d-57dd-47ef-b839-6f78a7c5c75f-srv-cert\") pod \"olm-operator-6b444d44fb-tqbl9\" (UID: \"ebf1040d-57dd-47ef-b839-6f78a7c5c75f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866100 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f660255f-8f78-4876-973d-db58f2ee7020-available-featuregates\") pod \"openshift-config-operator-7777fb866f-sk2l6\" (UID: \"f660255f-8f78-4876-973d-db58f2ee7020\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866121 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/49c520f1-fb05-48ca-8435-1985ce668451-tmpfs\") pod \"packageserver-d55dfcdfc-zgxcd\" (UID: \"49c520f1-fb05-48ca-8435-1985ce668451\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866154 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-bound-sa-token\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866177 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f0f88609-cbfe-4ccc-b5db-e5c1be771855-node-bootstrap-token\") pod \"machine-config-server-hlf9t\" (UID: \"f0f88609-cbfe-4ccc-b5db-e5c1be771855\") " pod="openshift-machine-config-operator/machine-config-server-hlf9t" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866214 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/32f62e32-732b-4646-85f0-45b8ea6544a6-srv-cert\") pod \"catalog-operator-68c6474976-r5v5p\" (UID: \"32f62e32-732b-4646-85f0-45b8ea6544a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866235 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/779b2915-e0d0-4e90-9c6d-af28f555fd7b-config\") pod \"service-ca-operator-777779d784-zh465\" (UID: \"779b2915-e0d0-4e90-9c6d-af28f555fd7b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zh465" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866257 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txsm4\" (UniqueName: \"kubernetes.io/projected/e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e-kube-api-access-txsm4\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb4c6\" (UID: \"e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866281 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-config\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866301 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866326 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2vc6\" (UniqueName: \"kubernetes.io/projected/779b2915-e0d0-4e90-9c6d-af28f555fd7b-kube-api-access-q2vc6\") pod \"service-ca-operator-777779d784-zh465\" (UID: \"779b2915-e0d0-4e90-9c6d-af28f555fd7b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zh465" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866351 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2493565c-3af9-4edf-a2f3-8a7a501e9305-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l6ndt\" (UID: \"2493565c-3af9-4edf-a2f3-8a7a501e9305\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866373 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22993daf-2b32-4be5-8eb7-f9194e903d62-config\") pod \"kube-apiserver-operator-766d6c64bb-946gp\" (UID: \"22993daf-2b32-4be5-8eb7-f9194e903d62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866395 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7b8ca1c-c3de-4829-ab9f-860f76033c63-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hd8rx\" (UID: \"b7b8ca1c-c3de-4829-ab9f-860f76033c63\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866417 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e0e1f142-2930-4f9b-b851-f7f7df22676b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4hmjh\" (UID: \"e0e1f142-2930-4f9b-b851-f7f7df22676b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4hmjh" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866439 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22993daf-2b32-4be5-8eb7-f9194e903d62-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-946gp\" (UID: \"22993daf-2b32-4be5-8eb7-f9194e903d62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866465 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvs6x\" (UniqueName: \"kubernetes.io/projected/bbb27a61-7407-4cd7-84df-4b66fbdcf82d-kube-api-access-pvs6x\") pod \"machine-config-operator-74547568cd-99hff\" (UID: \"bbb27a61-7407-4cd7-84df-4b66fbdcf82d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866486 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/32f62e32-732b-4646-85f0-45b8ea6544a6-profile-collector-cert\") pod \"catalog-operator-68c6474976-r5v5p\" (UID: \"32f62e32-732b-4646-85f0-45b8ea6544a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866507 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7b8ca1c-c3de-4829-ab9f-860f76033c63-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hd8rx\" (UID: \"b7b8ca1c-c3de-4829-ab9f-860f76033c63\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866530 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/797176c6-dd56-48d6-8004-ff1dd5353a50-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2n99d\" (UID: \"797176c6-dd56-48d6-8004-ff1dd5353a50\") " pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866555 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cclg\" (UniqueName: \"kubernetes.io/projected/ef269b18-ea84-43c2-971c-e772149acbf6-kube-api-access-2cclg\") pod \"console-operator-58897d9998-sbtn5\" (UID: \"ef269b18-ea84-43c2-971c-e772149acbf6\") " pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866577 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb4c6\" (UID: \"e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866598 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f781cb50-1e1b-4586-ba59-b204b1a6beec-config\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866617 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f781cb50-1e1b-4586-ba59-b204b1a6beec-etcd-service-ca\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866640 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xx2pr\" (UniqueName: \"kubernetes.io/projected/f0f88609-cbfe-4ccc-b5db-e5c1be771855-kube-api-access-xx2pr\") pod \"machine-config-server-hlf9t\" (UID: \"f0f88609-cbfe-4ccc-b5db-e5c1be771855\") " pod="openshift-machine-config-operator/machine-config-server-hlf9t" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866662 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qtrb\" (UniqueName: \"kubernetes.io/projected/797176c6-dd56-48d6-8004-ff1dd5353a50-kube-api-access-8qtrb\") pod \"marketplace-operator-79b997595-2n99d\" (UID: \"797176c6-dd56-48d6-8004-ff1dd5353a50\") " pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866686 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f781cb50-1e1b-4586-ba59-b204b1a6beec-etcd-ca\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866720 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bbb27a61-7407-4cd7-84df-4b66fbdcf82d-proxy-tls\") pod \"machine-config-operator-74547568cd-99hff\" (UID: \"bbb27a61-7407-4cd7-84df-4b66fbdcf82d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866741 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7mb4q\" (UniqueName: \"kubernetes.io/projected/c94773d8-a922-4778-b2ba-8937e9d6c19b-kube-api-access-7mb4q\") pod \"dns-operator-744455d44c-6fqf5\" (UID: \"c94773d8-a922-4778-b2ba-8937e9d6c19b\") " pod="openshift-dns-operator/dns-operator-744455d44c-6fqf5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866764 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9rrcn\" (UID: \"a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866786 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb4c6\" (UID: \"e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866808 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7b959a85-56a5-4296-9cf3-87741e1f9c39-stats-auth\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866829 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73-trusted-ca\") pod \"ingress-operator-5b745b69d9-jrkwc\" (UID: \"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866847 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7b959a85-56a5-4296-9cf3-87741e1f9c39-metrics-certs\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866861 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62kgq\" (UniqueName: \"kubernetes.io/projected/7b959a85-56a5-4296-9cf3-87741e1f9c39-kube-api-access-62kgq\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866882 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ebf1040d-57dd-47ef-b839-6f78a7c5c75f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tqbl9\" (UID: \"ebf1040d-57dd-47ef-b839-6f78a7c5c75f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866902 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/49c520f1-fb05-48ca-8435-1985ce668451-apiservice-cert\") pod \"packageserver-d55dfcdfc-zgxcd\" (UID: \"49c520f1-fb05-48ca-8435-1985ce668451\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866921 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78rbb\" (UniqueName: \"kubernetes.io/projected/58d59f3d-e656-4217-9472-62508a7ccc93-kube-api-access-78rbb\") pod \"dns-default-g2wxc\" (UID: \"58d59f3d-e656-4217-9472-62508a7ccc93\") " pod="openshift-dns/dns-default-g2wxc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866970 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef269b18-ea84-43c2-971c-e772149acbf6-trusted-ca\") pod \"console-operator-58897d9998-sbtn5\" (UID: \"ef269b18-ea84-43c2-971c-e772149acbf6\") " pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.866996 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f56fc09a-e2b7-46db-b938-f276df3f033e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867012 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sd756\" (UniqueName: \"kubernetes.io/projected/c822257d-9d2f-4b6f-87de-131de5cd0efe-kube-api-access-sd756\") pod \"auto-csr-approver-29556606-mkrp2\" (UID: \"c822257d-9d2f-4b6f-87de-131de5cd0efe\") " pod="openshift-infra/auto-csr-approver-29556606-mkrp2" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867027 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-plugins-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867044 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22993daf-2b32-4be5-8eb7-f9194e903d62-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-946gp\" (UID: \"22993daf-2b32-4be5-8eb7-f9194e903d62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867059 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/58d59f3d-e656-4217-9472-62508a7ccc93-metrics-tls\") pod \"dns-default-g2wxc\" (UID: \"58d59f3d-e656-4217-9472-62508a7ccc93\") " pod="openshift-dns/dns-default-g2wxc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867077 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f781cb50-1e1b-4586-ba59-b204b1a6beec-serving-cert\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867093 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c94773d8-a922-4778-b2ba-8937e9d6c19b-metrics-tls\") pod \"dns-operator-744455d44c-6fqf5\" (UID: \"c94773d8-a922-4778-b2ba-8937e9d6c19b\") " pod="openshift-dns-operator/dns-operator-744455d44c-6fqf5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867115 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4e100e6e-7259-4262-be47-9c2b5be7a53a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xfvsc\" (UID: \"4e100e6e-7259-4262-be47-9c2b5be7a53a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867139 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52zmx\" (UniqueName: \"kubernetes.io/projected/fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73-kube-api-access-52zmx\") pod \"ingress-operator-5b745b69d9-jrkwc\" (UID: \"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867159 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwlxd\" (UniqueName: \"kubernetes.io/projected/4e100e6e-7259-4262-be47-9c2b5be7a53a-kube-api-access-wwlxd\") pod \"package-server-manager-789f6589d5-xfvsc\" (UID: \"4e100e6e-7259-4262-be47-9c2b5be7a53a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867179 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-socket-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867222 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsqdg\" (UniqueName: \"kubernetes.io/projected/946f5fcb-dde4-4784-965d-75a47187e703-kube-api-access-vsqdg\") pod \"service-ca-9c57cc56f-68mjx\" (UID: \"946f5fcb-dde4-4784-965d-75a47187e703\") " pod="openshift-service-ca/service-ca-9c57cc56f-68mjx" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867240 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bbb27a61-7407-4cd7-84df-4b66fbdcf82d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-99hff\" (UID: \"bbb27a61-7407-4cd7-84df-4b66fbdcf82d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867255 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58d59f3d-e656-4217-9472-62508a7ccc93-config-volume\") pod \"dns-default-g2wxc\" (UID: \"58d59f3d-e656-4217-9472-62508a7ccc93\") " pod="openshift-dns/dns-default-g2wxc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867277 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73-bound-sa-token\") pod \"ingress-operator-5b745b69d9-jrkwc\" (UID: \"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867298 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f781cb50-1e1b-4586-ba59-b204b1a6beec-etcd-client\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867319 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cd4c3b3-6825-4bd2-97a5-330f91782d4b-config\") pod \"kube-controller-manager-operator-78b949d7b-rvkzz\" (UID: \"9cd4c3b3-6825-4bd2-97a5-330f91782d4b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867336 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bbb27a61-7407-4cd7-84df-4b66fbdcf82d-images\") pod \"machine-config-operator-74547568cd-99hff\" (UID: \"bbb27a61-7407-4cd7-84df-4b66fbdcf82d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867351 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7x7q\" (UniqueName: \"kubernetes.io/projected/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-kube-api-access-w7x7q\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867370 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zq92\" (UniqueName: \"kubernetes.io/projected/a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59-kube-api-access-8zq92\") pod \"cluster-image-registry-operator-dc59b4c8b-9rrcn\" (UID: \"a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867385 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-csi-data-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867400 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/797176c6-dd56-48d6-8004-ff1dd5353a50-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2n99d\" (UID: \"797176c6-dd56-48d6-8004-ff1dd5353a50\") " pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867416 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f56fc09a-e2b7-46db-b938-f276df3f033e-registry-certificates\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867432 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73-metrics-tls\") pod \"ingress-operator-5b745b69d9-jrkwc\" (UID: \"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867448 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9rrcn\" (UID: \"a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867462 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-mountpoint-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867481 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9cd4c3b3-6825-4bd2-97a5-330f91782d4b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rvkzz\" (UID: \"9cd4c3b3-6825-4bd2-97a5-330f91782d4b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867497 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/49c520f1-fb05-48ca-8435-1985ce668451-webhook-cert\") pod \"packageserver-d55dfcdfc-zgxcd\" (UID: \"49c520f1-fb05-48ca-8435-1985ce668451\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867511 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/528d3aa9-10bf-4029-a4d2-85768264fde8-config-volume\") pod \"collect-profiles-29556600-r9flg\" (UID: \"528d3aa9-10bf-4029-a4d2-85768264fde8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867526 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/779b2915-e0d0-4e90-9c6d-af28f555fd7b-serving-cert\") pod \"service-ca-operator-777779d784-zh465\" (UID: \"779b2915-e0d0-4e90-9c6d-af28f555fd7b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zh465" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867544 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f56fc09a-e2b7-46db-b938-f276df3f033e-trusted-ca\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867562 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9fc6g\" (UniqueName: \"kubernetes.io/projected/f660255f-8f78-4876-973d-db58f2ee7020-kube-api-access-9fc6g\") pod \"openshift-config-operator-7777fb866f-sk2l6\" (UID: \"f660255f-8f78-4876-973d-db58f2ee7020\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867578 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wc9r4\" (UniqueName: \"kubernetes.io/projected/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-kube-api-access-wc9r4\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.867594 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hw6wb\" (UniqueName: \"kubernetes.io/projected/2332524f-f990-4ef2-90b3-8b90c389d873-kube-api-access-hw6wb\") pod \"control-plane-machine-set-operator-78cbb6b69f-pvwll\" (UID: \"2332524f-f990-4ef2-90b3-8b90c389d873\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pvwll" Mar 13 10:07:45 crc kubenswrapper[4632]: E0313 10:07:45.867731 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:46.367716797 +0000 UTC m=+240.390246930 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.869830 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef269b18-ea84-43c2-971c-e772149acbf6-config\") pod \"console-operator-58897d9998-sbtn5\" (UID: \"ef269b18-ea84-43c2-971c-e772149acbf6\") " pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.871636 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f781cb50-1e1b-4586-ba59-b204b1a6beec-config\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.872300 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/f781cb50-1e1b-4586-ba59-b204b1a6beec-etcd-service-ca\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.872831 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/f781cb50-1e1b-4586-ba59-b204b1a6beec-etcd-ca\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.876726 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-service-ca-bundle\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.880762 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/f660255f-8f78-4876-973d-db58f2ee7020-available-featuregates\") pod \"openshift-config-operator-7777fb866f-sk2l6\" (UID: \"f660255f-8f78-4876-973d-db58f2ee7020\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.881249 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ef269b18-ea84-43c2-971c-e772149acbf6-trusted-ca\") pod \"console-operator-58897d9998-sbtn5\" (UID: \"ef269b18-ea84-43c2-971c-e772149acbf6\") " pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.883755 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/8be807d4-9bc2-41a1-b69f-1b0af031b5ab-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-rntsr\" (UID: \"8be807d4-9bc2-41a1-b69f-1b0af031b5ab\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.885110 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f56fc09a-e2b7-46db-b938-f276df3f033e-ca-trust-extracted\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.885757 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2493565c-3af9-4edf-a2f3-8a7a501e9305-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l6ndt\" (UID: \"2493565c-3af9-4edf-a2f3-8a7a501e9305\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.886533 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-registry-tls\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.887915 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bbb27a61-7407-4cd7-84df-4b66fbdcf82d-auth-proxy-config\") pod \"machine-config-operator-74547568cd-99hff\" (UID: \"bbb27a61-7407-4cd7-84df-4b66fbdcf82d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.888886 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f660255f-8f78-4876-973d-db58f2ee7020-serving-cert\") pod \"openshift-config-operator-7777fb866f-sk2l6\" (UID: \"f660255f-8f78-4876-973d-db58f2ee7020\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.890819 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/8be807d4-9bc2-41a1-b69f-1b0af031b5ab-proxy-tls\") pod \"machine-config-controller-84d6567774-rntsr\" (UID: \"8be807d4-9bc2-41a1-b69f-1b0af031b5ab\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.891455 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ef269b18-ea84-43c2-971c-e772149acbf6-serving-cert\") pod \"console-operator-58897d9998-sbtn5\" (UID: \"ef269b18-ea84-43c2-971c-e772149acbf6\") " pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.894164 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9cd4c3b3-6825-4bd2-97a5-330f91782d4b-config\") pod \"kube-controller-manager-operator-78b949d7b-rvkzz\" (UID: \"9cd4c3b3-6825-4bd2-97a5-330f91782d4b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.898753 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-config\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.899277 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73-trusted-ca\") pod \"ingress-operator-5b745b69d9-jrkwc\" (UID: \"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.899622 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22993daf-2b32-4be5-8eb7-f9194e903d62-config\") pod \"kube-apiserver-operator-766d6c64bb-946gp\" (UID: \"22993daf-2b32-4be5-8eb7-f9194e903d62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.899686 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-9rrcn\" (UID: \"a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.899772 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7b959a85-56a5-4296-9cf3-87741e1f9c39-service-ca-bundle\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.899973 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/bbb27a61-7407-4cd7-84df-4b66fbdcf82d-images\") pod \"machine-config-operator-74547568cd-99hff\" (UID: \"bbb27a61-7407-4cd7-84df-4b66fbdcf82d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.900105 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c94773d8-a922-4778-b2ba-8937e9d6c19b-metrics-tls\") pod \"dns-operator-744455d44c-6fqf5\" (UID: \"c94773d8-a922-4778-b2ba-8937e9d6c19b\") " pod="openshift-dns-operator/dns-operator-744455d44c-6fqf5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.903177 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.904116 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f56fc09a-e2b7-46db-b938-f276df3f033e-registry-certificates\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.907091 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb4c6\" (UID: \"e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.910089 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn"] Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.910153 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9v5nn"] Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.911298 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f56fc09a-e2b7-46db-b938-f276df3f033e-trusted-ca\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.911902 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7b8ca1c-c3de-4829-ab9f-860f76033c63-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-hd8rx\" (UID: \"b7b8ca1c-c3de-4829-ab9f-860f76033c63\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.917025 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7b8ca1c-c3de-4829-ab9f-860f76033c63-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-hd8rx\" (UID: \"b7b8ca1c-c3de-4829-ab9f-860f76033c63\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.918394 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f781cb50-1e1b-4586-ba59-b204b1a6beec-serving-cert\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.918545 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f56fc09a-e2b7-46db-b938-f276df3f033e-installation-pull-secrets\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.918592 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/7b959a85-56a5-4296-9cf3-87741e1f9c39-default-certificate\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.918866 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f781cb50-1e1b-4586-ba59-b204b1a6beec-etcd-client\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.919123 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/7b959a85-56a5-4296-9cf3-87741e1f9c39-stats-auth\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.919229 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-serving-cert\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.919645 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-9rrcn\" (UID: \"a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.921549 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/32f62e32-732b-4646-85f0-45b8ea6544a6-srv-cert\") pod \"catalog-operator-68c6474976-r5v5p\" (UID: \"32f62e32-732b-4646-85f0-45b8ea6544a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.922152 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k7tt\" (UniqueName: \"kubernetes.io/projected/32f62e32-732b-4646-85f0-45b8ea6544a6-kube-api-access-4k7tt\") pod \"catalog-operator-68c6474976-r5v5p\" (UID: \"32f62e32-732b-4646-85f0-45b8ea6544a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.924347 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2493565c-3af9-4edf-a2f3-8a7a501e9305-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l6ndt\" (UID: \"2493565c-3af9-4edf-a2f3-8a7a501e9305\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.924816 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb4c6\" (UID: \"e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.924818 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73-metrics-tls\") pod \"ingress-operator-5b745b69d9-jrkwc\" (UID: \"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.925318 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bbb27a61-7407-4cd7-84df-4b66fbdcf82d-proxy-tls\") pod \"machine-config-operator-74547568cd-99hff\" (UID: \"bbb27a61-7407-4cd7-84df-4b66fbdcf82d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.925458 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7b959a85-56a5-4296-9cf3-87741e1f9c39-metrics-certs\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.925486 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9cd4c3b3-6825-4bd2-97a5-330f91782d4b-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rvkzz\" (UID: \"9cd4c3b3-6825-4bd2-97a5-330f91782d4b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.936158 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9cd4c3b3-6825-4bd2-97a5-330f91782d4b-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rvkzz\" (UID: \"9cd4c3b3-6825-4bd2-97a5-330f91782d4b\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.938515 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/32f62e32-732b-4646-85f0-45b8ea6544a6-profile-collector-cert\") pod \"catalog-operator-68c6474976-r5v5p\" (UID: \"32f62e32-732b-4646-85f0-45b8ea6544a6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 10:07:45 crc kubenswrapper[4632]: W0313 10:07:45.938756 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c90710f_4595_425c_8be1_1436f43b5069.slice/crio-0fe3af461e6c99b0dddea09fd2fc17bc9781ce95850249ce166d4812660ac046 WatchSource:0}: Error finding container 0fe3af461e6c99b0dddea09fd2fc17bc9781ce95850249ce166d4812660ac046: Status 404 returned error can't find the container with id 0fe3af461e6c99b0dddea09fd2fc17bc9781ce95850249ce166d4812660ac046 Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.941735 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22993daf-2b32-4be5-8eb7-f9194e903d62-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-946gp\" (UID: \"22993daf-2b32-4be5-8eb7-f9194e903d62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.957482 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cclg\" (UniqueName: \"kubernetes.io/projected/ef269b18-ea84-43c2-971c-e772149acbf6-kube-api-access-2cclg\") pod \"console-operator-58897d9998-sbtn5\" (UID: \"ef269b18-ea84-43c2-971c-e772149acbf6\") " pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.969382 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2"] Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.970714 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/528d3aa9-10bf-4029-a4d2-85768264fde8-secret-volume\") pod \"collect-profiles-29556600-r9flg\" (UID: \"528d3aa9-10bf-4029-a4d2-85768264fde8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.970733 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62kgq\" (UniqueName: \"kubernetes.io/projected/7b959a85-56a5-4296-9cf3-87741e1f9c39-kube-api-access-62kgq\") pod \"router-default-5444994796-t9vht\" (UID: \"7b959a85-56a5-4296-9cf3-87741e1f9c39\") " pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.970760 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4db028f0-524e-46fc-aa33-da38ed7b8fa6-cert\") pod \"ingress-canary-qcb4l\" (UID: \"4db028f0-524e-46fc-aa33-da38ed7b8fa6\") " pod="openshift-ingress-canary/ingress-canary-qcb4l" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.970789 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-registration-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.970823 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hcvp\" (UniqueName: \"kubernetes.io/projected/49c520f1-fb05-48ca-8435-1985ce668451-kube-api-access-2hcvp\") pod \"packageserver-d55dfcdfc-zgxcd\" (UID: \"49c520f1-fb05-48ca-8435-1985ce668451\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.970849 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2332524f-f990-4ef2-90b3-8b90c389d873-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pvwll\" (UID: \"2332524f-f990-4ef2-90b3-8b90c389d873\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pvwll" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.970900 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm8lp\" (UniqueName: \"kubernetes.io/projected/528d3aa9-10bf-4029-a4d2-85768264fde8-kube-api-access-vm8lp\") pod \"collect-profiles-29556600-r9flg\" (UID: \"528d3aa9-10bf-4029-a4d2-85768264fde8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.970928 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5h9s\" (UniqueName: \"kubernetes.io/projected/4db028f0-524e-46fc-aa33-da38ed7b8fa6-kube-api-access-p5h9s\") pod \"ingress-canary-qcb4l\" (UID: \"4db028f0-524e-46fc-aa33-da38ed7b8fa6\") " pod="openshift-ingress-canary/ingress-canary-qcb4l" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971065 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ebf1040d-57dd-47ef-b839-6f78a7c5c75f-srv-cert\") pod \"olm-operator-6b444d44fb-tqbl9\" (UID: \"ebf1040d-57dd-47ef-b839-6f78a7c5c75f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971082 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/49c520f1-fb05-48ca-8435-1985ce668451-tmpfs\") pod \"packageserver-d55dfcdfc-zgxcd\" (UID: \"49c520f1-fb05-48ca-8435-1985ce668451\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971116 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971141 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f0f88609-cbfe-4ccc-b5db-e5c1be771855-node-bootstrap-token\") pod \"machine-config-server-hlf9t\" (UID: \"f0f88609-cbfe-4ccc-b5db-e5c1be771855\") " pod="openshift-machine-config-operator/machine-config-server-hlf9t" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971156 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/779b2915-e0d0-4e90-9c6d-af28f555fd7b-config\") pod \"service-ca-operator-777779d784-zh465\" (UID: \"779b2915-e0d0-4e90-9c6d-af28f555fd7b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zh465" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971201 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e0e1f142-2930-4f9b-b851-f7f7df22676b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4hmjh\" (UID: \"e0e1f142-2930-4f9b-b851-f7f7df22676b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4hmjh" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971218 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2vc6\" (UniqueName: \"kubernetes.io/projected/779b2915-e0d0-4e90-9c6d-af28f555fd7b-kube-api-access-q2vc6\") pod \"service-ca-operator-777779d784-zh465\" (UID: \"779b2915-e0d0-4e90-9c6d-af28f555fd7b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zh465" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971249 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/797176c6-dd56-48d6-8004-ff1dd5353a50-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2n99d\" (UID: \"797176c6-dd56-48d6-8004-ff1dd5353a50\") " pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971295 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xx2pr\" (UniqueName: \"kubernetes.io/projected/f0f88609-cbfe-4ccc-b5db-e5c1be771855-kube-api-access-xx2pr\") pod \"machine-config-server-hlf9t\" (UID: \"f0f88609-cbfe-4ccc-b5db-e5c1be771855\") " pod="openshift-machine-config-operator/machine-config-server-hlf9t" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971314 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qtrb\" (UniqueName: \"kubernetes.io/projected/797176c6-dd56-48d6-8004-ff1dd5353a50-kube-api-access-8qtrb\") pod \"marketplace-operator-79b997595-2n99d\" (UID: \"797176c6-dd56-48d6-8004-ff1dd5353a50\") " pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971361 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ebf1040d-57dd-47ef-b839-6f78a7c5c75f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tqbl9\" (UID: \"ebf1040d-57dd-47ef-b839-6f78a7c5c75f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971375 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/49c520f1-fb05-48ca-8435-1985ce668451-apiservice-cert\") pod \"packageserver-d55dfcdfc-zgxcd\" (UID: \"49c520f1-fb05-48ca-8435-1985ce668451\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971390 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78rbb\" (UniqueName: \"kubernetes.io/projected/58d59f3d-e656-4217-9472-62508a7ccc93-kube-api-access-78rbb\") pod \"dns-default-g2wxc\" (UID: \"58d59f3d-e656-4217-9472-62508a7ccc93\") " pod="openshift-dns/dns-default-g2wxc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971432 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sd756\" (UniqueName: \"kubernetes.io/projected/c822257d-9d2f-4b6f-87de-131de5cd0efe-kube-api-access-sd756\") pod \"auto-csr-approver-29556606-mkrp2\" (UID: \"c822257d-9d2f-4b6f-87de-131de5cd0efe\") " pod="openshift-infra/auto-csr-approver-29556606-mkrp2" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971447 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-plugins-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971465 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/58d59f3d-e656-4217-9472-62508a7ccc93-metrics-tls\") pod \"dns-default-g2wxc\" (UID: \"58d59f3d-e656-4217-9472-62508a7ccc93\") " pod="openshift-dns/dns-default-g2wxc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971482 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4e100e6e-7259-4262-be47-9c2b5be7a53a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xfvsc\" (UID: \"4e100e6e-7259-4262-be47-9c2b5be7a53a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971516 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwlxd\" (UniqueName: \"kubernetes.io/projected/4e100e6e-7259-4262-be47-9c2b5be7a53a-kube-api-access-wwlxd\") pod \"package-server-manager-789f6589d5-xfvsc\" (UID: \"4e100e6e-7259-4262-be47-9c2b5be7a53a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971532 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-socket-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971554 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsqdg\" (UniqueName: \"kubernetes.io/projected/946f5fcb-dde4-4784-965d-75a47187e703-kube-api-access-vsqdg\") pod \"service-ca-9c57cc56f-68mjx\" (UID: \"946f5fcb-dde4-4784-965d-75a47187e703\") " pod="openshift-service-ca/service-ca-9c57cc56f-68mjx" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971591 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58d59f3d-e656-4217-9472-62508a7ccc93-config-volume\") pod \"dns-default-g2wxc\" (UID: \"58d59f3d-e656-4217-9472-62508a7ccc93\") " pod="openshift-dns/dns-default-g2wxc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971629 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-mountpoint-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971644 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-csi-data-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971686 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/797176c6-dd56-48d6-8004-ff1dd5353a50-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2n99d\" (UID: \"797176c6-dd56-48d6-8004-ff1dd5353a50\") " pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971702 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/49c520f1-fb05-48ca-8435-1985ce668451-webhook-cert\") pod \"packageserver-d55dfcdfc-zgxcd\" (UID: \"49c520f1-fb05-48ca-8435-1985ce668451\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971716 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/528d3aa9-10bf-4029-a4d2-85768264fde8-config-volume\") pod \"collect-profiles-29556600-r9flg\" (UID: \"528d3aa9-10bf-4029-a4d2-85768264fde8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971749 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/779b2915-e0d0-4e90-9c6d-af28f555fd7b-serving-cert\") pod \"service-ca-operator-777779d784-zh465\" (UID: \"779b2915-e0d0-4e90-9c6d-af28f555fd7b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zh465" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971770 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wc9r4\" (UniqueName: \"kubernetes.io/projected/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-kube-api-access-wc9r4\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971787 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hw6wb\" (UniqueName: \"kubernetes.io/projected/2332524f-f990-4ef2-90b3-8b90c389d873-kube-api-access-hw6wb\") pod \"control-plane-machine-set-operator-78cbb6b69f-pvwll\" (UID: \"2332524f-f990-4ef2-90b3-8b90c389d873\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pvwll" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971803 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/946f5fcb-dde4-4784-965d-75a47187e703-signing-key\") pod \"service-ca-9c57cc56f-68mjx\" (UID: \"946f5fcb-dde4-4784-965d-75a47187e703\") " pod="openshift-service-ca/service-ca-9c57cc56f-68mjx" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971836 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/946f5fcb-dde4-4784-965d-75a47187e703-signing-cabundle\") pod \"service-ca-9c57cc56f-68mjx\" (UID: \"946f5fcb-dde4-4784-965d-75a47187e703\") " pod="openshift-service-ca/service-ca-9c57cc56f-68mjx" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971854 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l2w2\" (UniqueName: \"kubernetes.io/projected/e0e1f142-2930-4f9b-b851-f7f7df22676b-kube-api-access-8l2w2\") pod \"multus-admission-controller-857f4d67dd-4hmjh\" (UID: \"e0e1f142-2930-4f9b-b851-f7f7df22676b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4hmjh" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971871 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pst8h\" (UniqueName: \"kubernetes.io/projected/ebf1040d-57dd-47ef-b839-6f78a7c5c75f-kube-api-access-pst8h\") pod \"olm-operator-6b444d44fb-tqbl9\" (UID: \"ebf1040d-57dd-47ef-b839-6f78a7c5c75f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.971884 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f0f88609-cbfe-4ccc-b5db-e5c1be771855-certs\") pod \"machine-config-server-hlf9t\" (UID: \"f0f88609-cbfe-4ccc-b5db-e5c1be771855\") " pod="openshift-machine-config-operator/machine-config-server-hlf9t" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.979222 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-p9gp2"] Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.979780 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-registration-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.980771 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-plugins-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.981118 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/779b2915-e0d0-4e90-9c6d-af28f555fd7b-config\") pod \"service-ca-operator-777779d784-zh465\" (UID: \"779b2915-e0d0-4e90-9c6d-af28f555fd7b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zh465" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.981337 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.985219 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/528d3aa9-10bf-4029-a4d2-85768264fde8-config-volume\") pod \"collect-profiles-29556600-r9flg\" (UID: \"528d3aa9-10bf-4029-a4d2-85768264fde8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.985668 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/49c520f1-fb05-48ca-8435-1985ce668451-tmpfs\") pod \"packageserver-d55dfcdfc-zgxcd\" (UID: \"49c520f1-fb05-48ca-8435-1985ce668451\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:45 crc kubenswrapper[4632]: E0313 10:07:45.985968 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:46.485927977 +0000 UTC m=+240.508458110 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.992833 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-mountpoint-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.992931 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-csi-data-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.993832 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/797176c6-dd56-48d6-8004-ff1dd5353a50-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-2n99d\" (UID: \"797176c6-dd56-48d6-8004-ff1dd5353a50\") " pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:07:45 crc kubenswrapper[4632]: I0313 10:07:45.994324 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58d59f3d-e656-4217-9472-62508a7ccc93-config-volume\") pod \"dns-default-g2wxc\" (UID: \"58d59f3d-e656-4217-9472-62508a7ccc93\") " pod="openshift-dns/dns-default-g2wxc" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.000581 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/946f5fcb-dde4-4784-965d-75a47187e703-signing-cabundle\") pod \"service-ca-9c57cc56f-68mjx\" (UID: \"946f5fcb-dde4-4784-965d-75a47187e703\") " pod="openshift-service-ca/service-ca-9c57cc56f-68mjx" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.001207 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-socket-dir\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.006442 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/f0f88609-cbfe-4ccc-b5db-e5c1be771855-certs\") pod \"machine-config-server-hlf9t\" (UID: \"f0f88609-cbfe-4ccc-b5db-e5c1be771855\") " pod="openshift-machine-config-operator/machine-config-server-hlf9t" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.010789 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/e0e1f142-2930-4f9b-b851-f7f7df22676b-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-4hmjh\" (UID: \"e0e1f142-2930-4f9b-b851-f7f7df22676b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4hmjh" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.011414 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ebf1040d-57dd-47ef-b839-6f78a7c5c75f-srv-cert\") pod \"olm-operator-6b444d44fb-tqbl9\" (UID: \"ebf1040d-57dd-47ef-b839-6f78a7c5c75f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.015928 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-9rrcn\" (UID: \"a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.017076 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/49c520f1-fb05-48ca-8435-1985ce668451-apiservice-cert\") pod \"packageserver-d55dfcdfc-zgxcd\" (UID: \"49c520f1-fb05-48ca-8435-1985ce668451\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.018658 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/4e100e6e-7259-4262-be47-9c2b5be7a53a-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-xfvsc\" (UID: \"4e100e6e-7259-4262-be47-9c2b5be7a53a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.030797 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52zmx\" (UniqueName: \"kubernetes.io/projected/fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73-kube-api-access-52zmx\") pod \"ingress-operator-5b745b69d9-jrkwc\" (UID: \"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.036251 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/779b2915-e0d0-4e90-9c6d-af28f555fd7b-serving-cert\") pod \"service-ca-operator-777779d784-zh465\" (UID: \"779b2915-e0d0-4e90-9c6d-af28f555fd7b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zh465" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.040419 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/946f5fcb-dde4-4784-965d-75a47187e703-signing-key\") pod \"service-ca-9c57cc56f-68mjx\" (UID: \"946f5fcb-dde4-4784-965d-75a47187e703\") " pod="openshift-service-ca/service-ca-9c57cc56f-68mjx" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.051858 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mprmj\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-kube-api-access-mprmj\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.054646 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/528d3aa9-10bf-4029-a4d2-85768264fde8-secret-volume\") pod \"collect-profiles-29556600-r9flg\" (UID: \"528d3aa9-10bf-4029-a4d2-85768264fde8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.055092 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/f0f88609-cbfe-4ccc-b5db-e5c1be771855-node-bootstrap-token\") pod \"machine-config-server-hlf9t\" (UID: \"f0f88609-cbfe-4ccc-b5db-e5c1be771855\") " pod="openshift-machine-config-operator/machine-config-server-hlf9t" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.055218 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ebf1040d-57dd-47ef-b839-6f78a7c5c75f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tqbl9\" (UID: \"ebf1040d-57dd-47ef-b839-6f78a7c5c75f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.055407 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2332524f-f990-4ef2-90b3-8b90c389d873-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-pvwll\" (UID: \"2332524f-f990-4ef2-90b3-8b90c389d873\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pvwll" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.055627 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/797176c6-dd56-48d6-8004-ff1dd5353a50-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-2n99d\" (UID: \"797176c6-dd56-48d6-8004-ff1dd5353a50\") " pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.064138 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/49c520f1-fb05-48ca-8435-1985ce668451-webhook-cert\") pod \"packageserver-d55dfcdfc-zgxcd\" (UID: \"49c520f1-fb05-48ca-8435-1985ce668451\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.077127 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/4db028f0-524e-46fc-aa33-da38ed7b8fa6-cert\") pod \"ingress-canary-qcb4l\" (UID: \"4db028f0-524e-46fc-aa33-da38ed7b8fa6\") " pod="openshift-ingress-canary/ingress-canary-qcb4l" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.078804 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/58d59f3d-e656-4217-9472-62508a7ccc93-metrics-tls\") pod \"dns-default-g2wxc\" (UID: \"58d59f3d-e656-4217-9472-62508a7ccc93\") " pod="openshift-dns/dns-default-g2wxc" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.079383 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:46 crc kubenswrapper[4632]: E0313 10:07:46.080793 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:46.580761291 +0000 UTC m=+240.603291424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.083059 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:46 crc kubenswrapper[4632]: E0313 10:07:46.083726 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:46.58370773 +0000 UTC m=+240.606237863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.088504 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-bound-sa-token\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.098141 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.100377 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8sl88"] Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.102588 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73-bound-sa-token\") pod \"ingress-operator-5b745b69d9-jrkwc\" (UID: \"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.112460 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxh5f\" (UniqueName: \"kubernetes.io/projected/f781cb50-1e1b-4586-ba59-b204b1a6beec-kube-api-access-hxh5f\") pod \"etcd-operator-b45778765-k955n\" (UID: \"f781cb50-1e1b-4586-ba59-b204b1a6beec\") " pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.122435 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ghtc\" (UniqueName: \"kubernetes.io/projected/b7b8ca1c-c3de-4829-ab9f-860f76033c63-kube-api-access-2ghtc\") pod \"openshift-controller-manager-operator-756b6f6bc6-hd8rx\" (UID: \"b7b8ca1c-c3de-4829-ab9f-860f76033c63\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.127243 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.144274 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-c6jnc"] Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.157767 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l6kft\" (UniqueName: \"kubernetes.io/projected/8be807d4-9bc2-41a1-b69f-1b0af031b5ab-kube-api-access-l6kft\") pod \"machine-config-controller-84d6567774-rntsr\" (UID: \"8be807d4-9bc2-41a1-b69f-1b0af031b5ab\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.170104 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rq5z\" (UniqueName: \"kubernetes.io/projected/96067558-b20b-411c-b1af-b8fbb61df8f7-kube-api-access-9rq5z\") pod \"migrator-59844c95c7-9wxcs\" (UID: \"96067558-b20b-411c-b1af-b8fbb61df8f7\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9wxcs" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.170940 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz"] Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.181395 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.183155 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7mb4q\" (UniqueName: \"kubernetes.io/projected/c94773d8-a922-4778-b2ba-8937e9d6c19b-kube-api-access-7mb4q\") pod \"dns-operator-744455d44c-6fqf5\" (UID: \"c94773d8-a922-4778-b2ba-8937e9d6c19b\") " pod="openshift-dns-operator/dns-operator-744455d44c-6fqf5" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.188190 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:46 crc kubenswrapper[4632]: E0313 10:07:46.188397 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:46.688362014 +0000 UTC m=+240.710892157 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.188775 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:46 crc kubenswrapper[4632]: E0313 10:07:46.189152 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:46.689144931 +0000 UTC m=+240.711675064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.189942 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.201392 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.207180 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txsm4\" (UniqueName: \"kubernetes.io/projected/e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e-kube-api-access-txsm4\") pod \"kube-storage-version-migrator-operator-b67b599dd-hb4c6\" (UID: \"e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.209177 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.216665 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.247915 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2493565c-3af9-4edf-a2f3-8a7a501e9305-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-l6ndt\" (UID: \"2493565c-3af9-4edf-a2f3-8a7a501e9305\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.254363 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7x7q\" (UniqueName: \"kubernetes.io/projected/353e9ca9-cb3b-4c6e-b1ca-446611a12dca-kube-api-access-w7x7q\") pod \"authentication-operator-69f744f599-svhr5\" (UID: \"353e9ca9-cb3b-4c6e-b1ca-446611a12dca\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.257577 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-w2hhj"] Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.268408 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.270487 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zq92\" (UniqueName: \"kubernetes.io/projected/a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59-kube-api-access-8zq92\") pod \"cluster-image-registry-operator-dc59b4c8b-9rrcn\" (UID: \"a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.285729 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22993daf-2b32-4be5-8eb7-f9194e903d62-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-946gp\" (UID: \"22993daf-2b32-4be5-8eb7-f9194e903d62\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.289670 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:46 crc kubenswrapper[4632]: E0313 10:07:46.293474 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:46.793436147 +0000 UTC m=+240.815966280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.299283 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.302829 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvs6x\" (UniqueName: \"kubernetes.io/projected/bbb27a61-7407-4cd7-84df-4b66fbdcf82d-kube-api-access-pvs6x\") pod \"machine-config-operator-74547568cd-99hff\" (UID: \"bbb27a61-7407-4cd7-84df-4b66fbdcf82d\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.321512 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9fc6g\" (UniqueName: \"kubernetes.io/projected/f660255f-8f78-4876-973d-db58f2ee7020-kube-api-access-9fc6g\") pod \"openshift-config-operator-7777fb866f-sk2l6\" (UID: \"f660255f-8f78-4876-973d-db58f2ee7020\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.341068 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-zn7mn"] Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.350427 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sd756\" (UniqueName: \"kubernetes.io/projected/c822257d-9d2f-4b6f-87de-131de5cd0efe-kube-api-access-sd756\") pod \"auto-csr-approver-29556606-mkrp2\" (UID: \"c822257d-9d2f-4b6f-87de-131de5cd0efe\") " pod="openshift-infra/auto-csr-approver-29556606-mkrp2" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.364499 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xx2pr\" (UniqueName: \"kubernetes.io/projected/f0f88609-cbfe-4ccc-b5db-e5c1be771855-kube-api-access-xx2pr\") pod \"machine-config-server-hlf9t\" (UID: \"f0f88609-cbfe-4ccc-b5db-e5c1be771855\") " pod="openshift-machine-config-operator/machine-config-server-hlf9t" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.382808 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm8lp\" (UniqueName: \"kubernetes.io/projected/528d3aa9-10bf-4029-a4d2-85768264fde8-kube-api-access-vm8lp\") pod \"collect-profiles-29556600-r9flg\" (UID: \"528d3aa9-10bf-4029-a4d2-85768264fde8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" Mar 13 10:07:46 crc kubenswrapper[4632]: E0313 10:07:46.393224 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:46.893203511 +0000 UTC m=+240.915733644 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.392759 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.401346 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-6fqf5" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.411608 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.413173 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hcvp\" (UniqueName: \"kubernetes.io/projected/49c520f1-fb05-48ca-8435-1985ce668451-kube-api-access-2hcvp\") pod \"packageserver-d55dfcdfc-zgxcd\" (UID: \"49c520f1-fb05-48ca-8435-1985ce668451\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.419823 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2vc6\" (UniqueName: \"kubernetes.io/projected/779b2915-e0d0-4e90-9c6d-af28f555fd7b-kube-api-access-q2vc6\") pod \"service-ca-operator-777779d784-zh465\" (UID: \"779b2915-e0d0-4e90-9c6d-af28f555fd7b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-zh465" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.436887 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.446118 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.453303 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.471613 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9wxcs" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.475060 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.475198 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsqdg\" (UniqueName: \"kubernetes.io/projected/946f5fcb-dde4-4784-965d-75a47187e703-kube-api-access-vsqdg\") pod \"service-ca-9c57cc56f-68mjx\" (UID: \"946f5fcb-dde4-4784-965d-75a47187e703\") " pod="openshift-service-ca/service-ca-9c57cc56f-68mjx" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.491760 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qtrb\" (UniqueName: \"kubernetes.io/projected/797176c6-dd56-48d6-8004-ff1dd5353a50-kube-api-access-8qtrb\") pod \"marketplace-operator-79b997595-2n99d\" (UID: \"797176c6-dd56-48d6-8004-ff1dd5353a50\") " pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.496375 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:46 crc kubenswrapper[4632]: E0313 10:07:46.496802 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:46.996785543 +0000 UTC m=+241.019315666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.501848 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wc9r4\" (UniqueName: \"kubernetes.io/projected/09ddc697-7ac1-4896-b9e2-1ae6c59c6f47-kube-api-access-wc9r4\") pod \"csi-hostpathplugin-hvrrc\" (UID: \"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47\") " pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.529296 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.536862 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78rbb\" (UniqueName: \"kubernetes.io/projected/58d59f3d-e656-4217-9472-62508a7ccc93-kube-api-access-78rbb\") pod \"dns-default-g2wxc\" (UID: \"58d59f3d-e656-4217-9472-62508a7ccc93\") " pod="openshift-dns/dns-default-g2wxc" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.547329 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5h9s\" (UniqueName: \"kubernetes.io/projected/4db028f0-524e-46fc-aa33-da38ed7b8fa6-kube-api-access-p5h9s\") pod \"ingress-canary-qcb4l\" (UID: \"4db028f0-524e-46fc-aa33-da38ed7b8fa6\") " pod="openshift-ingress-canary/ingress-canary-qcb4l" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.562600 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.563463 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.563473 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l2w2\" (UniqueName: \"kubernetes.io/projected/e0e1f142-2930-4f9b-b851-f7f7df22676b-kube-api-access-8l2w2\") pod \"multus-admission-controller-857f4d67dd-4hmjh\" (UID: \"e0e1f142-2930-4f9b-b851-f7f7df22676b\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-4hmjh" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.579286 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zh465" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.591686 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-4hmjh" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.593007 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pst8h\" (UniqueName: \"kubernetes.io/projected/ebf1040d-57dd-47ef-b839-6f78a7c5c75f-kube-api-access-pst8h\") pod \"olm-operator-6b444d44fb-tqbl9\" (UID: \"ebf1040d-57dd-47ef-b839-6f78a7c5c75f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.598701 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:46 crc kubenswrapper[4632]: E0313 10:07:46.599151 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:47.09913465 +0000 UTC m=+241.121664783 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.604738 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwlxd\" (UniqueName: \"kubernetes.io/projected/4e100e6e-7259-4262-be47-9c2b5be7a53a-kube-api-access-wwlxd\") pod \"package-server-manager-789f6589d5-xfvsc\" (UID: \"4e100e6e-7259-4262-be47-9c2b5be7a53a\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.605526 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556606-mkrp2" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.606234 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hw6wb\" (UniqueName: \"kubernetes.io/projected/2332524f-f990-4ef2-90b3-8b90c389d873-kube-api-access-hw6wb\") pod \"control-plane-machine-set-operator-78cbb6b69f-pvwll\" (UID: \"2332524f-f990-4ef2-90b3-8b90c389d873\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pvwll" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.617342 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-sbtn5"] Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.618074 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.626921 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-68mjx" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.653291 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.654432 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-qcb4l" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.662655 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-hlf9t" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.675901 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-g2wxc" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.701682 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:46 crc kubenswrapper[4632]: E0313 10:07:46.702539 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:47.202520358 +0000 UTC m=+241.225050491 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.805055 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:46 crc kubenswrapper[4632]: E0313 10:07:46.805349 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:47.305336975 +0000 UTC m=+241.327867108 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.838571 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pvwll" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.861679 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.873015 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.908315 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2" event={"ID":"8966c5f5-d0a8-4533-842c-0930c1a97bd7","Type":"ContainerStarted","Data":"2237ac232e1fe4d9854f091db4771705644f579421680aecb051bafd7b457de5"} Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.910285 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:46 crc kubenswrapper[4632]: E0313 10:07:46.910846 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:47.410823215 +0000 UTC m=+241.433353348 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.919573 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-w2hhj" event={"ID":"7d155f24-9bfc-4039-9981-10e7f724fa51","Type":"ContainerStarted","Data":"6560d4d8f94cb50cc670c974ce19515844f3d9021206a20e645ccc7bea0024b1"} Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.926768 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-k955n"] Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.967578 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" event={"ID":"54b67d35-da46-4e38-9b9a-e91855d6d88d","Type":"ContainerStarted","Data":"dd2270d6f010487fdb464c24099d1d08bc8eb87013a5ecf0899f2f88527bb38a"} Mar 13 10:07:46 crc kubenswrapper[4632]: I0313 10:07:46.972410 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr"] Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.009206 4632 generic.go:334] "Generic (PLEG): container finished" podID="d19fca6e-5095-42b6-8590-32c5b2c73308" containerID="d29dfaad69b4c668d9514564ac1fac14021c2da5ea130b61ca7e86e2c34d5223" exitCode=0 Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.009364 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" event={"ID":"d19fca6e-5095-42b6-8590-32c5b2c73308","Type":"ContainerDied","Data":"d29dfaad69b4c668d9514564ac1fac14021c2da5ea130b61ca7e86e2c34d5223"} Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.009406 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" event={"ID":"d19fca6e-5095-42b6-8590-32c5b2c73308","Type":"ContainerStarted","Data":"7436d5dff5db5b7d40d786566aa2dea44858e5fe31b4b8aece1de8a3e88e87cf"} Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.011915 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:47 crc kubenswrapper[4632]: E0313 10:07:47.012364 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:47.512346876 +0000 UTC m=+241.534877009 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.015814 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc"] Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.022156 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" event={"ID":"ef269b18-ea84-43c2-971c-e772149acbf6","Type":"ContainerStarted","Data":"75c15ceccf80c00f0980ce9fe8061ee72b4cae63ba83d1ef9a1e704159958941"} Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.031722 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn" event={"ID":"2c90710f-4595-425c-8be1-1436f43b5069","Type":"ContainerStarted","Data":"2d58cd0f978c6bbcd3613f18efda5d2b07f6320b4bb672f96c06d6f1b9392d0c"} Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.031790 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn" event={"ID":"2c90710f-4595-425c-8be1-1436f43b5069","Type":"ContainerStarted","Data":"0fe3af461e6c99b0dddea09fd2fc17bc9781ce95850249ce166d4812660ac046"} Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.048222 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" event={"ID":"560e6c43-4285-4ca8-98b9-874e9dcb5810","Type":"ContainerStarted","Data":"e9aab0e9cd1796940dcc2818af221f5b388f490c5b2161fb3217fdbc24d92e66"} Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.054435 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-t9vht" event={"ID":"7b959a85-56a5-4296-9cf3-87741e1f9c39","Type":"ContainerStarted","Data":"15bfcb8fb0de56bcff8ed22a6dadbf140d5efb1811983961ddae8f45269a3699"} Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.061928 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" event={"ID":"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a","Type":"ContainerStarted","Data":"46d5cd8b5a8d1e4d5e145a625b40cd39a2bdcba910908f1195bf38b9cf2ad7c8"} Mar 13 10:07:47 crc kubenswrapper[4632]: W0313 10:07:47.093416 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8be807d4_9bc2_41a1_b69f_1b0af031b5ab.slice/crio-f39d4790900980d6f8b0fb35dd4baa9babd1e887665cd118c696b9ba33ee881c WatchSource:0}: Error finding container f39d4790900980d6f8b0fb35dd4baa9babd1e887665cd118c696b9ba33ee881c: Status 404 returned error can't find the container with id f39d4790900980d6f8b0fb35dd4baa9babd1e887665cd118c696b9ba33ee881c Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.094206 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" event={"ID":"70f440bb-5dd8-4863-9749-bc5f7c547750","Type":"ContainerStarted","Data":"cacc884e7672aacd612df662055c2d9769da0a235fec8c1ddc593601e1331830"} Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.094242 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" event={"ID":"70f440bb-5dd8-4863-9749-bc5f7c547750","Type":"ContainerStarted","Data":"365524316a4e3e846e005a856282706fac826be9337ec760f74d5dd19061bccd"} Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.095403 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.113914 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.119252 4632 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-9v5nn container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.119325 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" podUID="70f440bb-5dd8-4863-9749-bc5f7c547750" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 13 10:07:47 crc kubenswrapper[4632]: E0313 10:07:47.121586 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:47.614127671 +0000 UTC m=+241.636657804 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.121937 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:47 crc kubenswrapper[4632]: E0313 10:07:47.123880 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:47.623870119 +0000 UTC m=+241.646400242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.220433 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" event={"ID":"275c3112-6912-49f8-9d3f-8147662fb99f","Type":"ContainerStarted","Data":"b2f35c19a6f9bc4062fbc156a0e6a89b8f0bc286049907966124c4fa1d962cf0"} Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.222990 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:47 crc kubenswrapper[4632]: E0313 10:07:47.224380 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:47.724360178 +0000 UTC m=+241.746890321 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.234835 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zn7mn" event={"ID":"f5a50074-5531-442f-a0e9-0578f15634c1","Type":"ContainerStarted","Data":"c0f56571b6b9472de716bb190b1d68fe783e6f7b131b06ae9b0c01071f1d985f"} Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.253424 4632 generic.go:334] "Generic (PLEG): container finished" podID="37df1143-69fc-4d13-a5d3-790a9d14814a" containerID="1727fc9a0b7884510a5f00372a1dc955706d0b945e2fa4f057778a1cc32974e7" exitCode=0 Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.253514 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" event={"ID":"37df1143-69fc-4d13-a5d3-790a9d14814a","Type":"ContainerDied","Data":"1727fc9a0b7884510a5f00372a1dc955706d0b945e2fa4f057778a1cc32974e7"} Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.253556 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" event={"ID":"37df1143-69fc-4d13-a5d3-790a9d14814a","Type":"ContainerStarted","Data":"d1c99f569127b25b9fbcaede6867472b51e2252615cbf84c8b6a6df75564818b"} Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.344994 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:47 crc kubenswrapper[4632]: E0313 10:07:47.370238 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:47.870177367 +0000 UTC m=+241.892707500 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.437103 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p"] Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.471725 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:47 crc kubenswrapper[4632]: E0313 10:07:47.476842 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:47.976804481 +0000 UTC m=+241.999334614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.484719 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-svhr5"] Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.487268 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz"] Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.588508 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:47 crc kubenswrapper[4632]: E0313 10:07:47.588789 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:48.088776773 +0000 UTC m=+242.111306906 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.653044 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx"] Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.689876 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:47 crc kubenswrapper[4632]: E0313 10:07:47.691079 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:48.191057949 +0000 UTC m=+242.213588082 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.706644 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn"] Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.725544 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6"] Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.792971 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:47 crc kubenswrapper[4632]: E0313 10:07:47.794012 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:48.293986368 +0000 UTC m=+242.316516491 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.883121 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" podStartSLOduration=187.883103777 podStartE2EDuration="3m7.883103777s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:47.874052443 +0000 UTC m=+241.896582596" watchObservedRunningTime="2026-03-13 10:07:47.883103777 +0000 UTC m=+241.905633910" Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.885109 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-zh465"] Mar 13 10:07:47 crc kubenswrapper[4632]: W0313 10:07:47.891182 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb7b8ca1c_c3de_4829_ab9f_860f76033c63.slice/crio-9bd6a1e25880e0d50cb7206a62d253a15785eb5c750fcc9d272be545f04fec11 WatchSource:0}: Error finding container 9bd6a1e25880e0d50cb7206a62d253a15785eb5c750fcc9d272be545f04fec11: Status 404 returned error can't find the container with id 9bd6a1e25880e0d50cb7206a62d253a15785eb5c750fcc9d272be545f04fec11 Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.894185 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:47 crc kubenswrapper[4632]: E0313 10:07:47.894591 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:48.394542739 +0000 UTC m=+242.417072872 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:47 crc kubenswrapper[4632]: I0313 10:07:47.997838 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:48 crc kubenswrapper[4632]: E0313 10:07:48.028157 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:48.528072108 +0000 UTC m=+242.550602241 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.153153 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:48 crc kubenswrapper[4632]: E0313 10:07:48.153833 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:48.65380517 +0000 UTC m=+242.676335303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.256204 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:48 crc kubenswrapper[4632]: E0313 10:07:48.256623 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:48.756610906 +0000 UTC m=+242.779141039 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.310128 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt"] Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.353025 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" event={"ID":"560e6c43-4285-4ca8-98b9-874e9dcb5810","Type":"ContainerStarted","Data":"6158a81f875e1232b8d27bb41ad2531364da77cdf3704ac46d4ec2470ad3e550"} Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.353373 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.358974 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:48 crc kubenswrapper[4632]: E0313 10:07:48.359822 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:48.85980099 +0000 UTC m=+242.882331123 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.362147 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-6fqf5"] Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.370839 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-99hff"] Mar 13 10:07:48 crc kubenswrapper[4632]: W0313 10:07:48.375425 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod779b2915_e0d0_4e90_9c6d_af28f555fd7b.slice/crio-0b59f9c490eaefcaa625809b8cbeebb469b353d8229d8a4081413e9f101a689c WatchSource:0}: Error finding container 0b59f9c490eaefcaa625809b8cbeebb469b353d8229d8a4081413e9f101a689c: Status 404 returned error can't find the container with id 0b59f9c490eaefcaa625809b8cbeebb469b353d8229d8a4081413e9f101a689c Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.419822 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6"] Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.483000 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-9sqbn" podStartSLOduration=189.482970019 podStartE2EDuration="3m9.482970019s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:48.410891006 +0000 UTC m=+242.433421149" watchObservedRunningTime="2026-03-13 10:07:48.482970019 +0000 UTC m=+242.505500152" Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.494319 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:48 crc kubenswrapper[4632]: E0313 10:07:48.496380 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:48.996362182 +0000 UTC m=+243.018892315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.498608 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hlf9t" event={"ID":"f0f88609-cbfe-4ccc-b5db-e5c1be771855","Type":"ContainerStarted","Data":"ff16b2a3d9f3d7d99f8d4ce7d4d78b9e1cab2df5df81d37700d025fd29eb7322"} Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.510115 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" event={"ID":"f660255f-8f78-4876-973d-db58f2ee7020","Type":"ContainerStarted","Data":"aef5416b84402461bc252efa526c879aee119a858392680529c54a17da1ad089"} Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.540706 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" event={"ID":"32f62e32-732b-4646-85f0-45b8ea6544a6","Type":"ContainerStarted","Data":"90b7ec6506badd9443bd2435c386d50093b4439a3938b26453bc3448d5b17f87"} Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.588353 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" event={"ID":"275c3112-6912-49f8-9d3f-8147662fb99f","Type":"ContainerStarted","Data":"659c231578759d8866d439292e45b9c6aaafb089bba68dedf66a629dc8c40639"} Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.599598 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:48 crc kubenswrapper[4632]: E0313 10:07:48.600310 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:49.10028473 +0000 UTC m=+243.122814863 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.601858 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.602146 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-hvrrc"] Mar 13 10:07:48 crc kubenswrapper[4632]: E0313 10:07:48.602322 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:49.102306201 +0000 UTC m=+243.124836334 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.605135 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" event={"ID":"54b67d35-da46-4e38-9b9a-e91855d6d88d","Type":"ContainerStarted","Data":"8feffc3b7bf857c2072a56dc9b0fd9b862356263b01ceb8dccd59275afac9e52"} Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.628693 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz" event={"ID":"9cd4c3b3-6825-4bd2-97a5-330f91782d4b","Type":"ContainerStarted","Data":"9273fa4a8c3b1d68730f94261ebd33aa40699d16a884dcdb48f02b215427c5bb"} Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.630690 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" event={"ID":"8be807d4-9bc2-41a1-b69f-1b0af031b5ab","Type":"ContainerStarted","Data":"f39d4790900980d6f8b0fb35dd4baa9babd1e887665cd118c696b9ba33ee881c"} Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.631854 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" event={"ID":"a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59","Type":"ContainerStarted","Data":"2b9c22a14a4a7ae7fff9ce82d82b8b32f6520e8ebcbfa421349d5b033517659a"} Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.636629 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp"] Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.640268 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd"] Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.641230 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx" event={"ID":"b7b8ca1c-c3de-4829-ab9f-860f76033c63","Type":"ContainerStarted","Data":"9bd6a1e25880e0d50cb7206a62d253a15785eb5c750fcc9d272be545f04fec11"} Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.664483 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2n99d"] Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.675507 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" event={"ID":"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a","Type":"ContainerStarted","Data":"7595f698758f3f9ece4af82a45628aeb01bfa58cde4c80fefdeaf746f39aba12"} Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.677526 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.707877 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:48 crc kubenswrapper[4632]: E0313 10:07:48.708019 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:49.207994506 +0000 UTC m=+243.230524639 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.708224 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:48 crc kubenswrapper[4632]: E0313 10:07:48.709036 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:49.209027856 +0000 UTC m=+243.231557989 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.726272 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" podStartSLOduration=187.726237306 podStartE2EDuration="3m7.726237306s" podCreationTimestamp="2026-03-13 10:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:48.696515362 +0000 UTC m=+242.719045495" watchObservedRunningTime="2026-03-13 10:07:48.726237306 +0000 UTC m=+242.748767439" Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.756427 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-w2hhj" Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.802271 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" event={"ID":"f781cb50-1e1b-4586-ba59-b204b1a6beec","Type":"ContainerStarted","Data":"c52bdf516fa5eedcad8ea9ff2d5e00053c74f310f559396f45b05df9adae9a66"} Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.803842 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.803898 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.810525 4632 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8sl88 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.17:6443/healthz\": dial tcp 10.217.0.17:6443: connect: connection refused" start-of-body= Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.810594 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" podUID="f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.17:6443/healthz\": dial tcp 10.217.0.17:6443: connect: connection refused" Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.811985 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:48 crc kubenswrapper[4632]: E0313 10:07:48.812420 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:49.312399474 +0000 UTC m=+243.334929607 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.914341 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:48 crc kubenswrapper[4632]: E0313 10:07:48.919974 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:49.419958877 +0000 UTC m=+243.442489010 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:48 crc kubenswrapper[4632]: I0313 10:07:48.985575 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2" event={"ID":"8966c5f5-d0a8-4533-842c-0930c1a97bd7","Type":"ContainerStarted","Data":"1fb481567c320a802926357902cdf9e454a08bef3b34ac3ab075ac8849449faf"} Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.016269 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:49 crc kubenswrapper[4632]: E0313 10:07:49.077454 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:49.577426033 +0000 UTC m=+243.599956166 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.100468 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" event={"ID":"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73","Type":"ContainerStarted","Data":"5bb775d7a4d37ab105f66a4275bfb44cc48f9a56145a7c70d951017418982da0"} Mar 13 10:07:49 crc kubenswrapper[4632]: E0313 10:07:49.279048 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8be807d4_9bc2_41a1_b69f_1b0af031b5ab.slice/crio-4472e43626812cd6438ed2e942691abb4a297046dc1f20f89eb299f8aec4a1d2.scope\": RecentStats: unable to find data in memory cache]" Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.279575 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zn7mn" event={"ID":"f5a50074-5531-442f-a0e9-0578f15634c1","Type":"ContainerStarted","Data":"662793b7c27b62a99fd064350b3cd52eb21f393bbf5603bbcbf03a65855922bf"} Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.343797 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-hmljp" podStartSLOduration=190.343775878 podStartE2EDuration="3m10.343775878s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:49.327782333 +0000 UTC m=+243.350312466" watchObservedRunningTime="2026-03-13 10:07:49.343775878 +0000 UTC m=+243.366306011" Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.344336 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:49 crc kubenswrapper[4632]: E0313 10:07:49.344864 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:49.844799078 +0000 UTC m=+243.867329211 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.380636 4632 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-xthqz container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.380745 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" podUID="560e6c43-4285-4ca8-98b9-874e9dcb5810" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.394934 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" event={"ID":"353e9ca9-cb3b-4c6e-b1ca-446611a12dca","Type":"ContainerStarted","Data":"0238ab2e7a36fdf21574adc11d217b059ba17531c3b892fddcd341539fdf7844"} Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.444650 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-w2hhj" podStartSLOduration=189.444613785 podStartE2EDuration="3m9.444613785s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:49.411853919 +0000 UTC m=+243.434384052" watchObservedRunningTime="2026-03-13 10:07:49.444613785 +0000 UTC m=+243.467143928" Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.461176 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.463031 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-4hmjh"] Mar 13 10:07:49 crc kubenswrapper[4632]: E0313 10:07:49.468901 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:49.968831946 +0000 UTC m=+243.991362079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.485925 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.510628 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" podStartSLOduration=190.510601853 podStartE2EDuration="3m10.510601853s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:49.507773586 +0000 UTC m=+243.530303719" watchObservedRunningTime="2026-03-13 10:07:49.510601853 +0000 UTC m=+243.533131986" Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.554357 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-qcb4l"] Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.572640 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:49 crc kubenswrapper[4632]: E0313 10:07:49.573073 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:50.073054391 +0000 UTC m=+244.095584524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.618309 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pvwll"] Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.629225 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-zn7mn" podStartSLOduration=190.629208651 podStartE2EDuration="3m10.629208651s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:49.627575717 +0000 UTC m=+243.650105850" watchObservedRunningTime="2026-03-13 10:07:49.629208651 +0000 UTC m=+243.651738784" Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.675739 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:49 crc kubenswrapper[4632]: E0313 10:07:49.678724 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:50.178680524 +0000 UTC m=+244.201210657 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.682349 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:49 crc kubenswrapper[4632]: E0313 10:07:49.682971 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:50.182936291 +0000 UTC m=+244.205466424 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.714135 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-9wxcs"] Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.726149 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556606-mkrp2"] Mar 13 10:07:49 crc kubenswrapper[4632]: W0313 10:07:49.749151 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2332524f_f990_4ef2_90b3_8b90c389d873.slice/crio-e3d7f16736e0371190aa67ce95de1cd66666710393070685ef1ec194c35672ea WatchSource:0}: Error finding container e3d7f16736e0371190aa67ce95de1cd66666710393070685ef1ec194c35672ea: Status 404 returned error can't find the container with id e3d7f16736e0371190aa67ce95de1cd66666710393070685ef1ec194c35672ea Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.765816 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9"] Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.789812 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:49 crc kubenswrapper[4632]: E0313 10:07:49.790216 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:50.290197638 +0000 UTC m=+244.312727771 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.800459 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-68mjx"] Mar 13 10:07:49 crc kubenswrapper[4632]: W0313 10:07:49.844441 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0e1f142_2930_4f9b_b851_f7f7df22676b.slice/crio-94e0a638ee2ae51c16d5af96281e3d95c3d9ca2db3d7fd1fb46c71f1257c5158 WatchSource:0}: Error finding container 94e0a638ee2ae51c16d5af96281e3d95c3d9ca2db3d7fd1fb46c71f1257c5158: Status 404 returned error can't find the container with id 94e0a638ee2ae51c16d5af96281e3d95c3d9ca2db3d7fd1fb46c71f1257c5158 Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.885982 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.898753 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:49 crc kubenswrapper[4632]: E0313 10:07:49.899206 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:50.399191849 +0000 UTC m=+244.421721982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.903330 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-g2wxc"] Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.920808 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc"] Mar 13 10:07:49 crc kubenswrapper[4632]: I0313 10:07:49.984566 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg"] Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.001867 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:50 crc kubenswrapper[4632]: E0313 10:07:50.002223 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:50.50220402 +0000 UTC m=+244.524734153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:50 crc kubenswrapper[4632]: W0313 10:07:50.084622 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58d59f3d_e656_4217_9472_62508a7ccc93.slice/crio-9696f157835f576c90de9fc6fb04fe18f862a43e1e14381bfaf3ea5fa2f8c5df WatchSource:0}: Error finding container 9696f157835f576c90de9fc6fb04fe18f862a43e1e14381bfaf3ea5fa2f8c5df: Status 404 returned error can't find the container with id 9696f157835f576c90de9fc6fb04fe18f862a43e1e14381bfaf3ea5fa2f8c5df Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.106919 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:50 crc kubenswrapper[4632]: E0313 10:07:50.107509 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:50.607493576 +0000 UTC m=+244.630023709 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.225670 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:50 crc kubenswrapper[4632]: E0313 10:07:50.226079 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:50.726059353 +0000 UTC m=+244.748589486 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.276801 4632 ???:1] "http: TLS handshake error from 192.168.126.11:51414: no serving certificate available for the kubelet" Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.334674 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:50 crc kubenswrapper[4632]: E0313 10:07:50.335105 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:50.835092985 +0000 UTC m=+244.857623118 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.417901 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-68mjx" event={"ID":"946f5fcb-dde4-4784-965d-75a47187e703","Type":"ContainerStarted","Data":"2636a650497e2001cc9b3101d94e482e82f27bd0bc1ba4c4ff52faa01ca79e70"} Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.438490 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:50 crc kubenswrapper[4632]: E0313 10:07:50.439009 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:50.938992084 +0000 UTC m=+244.961522217 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.490141 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" event={"ID":"49c520f1-fb05-48ca-8435-1985ce668451","Type":"ContainerStarted","Data":"c54852cb281200c3dfab6510e8cb54c5924c24038a30081db00b84832819abcb"} Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.511719 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zh465" event={"ID":"779b2915-e0d0-4e90-9c6d-af28f555fd7b","Type":"ContainerStarted","Data":"0b59f9c490eaefcaa625809b8cbeebb469b353d8229d8a4081413e9f101a689c"} Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.544146 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:50 crc kubenswrapper[4632]: E0313 10:07:50.544527 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:51.044515164 +0000 UTC m=+245.067045297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.596698 4632 ???:1] "http: TLS handshake error from 192.168.126.11:51430: no serving certificate available for the kubelet" Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.602308 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" event={"ID":"bbb27a61-7407-4cd7-84df-4b66fbdcf82d","Type":"ContainerStarted","Data":"978d841db0ae703cb57743ebee1052df0e79457f163667ee122d8877204644c5"} Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.649725 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.650327 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-hlf9t" event={"ID":"f0f88609-cbfe-4ccc-b5db-e5c1be771855","Type":"ContainerStarted","Data":"37785bfa0feaa29bfa5e7bf2222e8c269922d184a39c7f35513f8811c878debd"} Mar 13 10:07:50 crc kubenswrapper[4632]: E0313 10:07:50.650648 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:51.150628338 +0000 UTC m=+245.173158471 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.670620 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" event={"ID":"528d3aa9-10bf-4029-a4d2-85768264fde8","Type":"ContainerStarted","Data":"1137745f79e5dd4b86f11690ff5ed0914b045872452dc8054e30e019f43d068c"} Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.717821 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pvwll" event={"ID":"2332524f-f990-4ef2-90b3-8b90c389d873","Type":"ContainerStarted","Data":"e3d7f16736e0371190aa67ce95de1cd66666710393070685ef1ec194c35672ea"} Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.718751 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9wxcs" event={"ID":"96067558-b20b-411c-b1af-b8fbb61df8f7","Type":"ContainerStarted","Data":"cf854e7992ea940d3b22d1d6f00a4593b8bc5d524fdb7929fd54bc70cd27491d"} Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.719507 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" event={"ID":"ebf1040d-57dd-47ef-b839-6f78a7c5c75f","Type":"ContainerStarted","Data":"a8ecc63c9ed7db7b994c34ee5fa665d5251ee69358655fb999be7016a6cf0616"} Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.734185 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" event={"ID":"353e9ca9-cb3b-4c6e-b1ca-446611a12dca","Type":"ContainerStarted","Data":"696b18c58833c0581e6bf36ae1881e00a6717c6dc6b1a5150c21fe634a2b6edb"} Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.737795 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6fqf5" event={"ID":"c94773d8-a922-4778-b2ba-8937e9d6c19b","Type":"ContainerStarted","Data":"fff7b0aa35e9d16e30d04426b48150081a0880479b641def5d6a9bcfb7f47cd6"} Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.751202 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:50 crc kubenswrapper[4632]: E0313 10:07:50.755911 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:51.255893654 +0000 UTC m=+245.278423787 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.793891 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" event={"ID":"4e100e6e-7259-4262-be47-9c2b5be7a53a","Type":"ContainerStarted","Data":"b9e3d723e372f0843b9d5d3305252118cb4fcdcab3fc206a25062c241fe435f9"} Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.797232 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-hlf9t" podStartSLOduration=7.797214583 podStartE2EDuration="7.797214583s" podCreationTimestamp="2026-03-13 10:07:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:50.717286551 +0000 UTC m=+244.739816684" watchObservedRunningTime="2026-03-13 10:07:50.797214583 +0000 UTC m=+244.819744716" Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.842149 4632 ???:1] "http: TLS handshake error from 192.168.126.11:51446: no serving certificate available for the kubelet" Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.853774 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:50 crc kubenswrapper[4632]: E0313 10:07:50.854294 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:51.354270191 +0000 UTC m=+245.376800324 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.866654 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-g2wxc" event={"ID":"58d59f3d-e656-4217-9472-62508a7ccc93","Type":"ContainerStarted","Data":"9696f157835f576c90de9fc6fb04fe18f862a43e1e14381bfaf3ea5fa2f8c5df"} Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.918617 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp" event={"ID":"22993daf-2b32-4be5-8eb7-f9194e903d62","Type":"ContainerStarted","Data":"a1638203d2af94586ec485f42c4fbf775571059e1c2de85a3a1d38eac9847322"} Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.955608 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" podStartSLOduration=191.955581967 podStartE2EDuration="3m11.955581967s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:50.799925328 +0000 UTC m=+244.822455461" watchObservedRunningTime="2026-03-13 10:07:50.955581967 +0000 UTC m=+244.978112110" Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.956812 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.958134 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" podStartSLOduration=190.958117188 podStartE2EDuration="3m10.958117188s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:50.955435924 +0000 UTC m=+244.977966057" watchObservedRunningTime="2026-03-13 10:07:50.958117188 +0000 UTC m=+244.980647321" Mar 13 10:07:50 crc kubenswrapper[4632]: E0313 10:07:50.959372 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:51.459351494 +0000 UTC m=+245.481881627 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:50 crc kubenswrapper[4632]: I0313 10:07:50.960319 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-t9vht" event={"ID":"7b959a85-56a5-4296-9cf3-87741e1f9c39","Type":"ContainerStarted","Data":"c0db1ffabe3d33862c8266179a821f8fd8c1a4906081849cc73b575a98544e3b"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.030035 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2" event={"ID":"8966c5f5-d0a8-4533-842c-0930c1a97bd7","Type":"ContainerStarted","Data":"ec9be579ef24f67bc9a4d0bd07390f0db6e2e0a609ceb2fe6ac4a52cee11b067"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.044285 4632 ???:1] "http: TLS handshake error from 192.168.126.11:51450: no serving certificate available for the kubelet" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.092985 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:51 crc kubenswrapper[4632]: E0313 10:07:51.093650 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:51.593634279 +0000 UTC m=+245.616164412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.098432 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" event={"ID":"797176c6-dd56-48d6-8004-ff1dd5353a50","Type":"ContainerStarted","Data":"f11fbb0ec92177c2b8cb772cacb63ff7d8a26b02bee6907aaa00dedbedf68d98"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.131461 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6" event={"ID":"e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e","Type":"ContainerStarted","Data":"eac967ea2e870d8d27ce821cf58f72bcea7cf75f33059be8e3c88ba059d1a1ac"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.150510 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-t9vht" podStartSLOduration=191.150481592 podStartE2EDuration="3m11.150481592s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:51.094995846 +0000 UTC m=+245.117525989" watchObservedRunningTime="2026-03-13 10:07:51.150481592 +0000 UTC m=+245.173011725" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.152861 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-qtrc2" podStartSLOduration=192.15285124 podStartE2EDuration="3m12.15285124s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:51.149614745 +0000 UTC m=+245.172144908" watchObservedRunningTime="2026-03-13 10:07:51.15285124 +0000 UTC m=+245.175381373" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.200176 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.200431 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:51 crc kubenswrapper[4632]: E0313 10:07:51.200852 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:51.700835994 +0000 UTC m=+245.723366127 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.207381 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:07:51 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:07:51 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:07:51 crc kubenswrapper[4632]: healthz check failed Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.207443 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.213869 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" event={"ID":"275c3112-6912-49f8-9d3f-8147662fb99f","Type":"ContainerStarted","Data":"16e1c9b7987925e139c95cc985e936b717e71be620d5b4e52242b0526d6a2335"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.237607 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556606-mkrp2" event={"ID":"c822257d-9d2f-4b6f-87de-131de5cd0efe","Type":"ContainerStarted","Data":"4b486b426e38ba0d310d07052394a9d5bdba25cfa8d2705294f114f94eaedc81"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.258654 4632 ???:1] "http: TLS handshake error from 192.168.126.11:51462: no serving certificate available for the kubelet" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.272804 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-c6jnc" podStartSLOduration=190.272775033 podStartE2EDuration="3m10.272775033s" podCreationTimestamp="2026-03-13 10:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:51.270443566 +0000 UTC m=+245.292973699" watchObservedRunningTime="2026-03-13 10:07:51.272775033 +0000 UTC m=+245.295305166" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.278920 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-w2hhj" event={"ID":"7d155f24-9bfc-4039-9981-10e7f724fa51","Type":"ContainerStarted","Data":"dc3a73e428d40f73e3034fb8b7d18fcfe7453c6673209f7b70847a9a508f90d4"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.280702 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.280756 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.283555 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" event={"ID":"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47","Type":"ContainerStarted","Data":"6bb360edaeb98a0ce0225fba59f7c71dd52dfa0c38be17890b61097dd0c283b3"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.323553 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:51 crc kubenswrapper[4632]: E0313 10:07:51.323874 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:51.82385677 +0000 UTC m=+245.846386913 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.365460 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" event={"ID":"ef269b18-ea84-43c2-971c-e772149acbf6","Type":"ContainerStarted","Data":"faaf308e22c1a8d08431430b330cacf53efc9923cc70f0515be295533e608c79"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.367080 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.395273 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" event={"ID":"32f62e32-732b-4646-85f0-45b8ea6544a6","Type":"ContainerStarted","Data":"4f41aedb607002fa771d4b82bf1fb15a527c048ee3048ce7cd9db7dc1d8b7961"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.396658 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.408855 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qcb4l" event={"ID":"4db028f0-524e-46fc-aa33-da38ed7b8fa6","Type":"ContainerStarted","Data":"05d9328f5993ff28efe917f43052a2d7d2b56f187f221997b229d80138a9668c"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.419956 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4hmjh" event={"ID":"e0e1f142-2930-4f9b-b851-f7f7df22676b","Type":"ContainerStarted","Data":"94e0a638ee2ae51c16d5af96281e3d95c3d9ca2db3d7fd1fb46c71f1257c5158"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.421109 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt" event={"ID":"2493565c-3af9-4edf-a2f3-8a7a501e9305","Type":"ContainerStarted","Data":"1fc8ce799e02b250ba115dbeb135562734f9d1f478f83e9ebd8c897a9aa0f527"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.422577 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" event={"ID":"8be807d4-9bc2-41a1-b69f-1b0af031b5ab","Type":"ContainerStarted","Data":"4472e43626812cd6438ed2e942691abb4a297046dc1f20f89eb299f8aec4a1d2"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.425109 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:51 crc kubenswrapper[4632]: E0313 10:07:51.428201 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:51.928173937 +0000 UTC m=+245.950704300 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.431423 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" event={"ID":"d19fca6e-5095-42b6-8590-32c5b2c73308","Type":"ContainerStarted","Data":"e308322021135c5ed9ce8a32583947cf382a4bc2a981fcbfb12a53b54fda4790"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.483251 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx" event={"ID":"b7b8ca1c-c3de-4829-ab9f-860f76033c63","Type":"ContainerStarted","Data":"afd7707f1200eec84045aec7e26bcc717b636a687a21f4ec5c7a205ad80ec7f3"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.485563 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podStartSLOduration=192.485542961 podStartE2EDuration="3m12.485542961s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:51.484160413 +0000 UTC m=+245.506690546" watchObservedRunningTime="2026-03-13 10:07:51.485542961 +0000 UTC m=+245.508073084" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.496255 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" event={"ID":"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73","Type":"ContainerStarted","Data":"b19f68b6880ea930ca87409ab0f966556f18774e7a975f90f350a496f8371831"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.508342 4632 ???:1] "http: TLS handshake error from 192.168.126.11:51466: no serving certificate available for the kubelet" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.523990 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" event={"ID":"37df1143-69fc-4d13-a5d3-790a9d14814a","Type":"ContainerStarted","Data":"671ef1ed2b00036ec2e404981c153eb4a8e77da75376b49caef1e2b96bc79aec"} Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.529195 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.529266 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.531239 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:51 crc kubenswrapper[4632]: E0313 10:07:51.533297 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:52.03327393 +0000 UTC m=+246.055804223 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.569662 4632 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-r5v5p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.569786 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" podUID="32f62e32-732b-4646-85f0-45b8ea6544a6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.570038 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.661547 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:51 crc kubenswrapper[4632]: E0313 10:07:51.662126 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:52.162107905 +0000 UTC m=+246.184638028 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.729218 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" podStartSLOduration=190.729198655 podStartE2EDuration="3m10.729198655s" podCreationTimestamp="2026-03-13 10:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:51.728824578 +0000 UTC m=+245.751354731" watchObservedRunningTime="2026-03-13 10:07:51.729198655 +0000 UTC m=+245.751728788" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.730405 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-hd8rx" podStartSLOduration=191.73039721 podStartE2EDuration="3m11.73039721s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:51.601732409 +0000 UTC m=+245.624262542" watchObservedRunningTime="2026-03-13 10:07:51.73039721 +0000 UTC m=+245.752927353" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.772522 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:51 crc kubenswrapper[4632]: E0313 10:07:51.773112 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:52.273067066 +0000 UTC m=+246.295597209 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.773269 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.773312 4632 ???:1] "http: TLS handshake error from 192.168.126.11:51482: no serving certificate available for the kubelet" Mar 13 10:07:51 crc kubenswrapper[4632]: E0313 10:07:51.773840 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:52.273828481 +0000 UTC m=+246.296358614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.844109 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" podStartSLOduration=190.844093518 podStartE2EDuration="3m10.844093518s" podCreationTimestamp="2026-03-13 10:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:51.842918364 +0000 UTC m=+245.865448497" watchObservedRunningTime="2026-03-13 10:07:51.844093518 +0000 UTC m=+245.866623651" Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.875387 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:51 crc kubenswrapper[4632]: E0313 10:07:51.875750 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:52.375731599 +0000 UTC m=+246.398261742 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.983294 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:51 crc kubenswrapper[4632]: E0313 10:07:51.983901 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:52.483886844 +0000 UTC m=+246.506416977 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:51 crc kubenswrapper[4632]: I0313 10:07:51.987397 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.088765 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:52 crc kubenswrapper[4632]: E0313 10:07:52.089075 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:52.589020737 +0000 UTC m=+246.611550870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.089494 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:52 crc kubenswrapper[4632]: E0313 10:07:52.089903 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:52.589895365 +0000 UTC m=+246.612425498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.210634 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:52 crc kubenswrapper[4632]: E0313 10:07:52.210757 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:52.710742098 +0000 UTC m=+246.733272231 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.210992 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:52 crc kubenswrapper[4632]: E0313 10:07:52.211266 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:52.711256419 +0000 UTC m=+246.733786552 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.261050 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:07:52 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:07:52 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:07:52 crc kubenswrapper[4632]: healthz check failed Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.261132 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.311522 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:52 crc kubenswrapper[4632]: E0313 10:07:52.312198 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:52.812182357 +0000 UTC m=+246.834712490 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.362651 4632 ???:1] "http: TLS handshake error from 192.168.126.11:51492: no serving certificate available for the kubelet" Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.414469 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:52 crc kubenswrapper[4632]: E0313 10:07:52.414913 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:52.914897091 +0000 UTC m=+246.937427214 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.515553 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:52 crc kubenswrapper[4632]: E0313 10:07:52.516017 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:53.016000182 +0000 UTC m=+247.038530315 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.560678 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" event={"ID":"a4c5a906-1d0b-40e7-aa4f-bc945e9f1f59","Type":"ContainerStarted","Data":"7a5bcf9f212f5a2b310a7efd17456057f682cd2aa737d0333fa5a860860efb95"} Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.581923 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" event={"ID":"528d3aa9-10bf-4029-a4d2-85768264fde8","Type":"ContainerStarted","Data":"da165dd4ae62fa2ea1c777c8125fcd4bfe4bd102f508da056f1a058689bba35e"} Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.597620 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp" event={"ID":"22993daf-2b32-4be5-8eb7-f9194e903d62","Type":"ContainerStarted","Data":"448dab19b144eab73c92a38dec4aa7a678df60a74c15d3992f8cfe45580486ed"} Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.623058 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz" event={"ID":"9cd4c3b3-6825-4bd2-97a5-330f91782d4b","Type":"ContainerStarted","Data":"ba6ebfd612cba7e001fb3e96528df7785ea1937329f26e6ca8b2ffa9099a0267"} Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.625763 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:52 crc kubenswrapper[4632]: E0313 10:07:52.628199 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:53.12818158 +0000 UTC m=+247.150711713 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.638286 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pvwll" event={"ID":"2332524f-f990-4ef2-90b3-8b90c389d873","Type":"ContainerStarted","Data":"d31f3a76795fd3df921747cc5a2960017c10149aa7e4a878019746d93c54cc18"} Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.653901 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-9rrcn" podStartSLOduration=192.653869351 podStartE2EDuration="3m12.653869351s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:52.651456891 +0000 UTC m=+246.673987044" watchObservedRunningTime="2026-03-13 10:07:52.653869351 +0000 UTC m=+246.676399504" Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.663477 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt" event={"ID":"2493565c-3af9-4edf-a2f3-8a7a501e9305","Type":"ContainerStarted","Data":"1afce30866e9f140a28fff0a431440447ec493a74037d1bf14031a4607047deb"} Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.707712 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" event={"ID":"8be807d4-9bc2-41a1-b69f-1b0af031b5ab","Type":"ContainerStarted","Data":"99faffe7be689eda64884a72770d491d25afa5bde0cae33d38c8950ba26f6a7f"} Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.737691 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:52 crc kubenswrapper[4632]: E0313 10:07:52.739403 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:53.239386886 +0000 UTC m=+247.261917019 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.765769 4632 generic.go:334] "Generic (PLEG): container finished" podID="f660255f-8f78-4876-973d-db58f2ee7020" containerID="8784ad9d9edf3b0167053e3050a765f4ed5e301efeda89b7117d2a334a743a5e" exitCode=0 Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.765881 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" event={"ID":"f660255f-8f78-4876-973d-db58f2ee7020","Type":"ContainerDied","Data":"8784ad9d9edf3b0167053e3050a765f4ed5e301efeda89b7117d2a334a743a5e"} Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.787673 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" podStartSLOduration=193.787638075 podStartE2EDuration="3m13.787638075s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:52.750020782 +0000 UTC m=+246.772550915" watchObservedRunningTime="2026-03-13 10:07:52.787638075 +0000 UTC m=+246.810168208" Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.817099 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" event={"ID":"49c520f1-fb05-48ca-8435-1985ce668451","Type":"ContainerStarted","Data":"35b32e739ccce4a6f84a62ef541fb840a3cf0ce2a60fb788f618073e6f79bd60"} Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.817159 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.833894 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" start-of-body= Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.833977 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.843740 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:52 crc kubenswrapper[4632]: E0313 10:07:52.849802 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:53.349780766 +0000 UTC m=+247.372310899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.865897 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zh465" event={"ID":"779b2915-e0d0-4e90-9c6d-af28f555fd7b","Type":"ContainerStarted","Data":"b1cb38909f02af1a3486fbdd45dec5e71ff2321e667af9738833b913d588a2ea"} Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.919726 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rvkzz" podStartSLOduration=192.919692755 podStartE2EDuration="3m12.919692755s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:52.848060192 +0000 UTC m=+246.870590335" watchObservedRunningTime="2026-03-13 10:07:52.919692755 +0000 UTC m=+246.942222888" Mar 13 10:07:52 crc kubenswrapper[4632]: I0313 10:07:52.950818 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:52 crc kubenswrapper[4632]: E0313 10:07:52.952793 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:53.452772826 +0000 UTC m=+247.475302959 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.056263 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-k955n" event={"ID":"f781cb50-1e1b-4586-ba59-b204b1a6beec","Type":"ContainerStarted","Data":"883d07c6c6d265d059ef0c146e608f387ce1adfcf7af4157abcb8ce50bb4dff3"} Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.057551 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:53 crc kubenswrapper[4632]: E0313 10:07:53.057899 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:53.557884489 +0000 UTC m=+247.580414612 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.082356 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" event={"ID":"bbb27a61-7407-4cd7-84df-4b66fbdcf82d","Type":"ContainerStarted","Data":"30e6bdf5c51b5f0b537bb2a7b181b4e9345f94e30109bbcf3545aa94fd680a70"} Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.098002 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-946gp" podStartSLOduration=193.097985283 podStartE2EDuration="3m13.097985283s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:53.095570564 +0000 UTC m=+247.118100697" watchObservedRunningTime="2026-03-13 10:07:53.097985283 +0000 UTC m=+247.120515416" Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.117262 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9wxcs" event={"ID":"96067558-b20b-411c-b1af-b8fbb61df8f7","Type":"ContainerStarted","Data":"cee136f6a40b1d254350a9976840583417034cd82c227204e53e60b7c5c6eca8"} Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.144116 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" event={"ID":"797176c6-dd56-48d6-8004-ff1dd5353a50","Type":"ContainerStarted","Data":"0fd5f07ae3c28f8c24cc66a585de93acc08f170fb621bbeb190cd66596980871"} Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.145332 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.159531 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:53 crc kubenswrapper[4632]: E0313 10:07:53.160340 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:53.660323948 +0000 UTC m=+247.682854081 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.160840 4632 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2n99d container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.160892 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" podUID="797176c6-dd56-48d6-8004-ff1dd5353a50" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.161753 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6" event={"ID":"e884f4d1-d4f3-4ef7-b1f2-c39ea2eee50e","Type":"ContainerStarted","Data":"e1f80f4408c7077e4880cbeb6be74d125578b1b416485e0937117937025c127c"} Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.194197 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" event={"ID":"fdaf1cb9-0ab4-477f-bbd5-d8d33ab56f73","Type":"ContainerStarted","Data":"42d99267513fb3271ca67710e3271f3c0526e8ee8db64d915dd6b44a1763398c"} Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.218038 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:07:53 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:07:53 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:07:53 crc kubenswrapper[4632]: healthz check failed Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.218093 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.221725 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.221767 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.261695 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:53 crc kubenswrapper[4632]: E0313 10:07:53.269255 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:53.769239368 +0000 UTC m=+247.791769501 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.304141 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.397799 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:53 crc kubenswrapper[4632]: E0313 10:07:53.402222 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:53.902185476 +0000 UTC m=+247.924715619 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.402785 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:53 crc kubenswrapper[4632]: E0313 10:07:53.405025 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:53.905012893 +0000 UTC m=+247.927543026 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.491551 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-pvwll" podStartSLOduration=192.491534529 podStartE2EDuration="3m12.491534529s" podCreationTimestamp="2026-03-13 10:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:53.211457976 +0000 UTC m=+247.233988109" watchObservedRunningTime="2026-03-13 10:07:53.491534529 +0000 UTC m=+247.514064662" Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.537217 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:53 crc kubenswrapper[4632]: E0313 10:07:53.537561 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:54.037542113 +0000 UTC m=+248.060072246 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.640480 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:53 crc kubenswrapper[4632]: E0313 10:07:53.641093 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:54.141080594 +0000 UTC m=+248.163610727 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.641662 4632 ???:1] "http: TLS handshake error from 192.168.126.11:51506: no serving certificate available for the kubelet" Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.719007 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" podStartSLOduration=192.718988555 podStartE2EDuration="3m12.718988555s" podCreationTimestamp="2026-03-13 10:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:53.707095164 +0000 UTC m=+247.729625297" watchObservedRunningTime="2026-03-13 10:07:53.718988555 +0000 UTC m=+247.741518678" Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.719245 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-jrkwc" podStartSLOduration=193.71923961 podStartE2EDuration="3m13.71923961s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:53.543893862 +0000 UTC m=+247.566424005" watchObservedRunningTime="2026-03-13 10:07:53.71923961 +0000 UTC m=+247.741769743" Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.742163 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:53 crc kubenswrapper[4632]: E0313 10:07:53.742666 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:54.242634464 +0000 UTC m=+248.265164737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.742873 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:53 crc kubenswrapper[4632]: E0313 10:07:53.743482 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:54.243470172 +0000 UTC m=+248.266000305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.838458 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-hb4c6" podStartSLOduration=193.838433919 podStartE2EDuration="3m13.838433919s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:53.830984538 +0000 UTC m=+247.853514671" watchObservedRunningTime="2026-03-13 10:07:53.838433919 +0000 UTC m=+247.860964052" Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.844308 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:53 crc kubenswrapper[4632]: E0313 10:07:53.844641 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:54.344620784 +0000 UTC m=+248.367150917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.885331 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-zh465" podStartSLOduration=192.88531112 podStartE2EDuration="3m12.88531112s" podCreationTimestamp="2026-03-13 10:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:53.877221206 +0000 UTC m=+247.899751339" watchObservedRunningTime="2026-03-13 10:07:53.88531112 +0000 UTC m=+247.907841253" Mar 13 10:07:53 crc kubenswrapper[4632]: I0313 10:07:53.950084 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:53 crc kubenswrapper[4632]: E0313 10:07:53.950599 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:54.450557474 +0000 UTC m=+248.473087607 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.053420 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:54 crc kubenswrapper[4632]: E0313 10:07:54.053589 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:54.553560955 +0000 UTC m=+248.576091088 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.053691 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:54 crc kubenswrapper[4632]: E0313 10:07:54.072745 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:54.572718453 +0000 UTC m=+248.595248586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.091815 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9wxcs" podStartSLOduration=194.09178793 podStartE2EDuration="3m14.09178793s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:53.971234544 +0000 UTC m=+247.993764677" watchObservedRunningTime="2026-03-13 10:07:54.09178793 +0000 UTC m=+248.114318063" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.091953 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podStartSLOduration=193.091932613 podStartE2EDuration="3m13.091932613s" podCreationTimestamp="2026-03-13 10:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:54.088602696 +0000 UTC m=+248.111132839" watchObservedRunningTime="2026-03-13 10:07:54.091932613 +0000 UTC m=+248.114462746" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.158463 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:54 crc kubenswrapper[4632]: E0313 10:07:54.158892 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:54.658871452 +0000 UTC m=+248.681401585 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.225120 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.225201 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.241538 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:07:54 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:07:54 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:07:54 crc kubenswrapper[4632]: healthz check failed Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.241644 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.243834 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-l6ndt" podStartSLOduration=194.243809575 podStartE2EDuration="3m14.243809575s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:54.241895506 +0000 UTC m=+248.264425639" watchObservedRunningTime="2026-03-13 10:07:54.243809575 +0000 UTC m=+248.266339708" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.271304 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:54 crc kubenswrapper[4632]: E0313 10:07:54.271770 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:54.771754022 +0000 UTC m=+248.794284155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.306879 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" event={"ID":"f660255f-8f78-4876-973d-db58f2ee7020","Type":"ContainerStarted","Data":"e98f0e8253db82d7fc1c628a628a0d9ea91c85c3796f3abe0d968983b3e782e2"} Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.307116 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.351856 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" event={"ID":"37df1143-69fc-4d13-a5d3-790a9d14814a","Type":"ContainerStarted","Data":"112e377136996e25010b35483b2e1f7104c5fa4408a57b127c0f29e9b4b1396a"} Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.365851 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-g2wxc" event={"ID":"58d59f3d-e656-4217-9472-62508a7ccc93","Type":"ContainerStarted","Data":"1fd32c57b88f0b8b4ead90868742916aabf06aaf3a5152f76aac591bcdbbd091"} Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.366519 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-g2wxc" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.373018 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:54 crc kubenswrapper[4632]: E0313 10:07:54.374091 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:54.874073178 +0000 UTC m=+248.896603311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.389444 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-qcb4l" event={"ID":"4db028f0-524e-46fc-aa33-da38ed7b8fa6","Type":"ContainerStarted","Data":"b3bb037ac8508fee6cfb9112e0d5912db93ffd0e474f14540ca64e50c35d91a0"} Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.407160 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" event={"ID":"bbb27a61-7407-4cd7-84df-4b66fbdcf82d","Type":"ContainerStarted","Data":"bf8af74d9b0f042318756359c4286a7a2e4768a4e8d05e51ae504c7fda6422fe"} Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.419265 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6fqf5" event={"ID":"c94773d8-a922-4778-b2ba-8937e9d6c19b","Type":"ContainerStarted","Data":"3fc5945ba70c7d6bff986b63e755d289379e65187e067a619b17e24a4f9716f1"} Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.436016 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-9wxcs" event={"ID":"96067558-b20b-411c-b1af-b8fbb61df8f7","Type":"ContainerStarted","Data":"cd7a89e2368215d8104acc2802924935cdf3abebfe9ce85efe2cac7fa0fb0fa7"} Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.456450 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4hmjh" event={"ID":"e0e1f142-2930-4f9b-b851-f7f7df22676b","Type":"ContainerStarted","Data":"ecbcedb42472f525c93e094d11c0b45f162c4052cd44889cc6b3056b5b130d62"} Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.472862 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-rntsr" podStartSLOduration=194.472828393 podStartE2EDuration="3m14.472828393s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:54.340728812 +0000 UTC m=+248.363258945" watchObservedRunningTime="2026-03-13 10:07:54.472828393 +0000 UTC m=+248.495358526" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.475229 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:54 crc kubenswrapper[4632]: E0313 10:07:54.477094 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:54.977076059 +0000 UTC m=+248.999606392 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.481147 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" event={"ID":"ebf1040d-57dd-47ef-b839-6f78a7c5c75f","Type":"ContainerStarted","Data":"956edf235abcb661b9044b0baa5051ccc57d0211ecc18378cd6b766dc4edf9bd"} Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.482098 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.485310 4632 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-tqbl9 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" start-of-body= Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.485378 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" podUID="ebf1040d-57dd-47ef-b839-6f78a7c5c75f" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": dial tcp 10.217.0.37:8443: connect: connection refused" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.497785 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" event={"ID":"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47","Type":"ContainerStarted","Data":"da106490637a2286eb470c5eb4cdab3ab51dc9c3d415ff2d01e1800f284ee2d9"} Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.508076 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" event={"ID":"4e100e6e-7259-4262-be47-9c2b5be7a53a","Type":"ContainerStarted","Data":"8686dc189770f7b6373b29069a26380efb52cc7fda6bcf92c2bb9d4fee9440fc"} Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.508135 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" event={"ID":"4e100e6e-7259-4262-be47-9c2b5be7a53a","Type":"ContainerStarted","Data":"5558b8a7760fdc6f72ff7d0488173d63a3f337d120614ea0027a0eb87a92cc2d"} Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.508853 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.510326 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-68mjx" event={"ID":"946f5fcb-dde4-4784-965d-75a47187e703","Type":"ContainerStarted","Data":"7b1701334cdec6b285605c47da84fb005fe850e1c337e7faaa2c84d60edc5fe2"} Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.511574 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" start-of-body= Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.511617 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.512419 4632 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2n99d container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.512452 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" podUID="797176c6-dd56-48d6-8004-ff1dd5353a50" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.577499 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:54 crc kubenswrapper[4632]: E0313 10:07:54.578041 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:55.078021317 +0000 UTC m=+249.100551460 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.627828 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-99hff" podStartSLOduration=194.627807568 podStartE2EDuration="3m14.627807568s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:54.541405464 +0000 UTC m=+248.563935607" watchObservedRunningTime="2026-03-13 10:07:54.627807568 +0000 UTC m=+248.650337701" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.682521 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:54 crc kubenswrapper[4632]: E0313 10:07:54.687073 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:55.1870568 +0000 UTC m=+249.209586933 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.699167 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" podStartSLOduration=193.699148356 podStartE2EDuration="3m13.699148356s" podCreationTimestamp="2026-03-13 10:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:54.681782593 +0000 UTC m=+248.704312726" watchObservedRunningTime="2026-03-13 10:07:54.699148356 +0000 UTC m=+248.721678489" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.700135 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" podStartSLOduration=193.700127135 podStartE2EDuration="3m13.700127135s" podCreationTimestamp="2026-03-13 10:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:54.629585794 +0000 UTC m=+248.652115937" watchObservedRunningTime="2026-03-13 10:07:54.700127135 +0000 UTC m=+248.722657268" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.727366 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-4hmjh" podStartSLOduration=194.727341898 podStartE2EDuration="3m14.727341898s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:54.727099723 +0000 UTC m=+248.749629856" watchObservedRunningTime="2026-03-13 10:07:54.727341898 +0000 UTC m=+248.749872031" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.784113 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:54 crc kubenswrapper[4632]: E0313 10:07:54.784502 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:55.284486778 +0000 UTC m=+249.307016911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.887656 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:54 crc kubenswrapper[4632]: E0313 10:07:54.888072 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:55.38805299 +0000 UTC m=+249.410583123 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.981278 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-qcb4l" podStartSLOduration=11.98125528 podStartE2EDuration="11.98125528s" podCreationTimestamp="2026-03-13 10:07:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:54.905612445 +0000 UTC m=+248.928142568" watchObservedRunningTime="2026-03-13 10:07:54.98125528 +0000 UTC m=+249.003785413" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.983962 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jvh86"] Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.984929 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.988222 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.988537 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:54 crc kubenswrapper[4632]: E0313 10:07:54.988701 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:55.488679791 +0000 UTC m=+249.511209924 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.988752 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.988929 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5kzjw\" (UniqueName: \"kubernetes.io/projected/bd46ae04-0610-4aa5-9385-dd45de66c5dd-kube-api-access-5kzjw\") pod \"community-operators-jvh86\" (UID: \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\") " pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.989054 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd46ae04-0610-4aa5-9385-dd45de66c5dd-utilities\") pod \"community-operators-jvh86\" (UID: \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\") " pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:07:54 crc kubenswrapper[4632]: E0313 10:07:54.989090 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:55.489077519 +0000 UTC m=+249.511607652 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:54 crc kubenswrapper[4632]: I0313 10:07:54.989149 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd46ae04-0610-4aa5-9385-dd45de66c5dd-catalog-content\") pod \"community-operators-jvh86\" (UID: \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\") " pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.090106 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.090278 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5kzjw\" (UniqueName: \"kubernetes.io/projected/bd46ae04-0610-4aa5-9385-dd45de66c5dd-kube-api-access-5kzjw\") pod \"community-operators-jvh86\" (UID: \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\") " pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.090337 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd46ae04-0610-4aa5-9385-dd45de66c5dd-utilities\") pod \"community-operators-jvh86\" (UID: \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\") " pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.090355 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd46ae04-0610-4aa5-9385-dd45de66c5dd-catalog-content\") pod \"community-operators-jvh86\" (UID: \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\") " pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.090826 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd46ae04-0610-4aa5-9385-dd45de66c5dd-catalog-content\") pod \"community-operators-jvh86\" (UID: \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\") " pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:07:55 crc kubenswrapper[4632]: E0313 10:07:55.090901 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:55.590886115 +0000 UTC m=+249.613416248 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.091377 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd46ae04-0610-4aa5-9385-dd45de66c5dd-utilities\") pod \"community-operators-jvh86\" (UID: \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\") " pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.115864 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" podStartSLOduration=196.115845321 podStartE2EDuration="3m16.115845321s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:55.076411261 +0000 UTC m=+249.098941394" watchObservedRunningTime="2026-03-13 10:07:55.115845321 +0000 UTC m=+249.138375454" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.127459 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jvh86"] Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.194533 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:55 crc kubenswrapper[4632]: E0313 10:07:55.194916 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:55.694901936 +0000 UTC m=+249.717432069 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.202474 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:07:55 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:07:55 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:07:55 crc kubenswrapper[4632]: healthz check failed Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.202546 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.228370 4632 ???:1] "http: TLS handshake error from 192.168.126.11:53496: no serving certificate available for the kubelet" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.237661 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5kzjw\" (UniqueName: \"kubernetes.io/projected/bd46ae04-0610-4aa5-9385-dd45de66c5dd-kube-api-access-5kzjw\") pod \"community-operators-jvh86\" (UID: \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\") " pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.296324 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:55 crc kubenswrapper[4632]: E0313 10:07:55.297212 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:55.797177592 +0000 UTC m=+249.819707725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.316058 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podStartSLOduration=196.316033054 podStartE2EDuration="3m16.316033054s" podCreationTimestamp="2026-03-13 10:04:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:55.307760176 +0000 UTC m=+249.330290309" watchObservedRunningTime="2026-03-13 10:07:55.316033054 +0000 UTC m=+249.338563187" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.318864 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xd455"] Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.320158 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xd455" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.333173 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.401928 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.402028 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kh6g9\" (UniqueName: \"kubernetes.io/projected/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-kube-api-access-kh6g9\") pod \"community-operators-xd455\" (UID: \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\") " pod="openshift-marketplace/community-operators-xd455" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.402104 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-utilities\") pod \"community-operators-xd455\" (UID: \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\") " pod="openshift-marketplace/community-operators-xd455" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.402210 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-catalog-content\") pod \"community-operators-xd455\" (UID: \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\") " pod="openshift-marketplace/community-operators-xd455" Mar 13 10:07:55 crc kubenswrapper[4632]: E0313 10:07:55.402737 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:55.902717453 +0000 UTC m=+249.925247586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.424383 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xd455"] Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.482640 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.484122 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.508908 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.509462 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-catalog-content\") pod \"community-operators-xd455\" (UID: \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\") " pod="openshift-marketplace/community-operators-xd455" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.509567 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kh6g9\" (UniqueName: \"kubernetes.io/projected/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-kube-api-access-kh6g9\") pod \"community-operators-xd455\" (UID: \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\") " pod="openshift-marketplace/community-operators-xd455" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.509637 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-utilities\") pod \"community-operators-xd455\" (UID: \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\") " pod="openshift-marketplace/community-operators-xd455" Mar 13 10:07:55 crc kubenswrapper[4632]: E0313 10:07:55.510243 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:56.010189534 +0000 UTC m=+250.032719667 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.511122 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-catalog-content\") pod \"community-operators-xd455\" (UID: \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\") " pod="openshift-marketplace/community-operators-xd455" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.511476 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-utilities\") pod \"community-operators-xd455\" (UID: \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\") " pod="openshift-marketplace/community-operators-xd455" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.511495 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-g2wxc" podStartSLOduration=12.511485401 podStartE2EDuration="12.511485401s" podCreationTimestamp="2026-03-13 10:07:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:55.443763576 +0000 UTC m=+249.466293709" watchObservedRunningTime="2026-03-13 10:07:55.511485401 +0000 UTC m=+249.534015534" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.511617 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.511667 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.512387 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p8wjg"] Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.513150 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-68mjx" podStartSLOduration=194.513141794 podStartE2EDuration="3m14.513141794s" podCreationTimestamp="2026-03-13 10:04:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:55.507895527 +0000 UTC m=+249.530425670" watchObservedRunningTime="2026-03-13 10:07:55.513141794 +0000 UTC m=+249.535671927" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.513598 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.517756 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.518319 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.530898 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:55 crc kubenswrapper[4632]: W0313 10:07:55.541552 4632 reflector.go:561] object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g": failed to list *v1.Secret: secrets "certified-operators-dockercfg-4rs5g" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Mar 13 10:07:55 crc kubenswrapper[4632]: E0313 10:07:55.541634 4632 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"certified-operators-dockercfg-4rs5g\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"certified-operators-dockercfg-4rs5g\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.570855 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kh6g9\" (UniqueName: \"kubernetes.io/projected/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-kube-api-access-kh6g9\") pod \"community-operators-xd455\" (UID: \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\") " pod="openshift-marketplace/community-operators-xd455" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.578889 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p8wjg"] Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.599237 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-6fqf5" event={"ID":"c94773d8-a922-4778-b2ba-8937e9d6c19b","Type":"ContainerStarted","Data":"7322cc652a565a3dbc9f2961b94879fa1c1cefc9d2f67a83d4d25ee10d30123a"} Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.611015 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.611076 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlsv8\" (UniqueName: \"kubernetes.io/projected/b11a7dff-bf08-44c3-b4f4-923119c13717-kube-api-access-wlsv8\") pod \"certified-operators-p8wjg\" (UID: \"b11a7dff-bf08-44c3-b4f4-923119c13717\") " pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.611162 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11a7dff-bf08-44c3-b4f4-923119c13717-utilities\") pod \"certified-operators-p8wjg\" (UID: \"b11a7dff-bf08-44c3-b4f4-923119c13717\") " pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.611187 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11a7dff-bf08-44c3-b4f4-923119c13717-catalog-content\") pod \"certified-operators-p8wjg\" (UID: \"b11a7dff-bf08-44c3-b4f4-923119c13717\") " pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:07:55 crc kubenswrapper[4632]: E0313 10:07:55.611586 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:56.111572251 +0000 UTC m=+250.134102384 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.613299 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-4hmjh" event={"ID":"e0e1f142-2930-4f9b-b851-f7f7df22676b","Type":"ContainerStarted","Data":"bc5e045ba766bfb2723324f723497a7fd48f730201e476208c152ec22d0fe530"} Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.632149 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.632200 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.632226 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.632275 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.646823 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-g2wxc" event={"ID":"58d59f3d-e656-4217-9472-62508a7ccc93","Type":"ContainerStarted","Data":"d2664a1ff1c3241483169e35dd71718d3944b4647d3547d3172180d62f38666f"} Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.652335 4632 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2n99d container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.652365 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" podUID="797176c6-dd56-48d6-8004-ff1dd5353a50" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.677208 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-7vrbc" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.691867 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8z668"] Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.699266 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.704843 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.706746 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xd455" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.715994 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.716437 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9845f384-2720-4d6a-aa73-1e66e30f7c2c-catalog-content\") pod \"certified-operators-8z668\" (UID: \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\") " pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:07:55 crc kubenswrapper[4632]: E0313 10:07:55.716784 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:56.216751145 +0000 UTC m=+250.239281468 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.717027 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.717074 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlsv8\" (UniqueName: \"kubernetes.io/projected/b11a7dff-bf08-44c3-b4f4-923119c13717-kube-api-access-wlsv8\") pod \"certified-operators-p8wjg\" (UID: \"b11a7dff-bf08-44c3-b4f4-923119c13717\") " pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.717162 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdp56\" (UniqueName: \"kubernetes.io/projected/9845f384-2720-4d6a-aa73-1e66e30f7c2c-kube-api-access-sdp56\") pod \"certified-operators-8z668\" (UID: \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\") " pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.717391 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9845f384-2720-4d6a-aa73-1e66e30f7c2c-utilities\") pod \"certified-operators-8z668\" (UID: \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\") " pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.717620 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11a7dff-bf08-44c3-b4f4-923119c13717-utilities\") pod \"certified-operators-p8wjg\" (UID: \"b11a7dff-bf08-44c3-b4f4-923119c13717\") " pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.717672 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11a7dff-bf08-44c3-b4f4-923119c13717-catalog-content\") pod \"certified-operators-p8wjg\" (UID: \"b11a7dff-bf08-44c3-b4f4-923119c13717\") " pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:07:55 crc kubenswrapper[4632]: E0313 10:07:55.736826 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:56.236797453 +0000 UTC m=+250.259327576 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.738625 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11a7dff-bf08-44c3-b4f4-923119c13717-utilities\") pod \"certified-operators-p8wjg\" (UID: \"b11a7dff-bf08-44c3-b4f4-923119c13717\") " pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.738833 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11a7dff-bf08-44c3-b4f4-923119c13717-catalog-content\") pod \"certified-operators-p8wjg\" (UID: \"b11a7dff-bf08-44c3-b4f4-923119c13717\") " pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.798224 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8z668"] Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.819162 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.819431 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9845f384-2720-4d6a-aa73-1e66e30f7c2c-utilities\") pod \"certified-operators-8z668\" (UID: \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\") " pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.819520 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9845f384-2720-4d6a-aa73-1e66e30f7c2c-catalog-content\") pod \"certified-operators-8z668\" (UID: \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\") " pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.819636 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdp56\" (UniqueName: \"kubernetes.io/projected/9845f384-2720-4d6a-aa73-1e66e30f7c2c-kube-api-access-sdp56\") pod \"certified-operators-8z668\" (UID: \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\") " pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:07:55 crc kubenswrapper[4632]: E0313 10:07:55.820103 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:56.320084823 +0000 UTC m=+250.342614956 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.820558 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9845f384-2720-4d6a-aa73-1e66e30f7c2c-utilities\") pod \"certified-operators-8z668\" (UID: \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\") " pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.820831 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9845f384-2720-4d6a-aa73-1e66e30f7c2c-catalog-content\") pod \"certified-operators-8z668\" (UID: \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\") " pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.840912 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlsv8\" (UniqueName: \"kubernetes.io/projected/b11a7dff-bf08-44c3-b4f4-923119c13717-kube-api-access-wlsv8\") pod \"certified-operators-p8wjg\" (UID: \"b11a7dff-bf08-44c3-b4f4-923119c13717\") " pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.888865 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdp56\" (UniqueName: \"kubernetes.io/projected/9845f384-2720-4d6a-aa73-1e66e30f7c2c-kube-api-access-sdp56\") pod \"certified-operators-8z668\" (UID: \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\") " pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.924400 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:55 crc kubenswrapper[4632]: E0313 10:07:55.924837 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:56.424819598 +0000 UTC m=+250.447349721 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.984187 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.985668 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.994542 4632 patch_prober.go:28] interesting pod/console-f9d7485db-zn7mn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Mar 13 10:07:55 crc kubenswrapper[4632]: I0313 10:07:55.994625 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zn7mn" podUID="f5a50074-5531-442f-a0e9-0578f15634c1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.026759 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:56 crc kubenswrapper[4632]: E0313 10:07:56.028898 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:56.528869359 +0000 UTC m=+250.551399492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.129653 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:56 crc kubenswrapper[4632]: E0313 10:07:56.130114 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:56.630102254 +0000 UTC m=+250.652632387 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.191380 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.197172 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:07:56 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:07:56 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:07:56 crc kubenswrapper[4632]: healthz check failed Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.197282 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.205105 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-6fqf5" podStartSLOduration=196.205069095 podStartE2EDuration="3m16.205069095s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:07:56.18360821 +0000 UTC m=+250.206138353" watchObservedRunningTime="2026-03-13 10:07:56.205069095 +0000 UTC m=+250.227599228" Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.235928 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:56 crc kubenswrapper[4632]: E0313 10:07:56.249247 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:56.749211091 +0000 UTC m=+250.771741224 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.396064 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:56 crc kubenswrapper[4632]: E0313 10:07:56.396810 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:56.896796686 +0000 UTC m=+250.919326819 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.503450 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:56 crc kubenswrapper[4632]: E0313 10:07:56.504075 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:57.004052533 +0000 UTC m=+251.026582666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.580464 4632 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2n99d container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.580501 4632 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-2n99d container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" start-of-body= Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.580551 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" podUID="797176c6-dd56-48d6-8004-ff1dd5353a50" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.580581 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" podUID="797176c6-dd56-48d6-8004-ff1dd5353a50" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.32:8080/healthz\": dial tcp 10.217.0.32:8080: connect: connection refused" Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.607227 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:56 crc kubenswrapper[4632]: E0313 10:07:56.607707 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:57.107689856 +0000 UTC m=+251.130219989 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.691158 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" event={"ID":"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47","Type":"ContainerStarted","Data":"3bf30655a86ac973647b131da3b6f942b8afa26f8c24f35c2fd8e55e5b065db4"} Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.709151 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:56 crc kubenswrapper[4632]: E0313 10:07:56.710005 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:57.209928491 +0000 UTC m=+251.232458614 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.785822 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.814292 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:56 crc kubenswrapper[4632]: E0313 10:07:56.820410 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:57.320385633 +0000 UTC m=+251.342915966 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.865089 4632 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-marketplace/certified-operators-p8wjg" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.865241 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:07:56 crc kubenswrapper[4632]: I0313 10:07:56.917613 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:56 crc kubenswrapper[4632]: E0313 10:07:56.918791 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:57.418769239 +0000 UTC m=+251.441299372 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.023552 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:57 crc kubenswrapper[4632]: E0313 10:07:57.024042 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:57.524030005 +0000 UTC m=+251.546560138 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.093667 4632 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-marketplace/certified-operators-8z668" secret="" err="failed to sync secret cache: timed out waiting for the condition" Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.093758 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.127029 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.132559 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:57 crc kubenswrapper[4632]: E0313 10:07:57.133968 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:57.633920625 +0000 UTC m=+251.656450758 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.192489 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jvh86"] Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.202687 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:07:57 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:07:57 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:07:57 crc kubenswrapper[4632]: healthz check failed Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.202777 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.235397 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:57 crc kubenswrapper[4632]: E0313 10:07:57.235887 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:57.735868683 +0000 UTC m=+251.758398816 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.266468 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xd455"] Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.340174 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:57 crc kubenswrapper[4632]: E0313 10:07:57.340663 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:57.84063984 +0000 UTC m=+251.863169963 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.466027 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:57 crc kubenswrapper[4632]: E0313 10:07:57.467180 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:57.967155268 +0000 UTC m=+251.989685401 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.538069 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.569706 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:57 crc kubenswrapper[4632]: E0313 10:07:57.570169 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:58.070153837 +0000 UTC m=+252.092683960 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.672088 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:57 crc kubenswrapper[4632]: E0313 10:07:57.672674 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:58.172661177 +0000 UTC m=+252.195191310 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.699422 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xd455" event={"ID":"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8","Type":"ContainerStarted","Data":"33259f14f07cee3a1d7261a44a8f74cbd0957ccef81016b9631c3a0a7ccd4085"} Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.701247 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvh86" event={"ID":"bd46ae04-0610-4aa5-9385-dd45de66c5dd","Type":"ContainerStarted","Data":"c1e7366f3326cfd08308453ff8a94a3f8d3ce8ebc6a33b2bfafadd960643927e"} Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.774204 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:57 crc kubenswrapper[4632]: E0313 10:07:57.774387 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:58.274360761 +0000 UTC m=+252.296890894 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.774642 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:57 crc kubenswrapper[4632]: E0313 10:07:57.775627 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:58.275616657 +0000 UTC m=+252.298147000 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.813119 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-txp2w"] Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.814593 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.862716 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.877594 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:57 crc kubenswrapper[4632]: E0313 10:07:57.878394 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:58.378377642 +0000 UTC m=+252.400907775 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.955382 4632 patch_prober.go:28] interesting pod/apiserver-76f77b778f-p9gp2 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 13 10:07:57 crc kubenswrapper[4632]: [+]log ok Mar 13 10:07:57 crc kubenswrapper[4632]: [+]etcd ok Mar 13 10:07:57 crc kubenswrapper[4632]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 13 10:07:57 crc kubenswrapper[4632]: [+]poststarthook/generic-apiserver-start-informers ok Mar 13 10:07:57 crc kubenswrapper[4632]: [+]poststarthook/max-in-flight-filter ok Mar 13 10:07:57 crc kubenswrapper[4632]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 13 10:07:57 crc kubenswrapper[4632]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 13 10:07:57 crc kubenswrapper[4632]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 13 10:07:57 crc kubenswrapper[4632]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Mar 13 10:07:57 crc kubenswrapper[4632]: [+]poststarthook/project.openshift.io-projectcache ok Mar 13 10:07:57 crc kubenswrapper[4632]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 13 10:07:57 crc kubenswrapper[4632]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Mar 13 10:07:57 crc kubenswrapper[4632]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 13 10:07:57 crc kubenswrapper[4632]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 13 10:07:57 crc kubenswrapper[4632]: livez check failed Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.955468 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" podUID="37df1143-69fc-4d13-a5d3-790a9d14814a" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.985040 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5jx8\" (UniqueName: \"kubernetes.io/projected/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-kube-api-access-z5jx8\") pod \"redhat-marketplace-txp2w\" (UID: \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\") " pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.985101 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.985182 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-utilities\") pod \"redhat-marketplace-txp2w\" (UID: \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\") " pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:07:57 crc kubenswrapper[4632]: I0313 10:07:57.985205 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-catalog-content\") pod \"redhat-marketplace-txp2w\" (UID: \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\") " pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:07:57 crc kubenswrapper[4632]: E0313 10:07:57.985824 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:58.485808193 +0000 UTC m=+252.508338326 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.028839 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-txp2w"] Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.141436 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.141703 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-utilities\") pod \"redhat-marketplace-txp2w\" (UID: \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\") " pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.141730 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-catalog-content\") pod \"redhat-marketplace-txp2w\" (UID: \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\") " pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.141797 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5jx8\" (UniqueName: \"kubernetes.io/projected/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-kube-api-access-z5jx8\") pod \"redhat-marketplace-txp2w\" (UID: \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\") " pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:07:58 crc kubenswrapper[4632]: E0313 10:07:58.142279 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:58.642265628 +0000 UTC m=+252.664795761 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.143127 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-catalog-content\") pod \"redhat-marketplace-txp2w\" (UID: \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\") " pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.143442 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-utilities\") pod \"redhat-marketplace-txp2w\" (UID: \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\") " pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.168693 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-t6bkt"] Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.169797 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.196154 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:07:58 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:07:58 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:07:58 crc kubenswrapper[4632]: healthz check failed Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.196217 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.232646 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t6bkt"] Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.239964 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5jx8\" (UniqueName: \"kubernetes.io/projected/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-kube-api-access-z5jx8\") pod \"redhat-marketplace-txp2w\" (UID: \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\") " pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.243211 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:58 crc kubenswrapper[4632]: E0313 10:07:58.243890 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:58.743871679 +0000 UTC m=+252.766401812 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.297284 4632 ???:1] "http: TLS handshake error from 192.168.126.11:53504: no serving certificate available for the kubelet" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.345060 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-z2gc7"] Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.350882 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:58 crc kubenswrapper[4632]: E0313 10:07:58.371137 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:58.871103111 +0000 UTC m=+252.893633254 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:58 crc kubenswrapper[4632]: E0313 10:07:58.386769 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:58.886752878 +0000 UTC m=+252.909283011 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.386350 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.386868 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-utilities\") pod \"redhat-marketplace-t6bkt\" (UID: \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\") " pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.386910 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-catalog-content\") pod \"redhat-marketplace-t6bkt\" (UID: \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\") " pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.387065 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc6ff\" (UniqueName: \"kubernetes.io/projected/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-kube-api-access-bc6ff\") pod \"redhat-marketplace-t6bkt\" (UID: \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\") " pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.397452 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.423007 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.507213 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.509600 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z2gc7"] Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.509828 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.511258 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc6ff\" (UniqueName: \"kubernetes.io/projected/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-kube-api-access-bc6ff\") pod \"redhat-marketplace-t6bkt\" (UID: \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\") " pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:07:58 crc kubenswrapper[4632]: E0313 10:07:58.511298 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:59.011270535 +0000 UTC m=+253.033800668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.511341 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a110c276-8516-4f9e-a6af-d6837cd0f387-catalog-content\") pod \"redhat-operators-z2gc7\" (UID: \"a110c276-8516-4f9e-a6af-d6837cd0f387\") " pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.511494 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a110c276-8516-4f9e-a6af-d6837cd0f387-utilities\") pod \"redhat-operators-z2gc7\" (UID: \"a110c276-8516-4f9e-a6af-d6837cd0f387\") " pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.511546 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.511576 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-utilities\") pod \"redhat-marketplace-t6bkt\" (UID: \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\") " pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.511599 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfdk8\" (UniqueName: \"kubernetes.io/projected/a110c276-8516-4f9e-a6af-d6837cd0f387-kube-api-access-tfdk8\") pod \"redhat-operators-z2gc7\" (UID: \"a110c276-8516-4f9e-a6af-d6837cd0f387\") " pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.511617 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-catalog-content\") pod \"redhat-marketplace-t6bkt\" (UID: \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\") " pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.512229 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-catalog-content\") pod \"redhat-marketplace-t6bkt\" (UID: \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\") " pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:07:58 crc kubenswrapper[4632]: E0313 10:07:58.512519 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:59.01251024 +0000 UTC m=+253.035040373 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.512807 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-utilities\") pod \"redhat-marketplace-t6bkt\" (UID: \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\") " pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.592631 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xr5l9"] Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.614428 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.615330 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.615524 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a110c276-8516-4f9e-a6af-d6837cd0f387-catalog-content\") pod \"redhat-operators-z2gc7\" (UID: \"a110c276-8516-4f9e-a6af-d6837cd0f387\") " pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.615595 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a110c276-8516-4f9e-a6af-d6837cd0f387-utilities\") pod \"redhat-operators-z2gc7\" (UID: \"a110c276-8516-4f9e-a6af-d6837cd0f387\") " pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.615626 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfdk8\" (UniqueName: \"kubernetes.io/projected/a110c276-8516-4f9e-a6af-d6837cd0f387-kube-api-access-tfdk8\") pod \"redhat-operators-z2gc7\" (UID: \"a110c276-8516-4f9e-a6af-d6837cd0f387\") " pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:07:58 crc kubenswrapper[4632]: E0313 10:07:58.616018 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:59.116002981 +0000 UTC m=+253.138533114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.616433 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a110c276-8516-4f9e-a6af-d6837cd0f387-catalog-content\") pod \"redhat-operators-z2gc7\" (UID: \"a110c276-8516-4f9e-a6af-d6837cd0f387\") " pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.618001 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a110c276-8516-4f9e-a6af-d6837cd0f387-utilities\") pod \"redhat-operators-z2gc7\" (UID: \"a110c276-8516-4f9e-a6af-d6837cd0f387\") " pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.681496 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xr5l9"] Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.722350 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc6ff\" (UniqueName: \"kubernetes.io/projected/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-kube-api-access-bc6ff\") pod \"redhat-marketplace-t6bkt\" (UID: \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\") " pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.723658 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87965e39-b879-4e26-9c8b-b78068c52aa0-utilities\") pod \"redhat-operators-xr5l9\" (UID: \"87965e39-b879-4e26-9c8b-b78068c52aa0\") " pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.723751 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hfz2\" (UniqueName: \"kubernetes.io/projected/87965e39-b879-4e26-9c8b-b78068c52aa0-kube-api-access-6hfz2\") pod \"redhat-operators-xr5l9\" (UID: \"87965e39-b879-4e26-9c8b-b78068c52aa0\") " pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.723778 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87965e39-b879-4e26-9c8b-b78068c52aa0-catalog-content\") pod \"redhat-operators-xr5l9\" (UID: \"87965e39-b879-4e26-9c8b-b78068c52aa0\") " pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.723824 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:58 crc kubenswrapper[4632]: E0313 10:07:58.724205 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:59.224191996 +0000 UTC m=+253.246722129 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.751672 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfdk8\" (UniqueName: \"kubernetes.io/projected/a110c276-8516-4f9e-a6af-d6837cd0f387-kube-api-access-tfdk8\") pod \"redhat-operators-z2gc7\" (UID: \"a110c276-8516-4f9e-a6af-d6837cd0f387\") " pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.770764 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" event={"ID":"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47","Type":"ContainerStarted","Data":"ad8f763365560329a7f776a97375e1fdcc2fb2412beb2907eed4da851706f983"} Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.775708 4632 generic.go:334] "Generic (PLEG): container finished" podID="cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" containerID="39c617653cdae12029a38a740d3aa9e4c08c056d9865caf4f87830fbf0817555" exitCode=0 Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.775773 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xd455" event={"ID":"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8","Type":"ContainerDied","Data":"39c617653cdae12029a38a740d3aa9e4c08c056d9865caf4f87830fbf0817555"} Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.805092 4632 generic.go:334] "Generic (PLEG): container finished" podID="bd46ae04-0610-4aa5-9385-dd45de66c5dd" containerID="eabb475f877c5898896f887fa631fab417c1e3579d0424b2b6c06f4278f091af" exitCode=0 Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.805140 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvh86" event={"ID":"bd46ae04-0610-4aa5-9385-dd45de66c5dd","Type":"ContainerDied","Data":"eabb475f877c5898896f887fa631fab417c1e3579d0424b2b6c06f4278f091af"} Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.824793 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.824999 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87965e39-b879-4e26-9c8b-b78068c52aa0-utilities\") pod \"redhat-operators-xr5l9\" (UID: \"87965e39-b879-4e26-9c8b-b78068c52aa0\") " pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.825090 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hfz2\" (UniqueName: \"kubernetes.io/projected/87965e39-b879-4e26-9c8b-b78068c52aa0-kube-api-access-6hfz2\") pod \"redhat-operators-xr5l9\" (UID: \"87965e39-b879-4e26-9c8b-b78068c52aa0\") " pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.825115 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87965e39-b879-4e26-9c8b-b78068c52aa0-catalog-content\") pod \"redhat-operators-xr5l9\" (UID: \"87965e39-b879-4e26-9c8b-b78068c52aa0\") " pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.825630 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87965e39-b879-4e26-9c8b-b78068c52aa0-catalog-content\") pod \"redhat-operators-xr5l9\" (UID: \"87965e39-b879-4e26-9c8b-b78068c52aa0\") " pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:07:58 crc kubenswrapper[4632]: E0313 10:07:58.825703 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:59.325686676 +0000 UTC m=+253.348216809 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.826106 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87965e39-b879-4e26-9c8b-b78068c52aa0-utilities\") pod \"redhat-operators-xr5l9\" (UID: \"87965e39-b879-4e26-9c8b-b78068c52aa0\") " pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.878766 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hfz2\" (UniqueName: \"kubernetes.io/projected/87965e39-b879-4e26-9c8b-b78068c52aa0-kube-api-access-6hfz2\") pod \"redhat-operators-xr5l9\" (UID: \"87965e39-b879-4e26-9c8b-b78068c52aa0\") " pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.929470 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:58 crc kubenswrapper[4632]: E0313 10:07:58.930379 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:59.430358931 +0000 UTC m=+253.452889064 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.978014 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:07:58 crc kubenswrapper[4632]: I0313 10:07:58.992412 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.050436 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.051231 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:59 crc kubenswrapper[4632]: E0313 10:07:59.051710 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:59.551688172 +0000 UTC m=+253.574218315 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.160958 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:59 crc kubenswrapper[4632]: E0313 10:07:59.161484 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:59.661468871 +0000 UTC m=+253.683999004 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.164243 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.211179 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:07:59 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:07:59 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:07:59 crc kubenswrapper[4632]: healthz check failed Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.211246 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.271600 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:59 crc kubenswrapper[4632]: E0313 10:07:59.272792 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:59.772772739 +0000 UTC m=+253.795302872 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.373146 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:59 crc kubenswrapper[4632]: E0313 10:07:59.373532 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:07:59.873510883 +0000 UTC m=+253.896041076 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.476515 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:59 crc kubenswrapper[4632]: E0313 10:07:59.476894 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:07:59.976876111 +0000 UTC m=+253.999406244 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.583896 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:59 crc kubenswrapper[4632]: E0313 10:07:59.584563 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:08:00.084550656 +0000 UTC m=+254.107080789 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.618463 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8z668"] Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.687974 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:59 crc kubenswrapper[4632]: E0313 10:07:59.688853 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:08:00.188616438 +0000 UTC m=+254.211146571 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.793720 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:07:59 crc kubenswrapper[4632]: E0313 10:07:59.794105 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:08:00.294092929 +0000 UTC m=+254.316623062 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.892459 4632 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.895484 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:07:59 crc kubenswrapper[4632]: E0313 10:07:59.901605 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:08:00.401580149 +0000 UTC m=+254.424110282 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.909691 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" event={"ID":"09ddc697-7ac1-4896-b9e2-1ae6c59c6f47","Type":"ContainerStarted","Data":"8d96e921fd05bbceb9759f1aec6352a154d23a6ac924ac647eb7c8c7bda71f68"} Mar 13 10:07:59 crc kubenswrapper[4632]: I0313 10:07:59.920162 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z668" event={"ID":"9845f384-2720-4d6a-aa73-1e66e30f7c2c","Type":"ContainerStarted","Data":"eb6537c579cc3249bae831f8164a219c024fbc6e74b0df55017ce52d6b143567"} Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.002422 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:08:00 crc kubenswrapper[4632]: E0313 10:08:00.002760 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:08:00.502722673 +0000 UTC m=+254.525252806 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.018094 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p8wjg"] Mar 13 10:08:00 crc kubenswrapper[4632]: W0313 10:08:00.069207 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb11a7dff_bf08_44c3_b4f4_923119c13717.slice/crio-67d236def43f1634b091443716f5df0abcd64ee4e8ef6768dd906ab3397df097 WatchSource:0}: Error finding container 67d236def43f1634b091443716f5df0abcd64ee4e8ef6768dd906ab3397df097: Status 404 returned error can't find the container with id 67d236def43f1634b091443716f5df0abcd64ee4e8ef6768dd906ab3397df097 Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.104584 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:08:00 crc kubenswrapper[4632]: E0313 10:08:00.105218 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:08:00.605195082 +0000 UTC m=+254.627725215 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.201969 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:00 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:00 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:00 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.202308 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.207011 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:08:00 crc kubenswrapper[4632]: E0313 10:08:00.207376 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:08:00.707363945 +0000 UTC m=+254.729894078 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.318180 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:08:00 crc kubenswrapper[4632]: E0313 10:08:00.318634 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:08:00.818600172 +0000 UTC m=+254.841130305 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.421004 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:08:00 crc kubenswrapper[4632]: E0313 10:08:00.421348 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:08:00.921335128 +0000 UTC m=+254.943865261 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.472099 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556608-9kzfk"] Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.473013 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556608-9kzfk" Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.487198 4632 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-13T10:07:59.892863333Z","Handler":null,"Name":""} Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.498810 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.518890 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xr5l9"] Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.525488 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:08:00 crc kubenswrapper[4632]: E0313 10:08:00.525822 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 10:08:01.025793343 +0000 UTC m=+255.048323476 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.525856 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.525914 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmxng\" (UniqueName: \"kubernetes.io/projected/37ab6711-478f-4cc7-b9a4-c9baa126b1a3-kube-api-access-dmxng\") pod \"auto-csr-approver-29556608-9kzfk\" (UID: \"37ab6711-478f-4cc7-b9a4-c9baa126b1a3\") " pod="openshift-infra/auto-csr-approver-29556608-9kzfk" Mar 13 10:08:00 crc kubenswrapper[4632]: E0313 10:08:00.526263 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 10:08:01.026246774 +0000 UTC m=+255.048776907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-fxs5z" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.547480 4632 patch_prober.go:28] interesting pod/apiserver-76f77b778f-p9gp2 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 13 10:08:00 crc kubenswrapper[4632]: [+]log ok Mar 13 10:08:00 crc kubenswrapper[4632]: [+]etcd ok Mar 13 10:08:00 crc kubenswrapper[4632]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 13 10:08:00 crc kubenswrapper[4632]: [+]poststarthook/generic-apiserver-start-informers ok Mar 13 10:08:00 crc kubenswrapper[4632]: [+]poststarthook/max-in-flight-filter ok Mar 13 10:08:00 crc kubenswrapper[4632]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 13 10:08:00 crc kubenswrapper[4632]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 13 10:08:00 crc kubenswrapper[4632]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 13 10:08:00 crc kubenswrapper[4632]: [+]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa ok Mar 13 10:08:00 crc kubenswrapper[4632]: [+]poststarthook/project.openshift.io-projectcache ok Mar 13 10:08:00 crc kubenswrapper[4632]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 13 10:08:00 crc kubenswrapper[4632]: [+]poststarthook/openshift.io-startinformers ok Mar 13 10:08:00 crc kubenswrapper[4632]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 13 10:08:00 crc kubenswrapper[4632]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 13 10:08:00 crc kubenswrapper[4632]: livez check failed Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.547547 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" podUID="37df1143-69fc-4d13-a5d3-790a9d14814a" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.568205 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556608-9kzfk"] Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.586767 4632 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.586814 4632 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.629352 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.629666 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmxng\" (UniqueName: \"kubernetes.io/projected/37ab6711-478f-4cc7-b9a4-c9baa126b1a3-kube-api-access-dmxng\") pod \"auto-csr-approver-29556608-9kzfk\" (UID: \"37ab6711-478f-4cc7-b9a4-c9baa126b1a3\") " pod="openshift-infra/auto-csr-approver-29556608-9kzfk" Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.639106 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.665772 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-z2gc7"] Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.700289 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmxng\" (UniqueName: \"kubernetes.io/projected/37ab6711-478f-4cc7-b9a4-c9baa126b1a3-kube-api-access-dmxng\") pod \"auto-csr-approver-29556608-9kzfk\" (UID: \"37ab6711-478f-4cc7-b9a4-c9baa126b1a3\") " pod="openshift-infra/auto-csr-approver-29556608-9kzfk" Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.719289 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-txp2w"] Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.732066 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.821324 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-t6bkt"] Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.860462 4632 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.860513 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.888310 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556608-9kzfk" Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.889502 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9v5nn"] Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.889742 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" podUID="70f440bb-5dd8-4863-9749-bc5f7c547750" containerName="controller-manager" containerID="cri-o://cacc884e7672aacd612df662055c2d9769da0a235fec8c1ddc593601e1331830" gracePeriod=30 Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.932202 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz"] Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.932412 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" podUID="560e6c43-4285-4ca8-98b9-874e9dcb5810" containerName="route-controller-manager" containerID="cri-o://6158a81f875e1232b8d27bb41ad2531364da77cdf3704ac46d4ec2470ad3e550" gracePeriod=30 Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.986629 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr5l9" event={"ID":"87965e39-b879-4e26-9c8b-b78068c52aa0","Type":"ContainerStarted","Data":"9d37b680fdc1d8687e48df9dab9cd8ad8fcee9b7cdb15c920f34a9cbf7bad5ef"} Mar 13 10:08:00 crc kubenswrapper[4632]: I0313 10:08:00.994647 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2gc7" event={"ID":"a110c276-8516-4f9e-a6af-d6837cd0f387","Type":"ContainerStarted","Data":"97491a7f994f5c8dffa29a28fb1914c53f3fb5687971c6cdb3d3b5b636967634"} Mar 13 10:08:01 crc kubenswrapper[4632]: I0313 10:08:01.017718 4632 generic.go:334] "Generic (PLEG): container finished" podID="b11a7dff-bf08-44c3-b4f4-923119c13717" containerID="31ed0687958629bbe6ae3de064bae07567e401a6f6f2576bf2e48b7390937742" exitCode=0 Mar 13 10:08:01 crc kubenswrapper[4632]: I0313 10:08:01.018058 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8wjg" event={"ID":"b11a7dff-bf08-44c3-b4f4-923119c13717","Type":"ContainerDied","Data":"31ed0687958629bbe6ae3de064bae07567e401a6f6f2576bf2e48b7390937742"} Mar 13 10:08:01 crc kubenswrapper[4632]: I0313 10:08:01.018110 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8wjg" event={"ID":"b11a7dff-bf08-44c3-b4f4-923119c13717","Type":"ContainerStarted","Data":"67d236def43f1634b091443716f5df0abcd64ee4e8ef6768dd906ab3397df097"} Mar 13 10:08:01 crc kubenswrapper[4632]: I0313 10:08:01.074460 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txp2w" event={"ID":"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87","Type":"ContainerStarted","Data":"3442fa414f5a2c2798e2a9a29c903f3acac1f4e2b61c872fefc305318ea1c556"} Mar 13 10:08:01 crc kubenswrapper[4632]: I0313 10:08:01.111971 4632 generic.go:334] "Generic (PLEG): container finished" podID="9845f384-2720-4d6a-aa73-1e66e30f7c2c" containerID="a609582fe9641518ec575a14e6a93f5bb1f502cb63d1e38602356c26bed99217" exitCode=0 Mar 13 10:08:01 crc kubenswrapper[4632]: I0313 10:08:01.112049 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z668" event={"ID":"9845f384-2720-4d6a-aa73-1e66e30f7c2c","Type":"ContainerDied","Data":"a609582fe9641518ec575a14e6a93f5bb1f502cb63d1e38602356c26bed99217"} Mar 13 10:08:01 crc kubenswrapper[4632]: I0313 10:08:01.124892 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-fxs5z\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:08:01 crc kubenswrapper[4632]: I0313 10:08:01.150995 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6bkt" event={"ID":"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e","Type":"ContainerStarted","Data":"3e3a79d99a0e6a35edab86938ccf523a35c4606e460775b549d1924f20dc4204"} Mar 13 10:08:01 crc kubenswrapper[4632]: I0313 10:08:01.233306 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:01 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:01 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:01 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:01 crc kubenswrapper[4632]: I0313 10:08:01.233770 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:01 crc kubenswrapper[4632]: I0313 10:08:01.443437 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 13 10:08:01 crc kubenswrapper[4632]: I0313 10:08:01.444852 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:08:01 crc kubenswrapper[4632]: I0313 10:08:01.942067 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:08:01 crc kubenswrapper[4632]: I0313 10:08:01.966837 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" podStartSLOduration=18.966820554999998 podStartE2EDuration="18.966820555s" podCreationTimestamp="2026-03-13 10:07:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:08:01.250843074 +0000 UTC m=+255.273373207" watchObservedRunningTime="2026-03-13 10:08:01.966820555 +0000 UTC m=+255.989350688" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.077503 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.083234 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556608-9kzfk"] Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.101046 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.117443 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-config\") pod \"70f440bb-5dd8-4863-9749-bc5f7c547750\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.117518 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-client-ca\") pod \"70f440bb-5dd8-4863-9749-bc5f7c547750\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.117589 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-proxy-ca-bundles\") pod \"70f440bb-5dd8-4863-9749-bc5f7c547750\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.117610 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vvbt\" (UniqueName: \"kubernetes.io/projected/70f440bb-5dd8-4863-9749-bc5f7c547750-kube-api-access-6vvbt\") pod \"70f440bb-5dd8-4863-9749-bc5f7c547750\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.117642 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70f440bb-5dd8-4863-9749-bc5f7c547750-serving-cert\") pod \"70f440bb-5dd8-4863-9749-bc5f7c547750\" (UID: \"70f440bb-5dd8-4863-9749-bc5f7c547750\") " Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.119015 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "70f440bb-5dd8-4863-9749-bc5f7c547750" (UID: "70f440bb-5dd8-4863-9749-bc5f7c547750"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.121647 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-client-ca" (OuterVolumeSpecName: "client-ca") pod "70f440bb-5dd8-4863-9749-bc5f7c547750" (UID: "70f440bb-5dd8-4863-9749-bc5f7c547750"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.126181 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-config" (OuterVolumeSpecName: "config") pod "70f440bb-5dd8-4863-9749-bc5f7c547750" (UID: "70f440bb-5dd8-4863-9749-bc5f7c547750"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.137357 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70f440bb-5dd8-4863-9749-bc5f7c547750-kube-api-access-6vvbt" (OuterVolumeSpecName: "kube-api-access-6vvbt") pod "70f440bb-5dd8-4863-9749-bc5f7c547750" (UID: "70f440bb-5dd8-4863-9749-bc5f7c547750"). InnerVolumeSpecName "kube-api-access-6vvbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.146109 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/70f440bb-5dd8-4863-9749-bc5f7c547750-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "70f440bb-5dd8-4863-9749-bc5f7c547750" (UID: "70f440bb-5dd8-4863-9749-bc5f7c547750"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.189447 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556608-9kzfk" event={"ID":"37ab6711-478f-4cc7-b9a4-c9baa126b1a3","Type":"ContainerStarted","Data":"ff362806bee1867b720f220a4cde4dbe8551207f73438d3af60407d151505f16"} Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.198087 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:02 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:02 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:02 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.198157 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.222416 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/560e6c43-4285-4ca8-98b9-874e9dcb5810-serving-cert\") pod \"560e6c43-4285-4ca8-98b9-874e9dcb5810\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.222499 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/560e6c43-4285-4ca8-98b9-874e9dcb5810-client-ca\") pod \"560e6c43-4285-4ca8-98b9-874e9dcb5810\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.222523 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/560e6c43-4285-4ca8-98b9-874e9dcb5810-config\") pod \"560e6c43-4285-4ca8-98b9-874e9dcb5810\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.222637 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdwcs\" (UniqueName: \"kubernetes.io/projected/560e6c43-4285-4ca8-98b9-874e9dcb5810-kube-api-access-sdwcs\") pod \"560e6c43-4285-4ca8-98b9-874e9dcb5810\" (UID: \"560e6c43-4285-4ca8-98b9-874e9dcb5810\") " Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.222983 4632 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.222997 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vvbt\" (UniqueName: \"kubernetes.io/projected/70f440bb-5dd8-4863-9749-bc5f7c547750-kube-api-access-6vvbt\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.223021 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/70f440bb-5dd8-4863-9749-bc5f7c547750-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.223031 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.223042 4632 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/70f440bb-5dd8-4863-9749-bc5f7c547750-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.224184 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/560e6c43-4285-4ca8-98b9-874e9dcb5810-client-ca" (OuterVolumeSpecName: "client-ca") pod "560e6c43-4285-4ca8-98b9-874e9dcb5810" (UID: "560e6c43-4285-4ca8-98b9-874e9dcb5810"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.224760 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/560e6c43-4285-4ca8-98b9-874e9dcb5810-config" (OuterVolumeSpecName: "config") pod "560e6c43-4285-4ca8-98b9-874e9dcb5810" (UID: "560e6c43-4285-4ca8-98b9-874e9dcb5810"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.233436 4632 generic.go:334] "Generic (PLEG): container finished" podID="70f440bb-5dd8-4863-9749-bc5f7c547750" containerID="cacc884e7672aacd612df662055c2d9769da0a235fec8c1ddc593601e1331830" exitCode=0 Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.233752 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" event={"ID":"70f440bb-5dd8-4863-9749-bc5f7c547750","Type":"ContainerDied","Data":"cacc884e7672aacd612df662055c2d9769da0a235fec8c1ddc593601e1331830"} Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.233830 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" event={"ID":"70f440bb-5dd8-4863-9749-bc5f7c547750","Type":"ContainerDied","Data":"365524316a4e3e846e005a856282706fac826be9337ec760f74d5dd19061bccd"} Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.233897 4632 scope.go:117] "RemoveContainer" containerID="cacc884e7672aacd612df662055c2d9769da0a235fec8c1ddc593601e1331830" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.235456 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9v5nn" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.239264 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/560e6c43-4285-4ca8-98b9-874e9dcb5810-kube-api-access-sdwcs" (OuterVolumeSpecName: "kube-api-access-sdwcs") pod "560e6c43-4285-4ca8-98b9-874e9dcb5810" (UID: "560e6c43-4285-4ca8-98b9-874e9dcb5810"). InnerVolumeSpecName "kube-api-access-sdwcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.263322 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/560e6c43-4285-4ca8-98b9-874e9dcb5810-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "560e6c43-4285-4ca8-98b9-874e9dcb5810" (UID: "560e6c43-4285-4ca8-98b9-874e9dcb5810"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.280662 4632 generic.go:334] "Generic (PLEG): container finished" podID="528d3aa9-10bf-4029-a4d2-85768264fde8" containerID="da165dd4ae62fa2ea1c777c8125fcd4bfe4bd102f508da056f1a058689bba35e" exitCode=0 Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.280749 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" event={"ID":"528d3aa9-10bf-4029-a4d2-85768264fde8","Type":"ContainerDied","Data":"da165dd4ae62fa2ea1c777c8125fcd4bfe4bd102f508da056f1a058689bba35e"} Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.309818 4632 generic.go:334] "Generic (PLEG): container finished" podID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" containerID="01ae56b596391f3b7877c67539058596dbfd086754ed8db1f4f40f76d82a4c4c" exitCode=0 Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.309929 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6bkt" event={"ID":"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e","Type":"ContainerDied","Data":"01ae56b596391f3b7877c67539058596dbfd086754ed8db1f4f40f76d82a4c4c"} Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.326019 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdwcs\" (UniqueName: \"kubernetes.io/projected/560e6c43-4285-4ca8-98b9-874e9dcb5810-kube-api-access-sdwcs\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.326074 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/560e6c43-4285-4ca8-98b9-874e9dcb5810-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.326797 4632 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/560e6c43-4285-4ca8-98b9-874e9dcb5810-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.326832 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/560e6c43-4285-4ca8-98b9-874e9dcb5810-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.334717 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9v5nn"] Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.348402 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9v5nn"] Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.367101 4632 generic.go:334] "Generic (PLEG): container finished" podID="87965e39-b879-4e26-9c8b-b78068c52aa0" containerID="bafd48184b1b00528329793cfe1af87f0aa9502582cddad85d784407e60c249d" exitCode=0 Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.367167 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr5l9" event={"ID":"87965e39-b879-4e26-9c8b-b78068c52aa0","Type":"ContainerDied","Data":"bafd48184b1b00528329793cfe1af87f0aa9502582cddad85d784407e60c249d"} Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.374177 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2gc7" event={"ID":"a110c276-8516-4f9e-a6af-d6837cd0f387","Type":"ContainerDied","Data":"0d073c1adaa82aa87cab8618a50587cfed8b79fe657e3f2878a87c7599c612fb"} Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.374662 4632 generic.go:334] "Generic (PLEG): container finished" podID="a110c276-8516-4f9e-a6af-d6837cd0f387" containerID="0d073c1adaa82aa87cab8618a50587cfed8b79fe657e3f2878a87c7599c612fb" exitCode=0 Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.384890 4632 generic.go:334] "Generic (PLEG): container finished" podID="f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" containerID="d1da7a7847a6ff5346add9e3ed943cdc6232146978e6161d764011992ac73c84" exitCode=0 Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.385000 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txp2w" event={"ID":"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87","Type":"ContainerDied","Data":"d1da7a7847a6ff5346add9e3ed943cdc6232146978e6161d764011992ac73c84"} Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.399512 4632 generic.go:334] "Generic (PLEG): container finished" podID="560e6c43-4285-4ca8-98b9-874e9dcb5810" containerID="6158a81f875e1232b8d27bb41ad2531364da77cdf3704ac46d4ec2470ad3e550" exitCode=0 Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.400428 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.403915 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" event={"ID":"560e6c43-4285-4ca8-98b9-874e9dcb5810","Type":"ContainerDied","Data":"6158a81f875e1232b8d27bb41ad2531364da77cdf3704ac46d4ec2470ad3e550"} Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.403994 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz" event={"ID":"560e6c43-4285-4ca8-98b9-874e9dcb5810","Type":"ContainerDied","Data":"e9aab0e9cd1796940dcc2818af221f5b388f490c5b2161fb3217fdbc24d92e66"} Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.438510 4632 scope.go:117] "RemoveContainer" containerID="cacc884e7672aacd612df662055c2d9769da0a235fec8c1ddc593601e1331830" Mar 13 10:08:02 crc kubenswrapper[4632]: E0313 10:08:02.441216 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cacc884e7672aacd612df662055c2d9769da0a235fec8c1ddc593601e1331830\": container with ID starting with cacc884e7672aacd612df662055c2d9769da0a235fec8c1ddc593601e1331830 not found: ID does not exist" containerID="cacc884e7672aacd612df662055c2d9769da0a235fec8c1ddc593601e1331830" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.441269 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cacc884e7672aacd612df662055c2d9769da0a235fec8c1ddc593601e1331830"} err="failed to get container status \"cacc884e7672aacd612df662055c2d9769da0a235fec8c1ddc593601e1331830\": rpc error: code = NotFound desc = could not find container \"cacc884e7672aacd612df662055c2d9769da0a235fec8c1ddc593601e1331830\": container with ID starting with cacc884e7672aacd612df662055c2d9769da0a235fec8c1ddc593601e1331830 not found: ID does not exist" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.441298 4632 scope.go:117] "RemoveContainer" containerID="6158a81f875e1232b8d27bb41ad2531364da77cdf3704ac46d4ec2470ad3e550" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.494343 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxs5z"] Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.522759 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz"] Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.534712 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-xthqz"] Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.554101 4632 scope.go:117] "RemoveContainer" containerID="6158a81f875e1232b8d27bb41ad2531364da77cdf3704ac46d4ec2470ad3e550" Mar 13 10:08:02 crc kubenswrapper[4632]: E0313 10:08:02.554662 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6158a81f875e1232b8d27bb41ad2531364da77cdf3704ac46d4ec2470ad3e550\": container with ID starting with 6158a81f875e1232b8d27bb41ad2531364da77cdf3704ac46d4ec2470ad3e550 not found: ID does not exist" containerID="6158a81f875e1232b8d27bb41ad2531364da77cdf3704ac46d4ec2470ad3e550" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.554704 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6158a81f875e1232b8d27bb41ad2531364da77cdf3704ac46d4ec2470ad3e550"} err="failed to get container status \"6158a81f875e1232b8d27bb41ad2531364da77cdf3704ac46d4ec2470ad3e550\": rpc error: code = NotFound desc = could not find container \"6158a81f875e1232b8d27bb41ad2531364da77cdf3704ac46d4ec2470ad3e550\": container with ID starting with 6158a81f875e1232b8d27bb41ad2531364da77cdf3704ac46d4ec2470ad3e550 not found: ID does not exist" Mar 13 10:08:02 crc kubenswrapper[4632]: W0313 10:08:02.592099 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf56fc09a_e2b7_46db_b938_f276df3f033e.slice/crio-6db13fe4cd83b1210971879bf1313cee58732376958e857687de7da1568c6519 WatchSource:0}: Error finding container 6db13fe4cd83b1210971879bf1313cee58732376958e857687de7da1568c6519: Status 404 returned error can't find the container with id 6db13fe4cd83b1210971879bf1313cee58732376958e857687de7da1568c6519 Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.828215 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-658cc96bdc-92bpr"] Mar 13 10:08:02 crc kubenswrapper[4632]: E0313 10:08:02.828753 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70f440bb-5dd8-4863-9749-bc5f7c547750" containerName="controller-manager" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.828774 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="70f440bb-5dd8-4863-9749-bc5f7c547750" containerName="controller-manager" Mar 13 10:08:02 crc kubenswrapper[4632]: E0313 10:08:02.828787 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="560e6c43-4285-4ca8-98b9-874e9dcb5810" containerName="route-controller-manager" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.828794 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="560e6c43-4285-4ca8-98b9-874e9dcb5810" containerName="route-controller-manager" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.828921 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="70f440bb-5dd8-4863-9749-bc5f7c547750" containerName="controller-manager" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.829015 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="560e6c43-4285-4ca8-98b9-874e9dcb5810" containerName="route-controller-manager" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.829416 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.835378 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.835588 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.839069 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.839292 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.839452 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.839564 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.863368 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp"] Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.865702 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.869101 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.882414 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.884848 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.886623 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.888384 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.889068 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.893613 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.926343 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-658cc96bdc-92bpr"] Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.937596 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwf9m\" (UniqueName: \"kubernetes.io/projected/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-kube-api-access-hwf9m\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.937679 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0f45c9a-e32c-420e-9106-fcb72dd59350-serving-cert\") pod \"route-controller-manager-86d66b8bfb-9wzpp\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.937793 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-client-ca\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.937821 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0f45c9a-e32c-420e-9106-fcb72dd59350-config\") pod \"route-controller-manager-86d66b8bfb-9wzpp\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.937995 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-proxy-ca-bundles\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.938093 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f0f45c9a-e32c-420e-9106-fcb72dd59350-client-ca\") pod \"route-controller-manager-86d66b8bfb-9wzpp\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.938211 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9pxp\" (UniqueName: \"kubernetes.io/projected/f0f45c9a-e32c-420e-9106-fcb72dd59350-kube-api-access-l9pxp\") pod \"route-controller-manager-86d66b8bfb-9wzpp\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.938258 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-config\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.938384 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-serving-cert\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.939263 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp"] Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.946265 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.947708 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.948846 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.952380 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 13 10:08:02 crc kubenswrapper[4632]: I0313 10:08:02.952737 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.041716 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9pxp\" (UniqueName: \"kubernetes.io/projected/f0f45c9a-e32c-420e-9106-fcb72dd59350-kube-api-access-l9pxp\") pod \"route-controller-manager-86d66b8bfb-9wzpp\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.047644 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-config\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.047755 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-serving-cert\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.047802 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwf9m\" (UniqueName: \"kubernetes.io/projected/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-kube-api-access-hwf9m\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.047856 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0f45c9a-e32c-420e-9106-fcb72dd59350-serving-cert\") pod \"route-controller-manager-86d66b8bfb-9wzpp\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.047978 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a4d2988-a460-407b-902a-aeb8eda619a1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5a4d2988-a460-407b-902a-aeb8eda619a1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.048022 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-client-ca\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.048058 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0f45c9a-e32c-420e-9106-fcb72dd59350-config\") pod \"route-controller-manager-86d66b8bfb-9wzpp\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.048083 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-proxy-ca-bundles\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.048111 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f0f45c9a-e32c-420e-9106-fcb72dd59350-client-ca\") pod \"route-controller-manager-86d66b8bfb-9wzpp\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.048151 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a4d2988-a460-407b-902a-aeb8eda619a1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5a4d2988-a460-407b-902a-aeb8eda619a1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.049117 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-client-ca\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.049308 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-config\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.050448 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-proxy-ca-bundles\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.051552 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0f45c9a-e32c-420e-9106-fcb72dd59350-config\") pod \"route-controller-manager-86d66b8bfb-9wzpp\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.052235 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f0f45c9a-e32c-420e-9106-fcb72dd59350-client-ca\") pod \"route-controller-manager-86d66b8bfb-9wzpp\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.069126 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0f45c9a-e32c-420e-9106-fcb72dd59350-serving-cert\") pod \"route-controller-manager-86d66b8bfb-9wzpp\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.069677 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-serving-cert\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.080617 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9pxp\" (UniqueName: \"kubernetes.io/projected/f0f45c9a-e32c-420e-9106-fcb72dd59350-kube-api-access-l9pxp\") pod \"route-controller-manager-86d66b8bfb-9wzpp\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.094743 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwf9m\" (UniqueName: \"kubernetes.io/projected/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-kube-api-access-hwf9m\") pod \"controller-manager-658cc96bdc-92bpr\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.156845 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a4d2988-a460-407b-902a-aeb8eda619a1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5a4d2988-a460-407b-902a-aeb8eda619a1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.156954 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a4d2988-a460-407b-902a-aeb8eda619a1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5a4d2988-a460-407b-902a-aeb8eda619a1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.157094 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a4d2988-a460-407b-902a-aeb8eda619a1-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5a4d2988-a460-407b-902a-aeb8eda619a1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.176877 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a4d2988-a460-407b-902a-aeb8eda619a1-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5a4d2988-a460-407b-902a-aeb8eda619a1\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.192661 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.195132 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:03 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:03 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:03 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.195204 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.227059 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.275046 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.414311 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.415156 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.415267 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.436531 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.436752 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.461129 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa472cd7-9575-473f-b6b6-709b644a5ec4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"aa472cd7-9575-473f-b6b6-709b644a5ec4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.461218 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa472cd7-9575-473f-b6b6-709b644a5ec4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"aa472cd7-9575-473f-b6b6-709b644a5ec4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.470181 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" event={"ID":"f56fc09a-e2b7-46db-b938-f276df3f033e","Type":"ContainerStarted","Data":"1081e88c7001e64d6f95133ef3938fcdaf6163c9ecf6555e86dc52149387161f"} Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.470230 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" event={"ID":"f56fc09a-e2b7-46db-b938-f276df3f033e","Type":"ContainerStarted","Data":"6db13fe4cd83b1210971879bf1313cee58732376958e857687de7da1568c6519"} Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.471067 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.500713 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" podStartSLOduration=203.500693275 podStartE2EDuration="3m23.500693275s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:08:03.496886344 +0000 UTC m=+257.519416477" watchObservedRunningTime="2026-03-13 10:08:03.500693275 +0000 UTC m=+257.523223408" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.551835 4632 ???:1] "http: TLS handshake error from 192.168.126.11:53520: no serving certificate available for the kubelet" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.563123 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa472cd7-9575-473f-b6b6-709b644a5ec4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"aa472cd7-9575-473f-b6b6-709b644a5ec4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.563289 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa472cd7-9575-473f-b6b6-709b644a5ec4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"aa472cd7-9575-473f-b6b6-709b644a5ec4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.564453 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa472cd7-9575-473f-b6b6-709b644a5ec4-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"aa472cd7-9575-473f-b6b6-709b644a5ec4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.586197 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa472cd7-9575-473f-b6b6-709b644a5ec4-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"aa472cd7-9575-473f-b6b6-709b644a5ec4\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 10:08:03 crc kubenswrapper[4632]: I0313 10:08:03.767243 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.080621 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="560e6c43-4285-4ca8-98b9-874e9dcb5810" path="/var/lib/kubelet/pods/560e6c43-4285-4ca8-98b9-874e9dcb5810/volumes" Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.082336 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70f440bb-5dd8-4863-9749-bc5f7c547750" path="/var/lib/kubelet/pods/70f440bb-5dd8-4863-9749-bc5f7c547750/volumes" Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.228416 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.247482 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:04 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:04 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:04 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.247564 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.320086 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/528d3aa9-10bf-4029-a4d2-85768264fde8-secret-volume\") pod \"528d3aa9-10bf-4029-a4d2-85768264fde8\" (UID: \"528d3aa9-10bf-4029-a4d2-85768264fde8\") " Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.320171 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/528d3aa9-10bf-4029-a4d2-85768264fde8-config-volume\") pod \"528d3aa9-10bf-4029-a4d2-85768264fde8\" (UID: \"528d3aa9-10bf-4029-a4d2-85768264fde8\") " Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.320218 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vm8lp\" (UniqueName: \"kubernetes.io/projected/528d3aa9-10bf-4029-a4d2-85768264fde8-kube-api-access-vm8lp\") pod \"528d3aa9-10bf-4029-a4d2-85768264fde8\" (UID: \"528d3aa9-10bf-4029-a4d2-85768264fde8\") " Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.324581 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/528d3aa9-10bf-4029-a4d2-85768264fde8-config-volume" (OuterVolumeSpecName: "config-volume") pod "528d3aa9-10bf-4029-a4d2-85768264fde8" (UID: "528d3aa9-10bf-4029-a4d2-85768264fde8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.339092 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/528d3aa9-10bf-4029-a4d2-85768264fde8-kube-api-access-vm8lp" (OuterVolumeSpecName: "kube-api-access-vm8lp") pod "528d3aa9-10bf-4029-a4d2-85768264fde8" (UID: "528d3aa9-10bf-4029-a4d2-85768264fde8"). InnerVolumeSpecName "kube-api-access-vm8lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.368074 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/528d3aa9-10bf-4029-a4d2-85768264fde8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "528d3aa9-10bf-4029-a4d2-85768264fde8" (UID: "528d3aa9-10bf-4029-a4d2-85768264fde8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.422664 4632 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/528d3aa9-10bf-4029-a4d2-85768264fde8-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.422775 4632 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/528d3aa9-10bf-4029-a4d2-85768264fde8-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.422791 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vm8lp\" (UniqueName: \"kubernetes.io/projected/528d3aa9-10bf-4029-a4d2-85768264fde8-kube-api-access-vm8lp\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.510318 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-658cc96bdc-92bpr"] Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.536218 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp"] Mar 13 10:08:04 crc kubenswrapper[4632]: W0313 10:08:04.581532 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8ffa9c9_d11d_46b5_ac51_6d38a8639d98.slice/crio-94bf0543076d266034298c32f3e17bb1e05e21aec66de3f477e64011186c779f WatchSource:0}: Error finding container 94bf0543076d266034298c32f3e17bb1e05e21aec66de3f477e64011186c779f: Status 404 returned error can't find the container with id 94bf0543076d266034298c32f3e17bb1e05e21aec66de3f477e64011186c779f Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.587867 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.638123 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" event={"ID":"528d3aa9-10bf-4029-a4d2-85768264fde8","Type":"ContainerDied","Data":"1137745f79e5dd4b86f11690ff5ed0914b045872452dc8054e30e019f43d068c"} Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.638184 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1137745f79e5dd4b86f11690ff5ed0914b045872452dc8054e30e019f43d068c" Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.638558 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg" Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.693724 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-g2wxc" Mar 13 10:08:04 crc kubenswrapper[4632]: W0313 10:08:04.785766 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf0f45c9a_e32c_420e_9106_fcb72dd59350.slice/crio-070e0df69129dffbb7be6eacfbcf901c3d12a23e3166a3f87eca493eaf3d27d9 WatchSource:0}: Error finding container 070e0df69129dffbb7be6eacfbcf901c3d12a23e3166a3f87eca493eaf3d27d9: Status 404 returned error can't find the container with id 070e0df69129dffbb7be6eacfbcf901c3d12a23e3166a3f87eca493eaf3d27d9 Mar 13 10:08:04 crc kubenswrapper[4632]: I0313 10:08:04.991210 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.196499 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:05 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:05 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:05 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.196784 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.522574 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.536073 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-p9gp2" Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.632674 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.633273 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.633748 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.633770 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.696513 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"aa472cd7-9575-473f-b6b6-709b644a5ec4","Type":"ContainerStarted","Data":"1e9a3663129374df63b373bcd43c56b70f7622faa20517540c6584e27d165001"} Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.699224 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" event={"ID":"f0f45c9a-e32c-420e-9106-fcb72dd59350","Type":"ContainerStarted","Data":"687f3a3904ba4bdede4a24f019fead48648120a9d8e838727216b0ec43fcb3a2"} Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.699277 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" event={"ID":"f0f45c9a-e32c-420e-9106-fcb72dd59350","Type":"ContainerStarted","Data":"070e0df69129dffbb7be6eacfbcf901c3d12a23e3166a3f87eca493eaf3d27d9"} Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.703616 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" event={"ID":"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98","Type":"ContainerStarted","Data":"fe1e770193ae7e14a37be92defae5c64d043b458e01244272a22574e7b2e1f74"} Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.703669 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" event={"ID":"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98","Type":"ContainerStarted","Data":"94bf0543076d266034298c32f3e17bb1e05e21aec66de3f477e64011186c779f"} Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.716300 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5a4d2988-a460-407b-902a-aeb8eda619a1","Type":"ContainerStarted","Data":"7fd2f4ed963308f80503bea68090613f7b4770625620e790958344e5d09eb8f5"} Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.807253 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" podStartSLOduration=4.807228306 podStartE2EDuration="4.807228306s" podCreationTimestamp="2026-03-13 10:08:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:08:05.804626447 +0000 UTC m=+259.827156580" watchObservedRunningTime="2026-03-13 10:08:05.807228306 +0000 UTC m=+259.829758439" Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.808515 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" podStartSLOduration=4.80850823 podStartE2EDuration="4.80850823s" podCreationTimestamp="2026-03-13 10:08:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:08:05.758512626 +0000 UTC m=+259.781042759" watchObservedRunningTime="2026-03-13 10:08:05.80850823 +0000 UTC m=+259.831038373" Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.983404 4632 patch_prober.go:28] interesting pod/console-f9d7485db-zn7mn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Mar 13 10:08:05 crc kubenswrapper[4632]: I0313 10:08:05.983475 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zn7mn" podUID="f5a50074-5531-442f-a0e9-0578f15634c1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" Mar 13 10:08:06 crc kubenswrapper[4632]: I0313 10:08:06.206156 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:06 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:06 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:06 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:06 crc kubenswrapper[4632]: I0313 10:08:06.206247 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:06 crc kubenswrapper[4632]: I0313 10:08:06.568775 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:08:06 crc kubenswrapper[4632]: I0313 10:08:06.808442 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5a4d2988-a460-407b-902a-aeb8eda619a1","Type":"ContainerStarted","Data":"8a2ad0a9e117d30bdffbdfe34a1c99a012b75db7ee1ca75eb636ee55b6520e15"} Mar 13 10:08:06 crc kubenswrapper[4632]: I0313 10:08:06.812608 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:06 crc kubenswrapper[4632]: I0313 10:08:06.813628 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:06 crc kubenswrapper[4632]: I0313 10:08:06.837206 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:06 crc kubenswrapper[4632]: I0313 10:08:06.839046 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:06 crc kubenswrapper[4632]: I0313 10:08:06.879816 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=4.879797186 podStartE2EDuration="4.879797186s" podCreationTimestamp="2026-03-13 10:08:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:08:06.845475588 +0000 UTC m=+260.868005741" watchObservedRunningTime="2026-03-13 10:08:06.879797186 +0000 UTC m=+260.902327329" Mar 13 10:08:07 crc kubenswrapper[4632]: I0313 10:08:07.196069 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:07 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:07 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:07 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:07 crc kubenswrapper[4632]: I0313 10:08:07.196135 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:07 crc kubenswrapper[4632]: I0313 10:08:07.937917 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"aa472cd7-9575-473f-b6b6-709b644a5ec4","Type":"ContainerStarted","Data":"bfd967ad660cb6670b66e0ac690d69022ebe216196a49d69939801b1e253860e"} Mar 13 10:08:07 crc kubenswrapper[4632]: I0313 10:08:07.945377 4632 generic.go:334] "Generic (PLEG): container finished" podID="5a4d2988-a460-407b-902a-aeb8eda619a1" containerID="8a2ad0a9e117d30bdffbdfe34a1c99a012b75db7ee1ca75eb636ee55b6520e15" exitCode=0 Mar 13 10:08:07 crc kubenswrapper[4632]: I0313 10:08:07.945583 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5a4d2988-a460-407b-902a-aeb8eda619a1","Type":"ContainerDied","Data":"8a2ad0a9e117d30bdffbdfe34a1c99a012b75db7ee1ca75eb636ee55b6520e15"} Mar 13 10:08:07 crc kubenswrapper[4632]: I0313 10:08:07.992034 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=4.991932752 podStartE2EDuration="4.991932752s" podCreationTimestamp="2026-03-13 10:08:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:08:07.963368456 +0000 UTC m=+261.985898599" watchObservedRunningTime="2026-03-13 10:08:07.991932752 +0000 UTC m=+262.014462885" Mar 13 10:08:08 crc kubenswrapper[4632]: I0313 10:08:08.196031 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:08 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:08 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:08 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:08 crc kubenswrapper[4632]: I0313 10:08:08.196142 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:09 crc kubenswrapper[4632]: I0313 10:08:09.002563 4632 generic.go:334] "Generic (PLEG): container finished" podID="aa472cd7-9575-473f-b6b6-709b644a5ec4" containerID="bfd967ad660cb6670b66e0ac690d69022ebe216196a49d69939801b1e253860e" exitCode=0 Mar 13 10:08:09 crc kubenswrapper[4632]: I0313 10:08:09.002673 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"aa472cd7-9575-473f-b6b6-709b644a5ec4","Type":"ContainerDied","Data":"bfd967ad660cb6670b66e0ac690d69022ebe216196a49d69939801b1e253860e"} Mar 13 10:08:09 crc kubenswrapper[4632]: I0313 10:08:09.194589 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:09 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:09 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:09 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:09 crc kubenswrapper[4632]: I0313 10:08:09.194719 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.108686 4632 ???:1] "http: TLS handshake error from 192.168.126.11:55198: no serving certificate available for the kubelet" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.118735 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5a4d2988-a460-407b-902a-aeb8eda619a1","Type":"ContainerDied","Data":"7fd2f4ed963308f80503bea68090613f7b4770625620e790958344e5d09eb8f5"} Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.118775 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fd2f4ed963308f80503bea68090613f7b4770625620e790958344e5d09eb8f5" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.135153 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.174432 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a4d2988-a460-407b-902a-aeb8eda619a1-kube-api-access\") pod \"5a4d2988-a460-407b-902a-aeb8eda619a1\" (UID: \"5a4d2988-a460-407b-902a-aeb8eda619a1\") " Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.174494 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a4d2988-a460-407b-902a-aeb8eda619a1-kubelet-dir\") pod \"5a4d2988-a460-407b-902a-aeb8eda619a1\" (UID: \"5a4d2988-a460-407b-902a-aeb8eda619a1\") " Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.174718 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a4d2988-a460-407b-902a-aeb8eda619a1-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5a4d2988-a460-407b-902a-aeb8eda619a1" (UID: "5a4d2988-a460-407b-902a-aeb8eda619a1"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.196214 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:10 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:10 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:10 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.196596 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.204680 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a4d2988-a460-407b-902a-aeb8eda619a1-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5a4d2988-a460-407b-902a-aeb8eda619a1" (UID: "5a4d2988-a460-407b-902a-aeb8eda619a1"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.275689 4632 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5a4d2988-a460-407b-902a-aeb8eda619a1-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.275730 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5a4d2988-a460-407b-902a-aeb8eda619a1-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.461127 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.461222 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.626464 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.675486 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.806085 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa472cd7-9575-473f-b6b6-709b644a5ec4-kube-api-access\") pod \"aa472cd7-9575-473f-b6b6-709b644a5ec4\" (UID: \"aa472cd7-9575-473f-b6b6-709b644a5ec4\") " Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.806133 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa472cd7-9575-473f-b6b6-709b644a5ec4-kubelet-dir\") pod \"aa472cd7-9575-473f-b6b6-709b644a5ec4\" (UID: \"aa472cd7-9575-473f-b6b6-709b644a5ec4\") " Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.813702 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa472cd7-9575-473f-b6b6-709b644a5ec4-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "aa472cd7-9575-473f-b6b6-709b644a5ec4" (UID: "aa472cd7-9575-473f-b6b6-709b644a5ec4"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.813853 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa472cd7-9575-473f-b6b6-709b644a5ec4-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "aa472cd7-9575-473f-b6b6-709b644a5ec4" (UID: "aa472cd7-9575-473f-b6b6-709b644a5ec4"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.907508 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/aa472cd7-9575-473f-b6b6-709b644a5ec4-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:10 crc kubenswrapper[4632]: I0313 10:08:10.907554 4632 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/aa472cd7-9575-473f-b6b6-709b644a5ec4-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:11 crc kubenswrapper[4632]: I0313 10:08:11.164154 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 10:08:11 crc kubenswrapper[4632]: I0313 10:08:11.164257 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"aa472cd7-9575-473f-b6b6-709b644a5ec4","Type":"ContainerDied","Data":"1e9a3663129374df63b373bcd43c56b70f7622faa20517540c6584e27d165001"} Mar 13 10:08:11 crc kubenswrapper[4632]: I0313 10:08:11.164295 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e9a3663129374df63b373bcd43c56b70f7622faa20517540c6584e27d165001" Mar 13 10:08:11 crc kubenswrapper[4632]: I0313 10:08:11.165158 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 10:08:11 crc kubenswrapper[4632]: I0313 10:08:11.197594 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:11 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:11 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:11 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:11 crc kubenswrapper[4632]: I0313 10:08:11.197710 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:12 crc kubenswrapper[4632]: I0313 10:08:12.195519 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:12 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:12 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:12 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:12 crc kubenswrapper[4632]: I0313 10:08:12.195621 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:13 crc kubenswrapper[4632]: I0313 10:08:13.200578 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:13 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:13 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:13 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:13 crc kubenswrapper[4632]: I0313 10:08:13.200670 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:13 crc kubenswrapper[4632]: I0313 10:08:13.823292 4632 ???:1] "http: TLS handshake error from 192.168.126.11:55212: no serving certificate available for the kubelet" Mar 13 10:08:14 crc kubenswrapper[4632]: I0313 10:08:14.194985 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:14 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:14 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:14 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:14 crc kubenswrapper[4632]: I0313 10:08:14.195111 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:15 crc kubenswrapper[4632]: I0313 10:08:15.219849 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:15 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:15 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:15 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:15 crc kubenswrapper[4632]: I0313 10:08:15.220049 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:15 crc kubenswrapper[4632]: I0313 10:08:15.633290 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:08:15 crc kubenswrapper[4632]: I0313 10:08:15.633431 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:08:15 crc kubenswrapper[4632]: I0313 10:08:15.636338 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:08:15 crc kubenswrapper[4632]: I0313 10:08:15.636402 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:08:15 crc kubenswrapper[4632]: I0313 10:08:15.636514 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-w2hhj" Mar 13 10:08:15 crc kubenswrapper[4632]: I0313 10:08:15.637484 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:08:15 crc kubenswrapper[4632]: I0313 10:08:15.637605 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:08:15 crc kubenswrapper[4632]: I0313 10:08:15.637833 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"dc3a73e428d40f73e3034fb8b7d18fcfe7453c6673209f7b70847a9a508f90d4"} pod="openshift-console/downloads-7954f5f757-w2hhj" containerMessage="Container download-server failed liveness probe, will be restarted" Mar 13 10:08:15 crc kubenswrapper[4632]: I0313 10:08:15.637886 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" containerID="cri-o://dc3a73e428d40f73e3034fb8b7d18fcfe7453c6673209f7b70847a9a508f90d4" gracePeriod=2 Mar 13 10:08:15 crc kubenswrapper[4632]: I0313 10:08:15.982742 4632 patch_prober.go:28] interesting pod/console-f9d7485db-zn7mn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Mar 13 10:08:15 crc kubenswrapper[4632]: I0313 10:08:15.983404 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zn7mn" podUID="f5a50074-5531-442f-a0e9-0578f15634c1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" Mar 13 10:08:16 crc kubenswrapper[4632]: I0313 10:08:16.197375 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:16 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:16 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:16 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:16 crc kubenswrapper[4632]: I0313 10:08:16.197468 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:17 crc kubenswrapper[4632]: I0313 10:08:17.193585 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:17 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:17 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:17 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:17 crc kubenswrapper[4632]: I0313 10:08:17.193683 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:18 crc kubenswrapper[4632]: I0313 10:08:18.195540 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:18 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:18 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:18 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:18 crc kubenswrapper[4632]: I0313 10:08:18.195648 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:18 crc kubenswrapper[4632]: I0313 10:08:18.374921 4632 generic.go:334] "Generic (PLEG): container finished" podID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerID="dc3a73e428d40f73e3034fb8b7d18fcfe7453c6673209f7b70847a9a508f90d4" exitCode=0 Mar 13 10:08:18 crc kubenswrapper[4632]: I0313 10:08:18.375021 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-w2hhj" event={"ID":"7d155f24-9bfc-4039-9981-10e7f724fa51","Type":"ContainerDied","Data":"dc3a73e428d40f73e3034fb8b7d18fcfe7453c6673209f7b70847a9a508f90d4"} Mar 13 10:08:19 crc kubenswrapper[4632]: I0313 10:08:19.060082 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-658cc96bdc-92bpr"] Mar 13 10:08:19 crc kubenswrapper[4632]: I0313 10:08:19.060651 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" podUID="c8ffa9c9-d11d-46b5-ac51-6d38a8639d98" containerName="controller-manager" containerID="cri-o://fe1e770193ae7e14a37be92defae5c64d043b458e01244272a22574e7b2e1f74" gracePeriod=30 Mar 13 10:08:19 crc kubenswrapper[4632]: I0313 10:08:19.098895 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp"] Mar 13 10:08:19 crc kubenswrapper[4632]: I0313 10:08:19.099397 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" podUID="f0f45c9a-e32c-420e-9106-fcb72dd59350" containerName="route-controller-manager" containerID="cri-o://687f3a3904ba4bdede4a24f019fead48648120a9d8e838727216b0ec43fcb3a2" gracePeriod=30 Mar 13 10:08:19 crc kubenswrapper[4632]: I0313 10:08:19.193702 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:19 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:19 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:19 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:19 crc kubenswrapper[4632]: I0313 10:08:19.193762 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:20 crc kubenswrapper[4632]: I0313 10:08:20.193688 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:20 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:20 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:20 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:20 crc kubenswrapper[4632]: I0313 10:08:20.193740 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:21 crc kubenswrapper[4632]: I0313 10:08:21.192851 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:21 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:21 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:21 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:21 crc kubenswrapper[4632]: I0313 10:08:21.192898 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:21 crc kubenswrapper[4632]: I0313 10:08:21.450156 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:08:22 crc kubenswrapper[4632]: I0313 10:08:22.193896 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:22 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:22 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:22 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:22 crc kubenswrapper[4632]: I0313 10:08:22.194222 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:22 crc kubenswrapper[4632]: I0313 10:08:22.469852 4632 generic.go:334] "Generic (PLEG): container finished" podID="f0f45c9a-e32c-420e-9106-fcb72dd59350" containerID="687f3a3904ba4bdede4a24f019fead48648120a9d8e838727216b0ec43fcb3a2" exitCode=0 Mar 13 10:08:22 crc kubenswrapper[4632]: I0313 10:08:22.469971 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" event={"ID":"f0f45c9a-e32c-420e-9106-fcb72dd59350","Type":"ContainerDied","Data":"687f3a3904ba4bdede4a24f019fead48648120a9d8e838727216b0ec43fcb3a2"} Mar 13 10:08:22 crc kubenswrapper[4632]: I0313 10:08:22.475915 4632 generic.go:334] "Generic (PLEG): container finished" podID="c8ffa9c9-d11d-46b5-ac51-6d38a8639d98" containerID="fe1e770193ae7e14a37be92defae5c64d043b458e01244272a22574e7b2e1f74" exitCode=0 Mar 13 10:08:22 crc kubenswrapper[4632]: I0313 10:08:22.476004 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" event={"ID":"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98","Type":"ContainerDied","Data":"fe1e770193ae7e14a37be92defae5c64d043b458e01244272a22574e7b2e1f74"} Mar 13 10:08:23 crc kubenswrapper[4632]: I0313 10:08:23.194372 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:23 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:23 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:23 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:23 crc kubenswrapper[4632]: I0313 10:08:23.194495 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:23 crc kubenswrapper[4632]: I0313 10:08:23.195140 4632 patch_prober.go:28] interesting pod/controller-manager-658cc96bdc-92bpr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: connect: connection refused" start-of-body= Mar 13 10:08:23 crc kubenswrapper[4632]: I0313 10:08:23.195261 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" podUID="c8ffa9c9-d11d-46b5-ac51-6d38a8639d98" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: connect: connection refused" Mar 13 10:08:23 crc kubenswrapper[4632]: I0313 10:08:23.229162 4632 patch_prober.go:28] interesting pod/route-controller-manager-86d66b8bfb-9wzpp container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" start-of-body= Mar 13 10:08:23 crc kubenswrapper[4632]: I0313 10:08:23.229308 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" podUID="f0f45c9a-e32c-420e-9106-fcb72dd59350" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": dial tcp 10.217.0.55:8443: connect: connection refused" Mar 13 10:08:24 crc kubenswrapper[4632]: I0313 10:08:24.193580 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:08:24 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 10:08:24 crc kubenswrapper[4632]: [+]process-running ok Mar 13 10:08:24 crc kubenswrapper[4632]: healthz check failed Mar 13 10:08:24 crc kubenswrapper[4632]: I0313 10:08:24.193676 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:08:25 crc kubenswrapper[4632]: I0313 10:08:25.199803 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:08:25 crc kubenswrapper[4632]: I0313 10:08:25.205716 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 10:08:25 crc kubenswrapper[4632]: I0313 10:08:25.632001 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:08:25 crc kubenswrapper[4632]: I0313 10:08:25.632054 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:08:26 crc kubenswrapper[4632]: I0313 10:08:25.983328 4632 patch_prober.go:28] interesting pod/console-f9d7485db-zn7mn container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" start-of-body= Mar 13 10:08:26 crc kubenswrapper[4632]: I0313 10:08:25.983703 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-zn7mn" podUID="f5a50074-5531-442f-a0e9-0578f15634c1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.22:8443/health\": dial tcp 10.217.0.22:8443: connect: connection refused" Mar 13 10:08:26 crc kubenswrapper[4632]: I0313 10:08:26.867154 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.890904 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.901346 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.941526 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb"] Mar 13 10:08:33 crc kubenswrapper[4632]: E0313 10:08:33.941876 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa472cd7-9575-473f-b6b6-709b644a5ec4" containerName="pruner" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.941894 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa472cd7-9575-473f-b6b6-709b644a5ec4" containerName="pruner" Mar 13 10:08:33 crc kubenswrapper[4632]: E0313 10:08:33.941924 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8ffa9c9-d11d-46b5-ac51-6d38a8639d98" containerName="controller-manager" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.941935 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8ffa9c9-d11d-46b5-ac51-6d38a8639d98" containerName="controller-manager" Mar 13 10:08:33 crc kubenswrapper[4632]: E0313 10:08:33.941964 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="528d3aa9-10bf-4029-a4d2-85768264fde8" containerName="collect-profiles" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.941972 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="528d3aa9-10bf-4029-a4d2-85768264fde8" containerName="collect-profiles" Mar 13 10:08:33 crc kubenswrapper[4632]: E0313 10:08:33.941988 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a4d2988-a460-407b-902a-aeb8eda619a1" containerName="pruner" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.941998 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a4d2988-a460-407b-902a-aeb8eda619a1" containerName="pruner" Mar 13 10:08:33 crc kubenswrapper[4632]: E0313 10:08:33.942008 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0f45c9a-e32c-420e-9106-fcb72dd59350" containerName="route-controller-manager" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.942015 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0f45c9a-e32c-420e-9106-fcb72dd59350" containerName="route-controller-manager" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.942146 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8ffa9c9-d11d-46b5-ac51-6d38a8639d98" containerName="controller-manager" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.942173 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="528d3aa9-10bf-4029-a4d2-85768264fde8" containerName="collect-profiles" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.942191 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa472cd7-9575-473f-b6b6-709b644a5ec4" containerName="pruner" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.942204 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0f45c9a-e32c-420e-9106-fcb72dd59350" containerName="route-controller-manager" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.942213 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a4d2988-a460-407b-902a-aeb8eda619a1" containerName="pruner" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.942736 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:33 crc kubenswrapper[4632]: I0313 10:08:33.953421 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb"] Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.084091 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0f45c9a-e32c-420e-9106-fcb72dd59350-serving-cert\") pod \"f0f45c9a-e32c-420e-9106-fcb72dd59350\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.084433 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-client-ca\") pod \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.084556 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9pxp\" (UniqueName: \"kubernetes.io/projected/f0f45c9a-e32c-420e-9106-fcb72dd59350-kube-api-access-l9pxp\") pod \"f0f45c9a-e32c-420e-9106-fcb72dd59350\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.084659 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-serving-cert\") pod \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.084772 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwf9m\" (UniqueName: \"kubernetes.io/projected/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-kube-api-access-hwf9m\") pod \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.084885 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-config\") pod \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.085156 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0f45c9a-e32c-420e-9106-fcb72dd59350-config\") pod \"f0f45c9a-e32c-420e-9106-fcb72dd59350\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.085318 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-proxy-ca-bundles\") pod \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\" (UID: \"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98\") " Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.085441 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f0f45c9a-e32c-420e-9106-fcb72dd59350-client-ca\") pod \"f0f45c9a-e32c-420e-9106-fcb72dd59350\" (UID: \"f0f45c9a-e32c-420e-9106-fcb72dd59350\") " Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.085667 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0982dbd-62a1-47c5-8510-5045b9ca5785-client-ca\") pod \"route-controller-manager-5c44777cb6-dkmdb\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.085862 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0f45c9a-e32c-420e-9106-fcb72dd59350-config" (OuterVolumeSpecName: "config") pod "f0f45c9a-e32c-420e-9106-fcb72dd59350" (UID: "f0f45c9a-e32c-420e-9106-fcb72dd59350"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.085896 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-config" (OuterVolumeSpecName: "config") pod "c8ffa9c9-d11d-46b5-ac51-6d38a8639d98" (UID: "c8ffa9c9-d11d-46b5-ac51-6d38a8639d98"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.086435 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c8ffa9c9-d11d-46b5-ac51-6d38a8639d98" (UID: "c8ffa9c9-d11d-46b5-ac51-6d38a8639d98"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.086661 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0f45c9a-e32c-420e-9106-fcb72dd59350-client-ca" (OuterVolumeSpecName: "client-ca") pod "f0f45c9a-e32c-420e-9106-fcb72dd59350" (UID: "f0f45c9a-e32c-420e-9106-fcb72dd59350"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.087161 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-client-ca" (OuterVolumeSpecName: "client-ca") pod "c8ffa9c9-d11d-46b5-ac51-6d38a8639d98" (UID: "c8ffa9c9-d11d-46b5-ac51-6d38a8639d98"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.087614 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0982dbd-62a1-47c5-8510-5045b9ca5785-config\") pod \"route-controller-manager-5c44777cb6-dkmdb\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.088065 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0982dbd-62a1-47c5-8510-5045b9ca5785-serving-cert\") pod \"route-controller-manager-5c44777cb6-dkmdb\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.088214 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cqqg\" (UniqueName: \"kubernetes.io/projected/a0982dbd-62a1-47c5-8510-5045b9ca5785-kube-api-access-2cqqg\") pod \"route-controller-manager-5c44777cb6-dkmdb\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.088392 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.088479 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0f45c9a-e32c-420e-9106-fcb72dd59350-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.088564 4632 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.088643 4632 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f0f45c9a-e32c-420e-9106-fcb72dd59350-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.088763 4632 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.089906 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c8ffa9c9-d11d-46b5-ac51-6d38a8639d98" (UID: "c8ffa9c9-d11d-46b5-ac51-6d38a8639d98"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.090769 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0f45c9a-e32c-420e-9106-fcb72dd59350-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f0f45c9a-e32c-420e-9106-fcb72dd59350" (UID: "f0f45c9a-e32c-420e-9106-fcb72dd59350"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.091161 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0f45c9a-e32c-420e-9106-fcb72dd59350-kube-api-access-l9pxp" (OuterVolumeSpecName: "kube-api-access-l9pxp") pod "f0f45c9a-e32c-420e-9106-fcb72dd59350" (UID: "f0f45c9a-e32c-420e-9106-fcb72dd59350"). InnerVolumeSpecName "kube-api-access-l9pxp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.114812 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-kube-api-access-hwf9m" (OuterVolumeSpecName: "kube-api-access-hwf9m") pod "c8ffa9c9-d11d-46b5-ac51-6d38a8639d98" (UID: "c8ffa9c9-d11d-46b5-ac51-6d38a8639d98"). InnerVolumeSpecName "kube-api-access-hwf9m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.189235 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cqqg\" (UniqueName: \"kubernetes.io/projected/a0982dbd-62a1-47c5-8510-5045b9ca5785-kube-api-access-2cqqg\") pod \"route-controller-manager-5c44777cb6-dkmdb\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.189332 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0982dbd-62a1-47c5-8510-5045b9ca5785-client-ca\") pod \"route-controller-manager-5c44777cb6-dkmdb\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.189371 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0982dbd-62a1-47c5-8510-5045b9ca5785-config\") pod \"route-controller-manager-5c44777cb6-dkmdb\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.189395 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0982dbd-62a1-47c5-8510-5045b9ca5785-serving-cert\") pod \"route-controller-manager-5c44777cb6-dkmdb\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.189428 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f0f45c9a-e32c-420e-9106-fcb72dd59350-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.189439 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9pxp\" (UniqueName: \"kubernetes.io/projected/f0f45c9a-e32c-420e-9106-fcb72dd59350-kube-api-access-l9pxp\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.189449 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.189458 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwf9m\" (UniqueName: \"kubernetes.io/projected/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98-kube-api-access-hwf9m\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.191212 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0982dbd-62a1-47c5-8510-5045b9ca5785-config\") pod \"route-controller-manager-5c44777cb6-dkmdb\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.194922 4632 patch_prober.go:28] interesting pod/controller-manager-658cc96bdc-92bpr container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: i/o timeout" start-of-body= Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.194999 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" podUID="c8ffa9c9-d11d-46b5-ac51-6d38a8639d98" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.54:8443/healthz\": dial tcp 10.217.0.54:8443: i/o timeout" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.203212 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0982dbd-62a1-47c5-8510-5045b9ca5785-serving-cert\") pod \"route-controller-manager-5c44777cb6-dkmdb\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.203858 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0982dbd-62a1-47c5-8510-5045b9ca5785-client-ca\") pod \"route-controller-manager-5c44777cb6-dkmdb\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.209438 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cqqg\" (UniqueName: \"kubernetes.io/projected/a0982dbd-62a1-47c5-8510-5045b9ca5785-kube-api-access-2cqqg\") pod \"route-controller-manager-5c44777cb6-dkmdb\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.228452 4632 patch_prober.go:28] interesting pod/route-controller-manager-86d66b8bfb-9wzpp container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.55:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.228527 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" podUID="f0f45c9a-e32c-420e-9106-fcb72dd59350" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.55:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.274596 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.583270 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.583528 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp" event={"ID":"f0f45c9a-e32c-420e-9106-fcb72dd59350","Type":"ContainerDied","Data":"070e0df69129dffbb7be6eacfbcf901c3d12a23e3166a3f87eca493eaf3d27d9"} Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.583570 4632 scope.go:117] "RemoveContainer" containerID="687f3a3904ba4bdede4a24f019fead48648120a9d8e838727216b0ec43fcb3a2" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.588379 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" event={"ID":"c8ffa9c9-d11d-46b5-ac51-6d38a8639d98","Type":"ContainerDied","Data":"94bf0543076d266034298c32f3e17bb1e05e21aec66de3f477e64011186c779f"} Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.588444 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-658cc96bdc-92bpr" Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.628134 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-658cc96bdc-92bpr"] Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.635712 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-658cc96bdc-92bpr"] Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.641396 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp"] Mar 13 10:08:34 crc kubenswrapper[4632]: I0313 10:08:34.644026 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-86d66b8bfb-9wzpp"] Mar 13 10:08:35 crc kubenswrapper[4632]: I0313 10:08:35.632905 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:08:35 crc kubenswrapper[4632]: I0313 10:08:35.633227 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:08:35 crc kubenswrapper[4632]: E0313 10:08:35.636127 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 13 10:08:35 crc kubenswrapper[4632]: E0313 10:08:35.636293 4632 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 10:08:35 crc kubenswrapper[4632]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 13 10:08:35 crc kubenswrapper[4632]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-sd756,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29556606-mkrp2_openshift-infra(c822257d-9d2f-4b6f-87de-131de5cd0efe): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 13 10:08:35 crc kubenswrapper[4632]: > logger="UnhandledError" Mar 13 10:08:35 crc kubenswrapper[4632]: E0313 10:08:35.637653 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29556606-mkrp2" podUID="c822257d-9d2f-4b6f-87de-131de5cd0efe" Mar 13 10:08:35 crc kubenswrapper[4632]: I0313 10:08:35.994259 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:08:35 crc kubenswrapper[4632]: I0313 10:08:35.998146 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.056963 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c8ffa9c9-d11d-46b5-ac51-6d38a8639d98" path="/var/lib/kubelet/pods/c8ffa9c9-d11d-46b5-ac51-6d38a8639d98/volumes" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.057994 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0f45c9a-e32c-420e-9106-fcb72dd59350" path="/var/lib/kubelet/pods/f0f45c9a-e32c-420e-9106-fcb72dd59350/volumes" Mar 13 10:08:36 crc kubenswrapper[4632]: E0313 10:08:36.607679 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29556606-mkrp2" podUID="c822257d-9d2f-4b6f-87de-131de5cd0efe" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.865426 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-747c9765b-lqcx6"] Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.866473 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.874257 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.881716 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.883912 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.883985 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.885909 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.889629 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.892680 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-747c9765b-lqcx6"] Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.897483 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.937349 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-config\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.937468 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-proxy-ca-bundles\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.937501 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2prq4\" (UniqueName: \"kubernetes.io/projected/0816f595-2f7f-425a-9a6b-1022e2a4ca04-kube-api-access-2prq4\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.937577 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-client-ca\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:36 crc kubenswrapper[4632]: I0313 10:08:36.937630 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0816f595-2f7f-425a-9a6b-1022e2a4ca04-serving-cert\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:37 crc kubenswrapper[4632]: I0313 10:08:37.039606 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-config\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:37 crc kubenswrapper[4632]: I0313 10:08:37.039736 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-proxy-ca-bundles\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:37 crc kubenswrapper[4632]: I0313 10:08:37.039776 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2prq4\" (UniqueName: \"kubernetes.io/projected/0816f595-2f7f-425a-9a6b-1022e2a4ca04-kube-api-access-2prq4\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:37 crc kubenswrapper[4632]: I0313 10:08:37.039822 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-client-ca\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:37 crc kubenswrapper[4632]: I0313 10:08:37.039859 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0816f595-2f7f-425a-9a6b-1022e2a4ca04-serving-cert\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:37 crc kubenswrapper[4632]: I0313 10:08:37.048602 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-client-ca\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:37 crc kubenswrapper[4632]: I0313 10:08:37.050594 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-proxy-ca-bundles\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:37 crc kubenswrapper[4632]: I0313 10:08:37.052229 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-config\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:37 crc kubenswrapper[4632]: I0313 10:08:37.077073 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0816f595-2f7f-425a-9a6b-1022e2a4ca04-serving-cert\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:37 crc kubenswrapper[4632]: I0313 10:08:37.085594 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2prq4\" (UniqueName: \"kubernetes.io/projected/0816f595-2f7f-425a-9a6b-1022e2a4ca04-kube-api-access-2prq4\") pod \"controller-manager-747c9765b-lqcx6\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:37 crc kubenswrapper[4632]: I0313 10:08:37.190570 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:08:39 crc kubenswrapper[4632]: I0313 10:08:39.023879 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-747c9765b-lqcx6"] Mar 13 10:08:39 crc kubenswrapper[4632]: I0313 10:08:39.059078 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb"] Mar 13 10:08:39 crc kubenswrapper[4632]: I0313 10:08:39.404555 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 10:08:39 crc kubenswrapper[4632]: I0313 10:08:39.406081 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 10:08:39 crc kubenswrapper[4632]: I0313 10:08:39.410991 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 13 10:08:39 crc kubenswrapper[4632]: I0313 10:08:39.412409 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 13 10:08:39 crc kubenswrapper[4632]: I0313 10:08:39.415447 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 10:08:39 crc kubenswrapper[4632]: I0313 10:08:39.554003 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a93bf50-5608-4b34-aea5-2f027d469fe7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7a93bf50-5608-4b34-aea5-2f027d469fe7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 10:08:39 crc kubenswrapper[4632]: I0313 10:08:39.554312 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a93bf50-5608-4b34-aea5-2f027d469fe7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7a93bf50-5608-4b34-aea5-2f027d469fe7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 10:08:39 crc kubenswrapper[4632]: I0313 10:08:39.654802 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a93bf50-5608-4b34-aea5-2f027d469fe7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7a93bf50-5608-4b34-aea5-2f027d469fe7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 10:08:39 crc kubenswrapper[4632]: I0313 10:08:39.654977 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a93bf50-5608-4b34-aea5-2f027d469fe7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7a93bf50-5608-4b34-aea5-2f027d469fe7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 10:08:39 crc kubenswrapper[4632]: I0313 10:08:39.654987 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a93bf50-5608-4b34-aea5-2f027d469fe7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7a93bf50-5608-4b34-aea5-2f027d469fe7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 10:08:39 crc kubenswrapper[4632]: I0313 10:08:39.678073 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a93bf50-5608-4b34-aea5-2f027d469fe7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7a93bf50-5608-4b34-aea5-2f027d469fe7\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 10:08:39 crc kubenswrapper[4632]: I0313 10:08:39.729826 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 10:08:40 crc kubenswrapper[4632]: I0313 10:08:40.461152 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:08:40 crc kubenswrapper[4632]: I0313 10:08:40.461227 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:08:40 crc kubenswrapper[4632]: I0313 10:08:40.461286 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:08:40 crc kubenswrapper[4632]: I0313 10:08:40.461998 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:08:40 crc kubenswrapper[4632]: I0313 10:08:40.462058 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7" gracePeriod=600 Mar 13 10:08:40 crc kubenswrapper[4632]: I0313 10:08:40.646339 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7" exitCode=0 Mar 13 10:08:40 crc kubenswrapper[4632]: I0313 10:08:40.646680 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7"} Mar 13 10:08:43 crc kubenswrapper[4632]: E0313 10:08:43.287443 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 13 10:08:43 crc kubenswrapper[4632]: E0313 10:08:43.287742 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kh6g9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-xd455_openshift-marketplace(cd6e3c73-fbc1-4213-bbef-02dd2b0587f8): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 10:08:43 crc kubenswrapper[4632]: E0313 10:08:43.288999 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-xd455" podUID="cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" Mar 13 10:08:43 crc kubenswrapper[4632]: E0313 10:08:43.325135 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 13 10:08:43 crc kubenswrapper[4632]: E0313 10:08:43.325393 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5kzjw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-jvh86_openshift-marketplace(bd46ae04-0610-4aa5-9385-dd45de66c5dd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 10:08:43 crc kubenswrapper[4632]: E0313 10:08:43.328513 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-jvh86" podUID="bd46ae04-0610-4aa5-9385-dd45de66c5dd" Mar 13 10:08:43 crc kubenswrapper[4632]: I0313 10:08:43.610116 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 10:08:43 crc kubenswrapper[4632]: I0313 10:08:43.611073 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 10:08:43 crc kubenswrapper[4632]: I0313 10:08:43.623980 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 10:08:43 crc kubenswrapper[4632]: I0313 10:08:43.651286 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc39d207-84a2-4a28-9296-bed684aa308d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"dc39d207-84a2-4a28-9296-bed684aa308d\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 10:08:43 crc kubenswrapper[4632]: I0313 10:08:43.651343 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dc39d207-84a2-4a28-9296-bed684aa308d-var-lock\") pod \"installer-9-crc\" (UID: \"dc39d207-84a2-4a28-9296-bed684aa308d\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 10:08:43 crc kubenswrapper[4632]: I0313 10:08:43.651391 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc39d207-84a2-4a28-9296-bed684aa308d-kube-api-access\") pod \"installer-9-crc\" (UID: \"dc39d207-84a2-4a28-9296-bed684aa308d\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 10:08:43 crc kubenswrapper[4632]: I0313 10:08:43.752832 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc39d207-84a2-4a28-9296-bed684aa308d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"dc39d207-84a2-4a28-9296-bed684aa308d\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 10:08:43 crc kubenswrapper[4632]: I0313 10:08:43.752887 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dc39d207-84a2-4a28-9296-bed684aa308d-var-lock\") pod \"installer-9-crc\" (UID: \"dc39d207-84a2-4a28-9296-bed684aa308d\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 10:08:43 crc kubenswrapper[4632]: I0313 10:08:43.752969 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc39d207-84a2-4a28-9296-bed684aa308d-kube-api-access\") pod \"installer-9-crc\" (UID: \"dc39d207-84a2-4a28-9296-bed684aa308d\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 10:08:43 crc kubenswrapper[4632]: I0313 10:08:43.753008 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc39d207-84a2-4a28-9296-bed684aa308d-kubelet-dir\") pod \"installer-9-crc\" (UID: \"dc39d207-84a2-4a28-9296-bed684aa308d\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 10:08:43 crc kubenswrapper[4632]: I0313 10:08:43.753092 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dc39d207-84a2-4a28-9296-bed684aa308d-var-lock\") pod \"installer-9-crc\" (UID: \"dc39d207-84a2-4a28-9296-bed684aa308d\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 10:08:43 crc kubenswrapper[4632]: I0313 10:08:43.772768 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc39d207-84a2-4a28-9296-bed684aa308d-kube-api-access\") pod \"installer-9-crc\" (UID: \"dc39d207-84a2-4a28-9296-bed684aa308d\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 10:08:43 crc kubenswrapper[4632]: I0313 10:08:43.941692 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 10:08:45 crc kubenswrapper[4632]: I0313 10:08:45.633189 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:08:45 crc kubenswrapper[4632]: I0313 10:08:45.633244 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:08:48 crc kubenswrapper[4632]: E0313 10:08:48.663136 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 13 10:08:48 crc kubenswrapper[4632]: E0313 10:08:48.664000 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6hfz2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-xr5l9_openshift-marketplace(87965e39-b879-4e26-9c8b-b78068c52aa0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 10:08:48 crc kubenswrapper[4632]: E0313 10:08:48.665191 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-xr5l9" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" Mar 13 10:08:52 crc kubenswrapper[4632]: E0313 10:08:52.258780 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 13 10:08:52 crc kubenswrapper[4632]: E0313 10:08:52.260272 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wlsv8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-p8wjg_openshift-marketplace(b11a7dff-bf08-44c3-b4f4-923119c13717): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 10:08:52 crc kubenswrapper[4632]: E0313 10:08:52.263281 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-p8wjg" podUID="b11a7dff-bf08-44c3-b4f4-923119c13717" Mar 13 10:08:52 crc kubenswrapper[4632]: E0313 10:08:52.320162 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 13 10:08:52 crc kubenswrapper[4632]: E0313 10:08:52.320337 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tfdk8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-z2gc7_openshift-marketplace(a110c276-8516-4f9e-a6af-d6837cd0f387): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 10:08:52 crc kubenswrapper[4632]: E0313 10:08:52.321631 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-z2gc7" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" Mar 13 10:08:52 crc kubenswrapper[4632]: E0313 10:08:52.508207 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 13 10:08:52 crc kubenswrapper[4632]: E0313 10:08:52.508332 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sdp56,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-8z668_openshift-marketplace(9845f384-2720-4d6a-aa73-1e66e30f7c2c): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 10:08:52 crc kubenswrapper[4632]: E0313 10:08:52.509950 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-8z668" podUID="9845f384-2720-4d6a-aa73-1e66e30f7c2c" Mar 13 10:08:54 crc kubenswrapper[4632]: I0313 10:08:54.813880 4632 ???:1] "http: TLS handshake error from 192.168.126.11:40032: no serving certificate available for the kubelet" Mar 13 10:08:55 crc kubenswrapper[4632]: I0313 10:08:55.632439 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:08:55 crc kubenswrapper[4632]: I0313 10:08:55.632551 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:08:56 crc kubenswrapper[4632]: I0313 10:08:56.254608 4632 scope.go:117] "RemoveContainer" containerID="fe1e770193ae7e14a37be92defae5c64d043b458e01244272a22574e7b2e1f74" Mar 13 10:08:56 crc kubenswrapper[4632]: E0313 10:08:56.275960 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-8z668" podUID="9845f384-2720-4d6a-aa73-1e66e30f7c2c" Mar 13 10:08:56 crc kubenswrapper[4632]: E0313 10:08:56.276047 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-xr5l9" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" Mar 13 10:08:56 crc kubenswrapper[4632]: E0313 10:08:56.276110 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-p8wjg" podUID="b11a7dff-bf08-44c3-b4f4-923119c13717" Mar 13 10:08:56 crc kubenswrapper[4632]: E0313 10:08:56.276162 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-z2gc7" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" Mar 13 10:08:56 crc kubenswrapper[4632]: I0313 10:08:56.796630 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 10:08:56 crc kubenswrapper[4632]: I0313 10:08:56.799379 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb"] Mar 13 10:08:56 crc kubenswrapper[4632]: W0313 10:08:56.806678 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda0982dbd_62a1_47c5_8510_5045b9ca5785.slice/crio-e2910cae586a594472f36055ddd013b1c782ed1e61c6b7e0b1e88de17b0d81e9 WatchSource:0}: Error finding container e2910cae586a594472f36055ddd013b1c782ed1e61c6b7e0b1e88de17b0d81e9: Status 404 returned error can't find the container with id e2910cae586a594472f36055ddd013b1c782ed1e61c6b7e0b1e88de17b0d81e9 Mar 13 10:08:56 crc kubenswrapper[4632]: W0313 10:08:56.812414 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7a93bf50_5608_4b34_aea5_2f027d469fe7.slice/crio-8ca4d20a3432f315af224e18c181a59d62f280063326feb2da6b42b782268a63 WatchSource:0}: Error finding container 8ca4d20a3432f315af224e18c181a59d62f280063326feb2da6b42b782268a63: Status 404 returned error can't find the container with id 8ca4d20a3432f315af224e18c181a59d62f280063326feb2da6b42b782268a63 Mar 13 10:08:56 crc kubenswrapper[4632]: I0313 10:08:56.877790 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 10:08:56 crc kubenswrapper[4632]: I0313 10:08:56.883762 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-747c9765b-lqcx6"] Mar 13 10:08:57 crc kubenswrapper[4632]: I0313 10:08:57.758176 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"dc39d207-84a2-4a28-9296-bed684aa308d","Type":"ContainerStarted","Data":"c463b62ee1a6928ceb028fe480183c5ca7bb846ec47d4163fa232376e05db524"} Mar 13 10:08:57 crc kubenswrapper[4632]: I0313 10:08:57.758850 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"dc39d207-84a2-4a28-9296-bed684aa308d","Type":"ContainerStarted","Data":"7017997794d887d37a83222e44995b72d5d076c42028b8e1498fdb1f2cb4d188"} Mar 13 10:08:57 crc kubenswrapper[4632]: I0313 10:08:57.762521 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-w2hhj" event={"ID":"7d155f24-9bfc-4039-9981-10e7f724fa51","Type":"ContainerStarted","Data":"8267a2aa0ccf8bef0a1fb4ed1acbdc94b5e3909c757e1794b32f16ebf1d938e7"} Mar 13 10:08:57 crc kubenswrapper[4632]: I0313 10:08:57.762590 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-w2hhj" Mar 13 10:08:57 crc kubenswrapper[4632]: I0313 10:08:57.763339 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:08:57 crc kubenswrapper[4632]: I0313 10:08:57.763390 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:08:57 crc kubenswrapper[4632]: I0313 10:08:57.763855 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7a93bf50-5608-4b34-aea5-2f027d469fe7","Type":"ContainerStarted","Data":"9397ca1f1b655dbac921f1df09ecbfaa16d86c267f0a805db268730c6e1431c8"} Mar 13 10:08:57 crc kubenswrapper[4632]: I0313 10:08:57.763908 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7a93bf50-5608-4b34-aea5-2f027d469fe7","Type":"ContainerStarted","Data":"8ca4d20a3432f315af224e18c181a59d62f280063326feb2da6b42b782268a63"} Mar 13 10:08:57 crc kubenswrapper[4632]: I0313 10:08:57.767243 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556608-9kzfk" event={"ID":"37ab6711-478f-4cc7-b9a4-c9baa126b1a3","Type":"ContainerStarted","Data":"24d957ae4862987ed76c21db8796ae914a7d2beca83397bc3f90816dc051c956"} Mar 13 10:08:57 crc kubenswrapper[4632]: I0313 10:08:57.772190 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" event={"ID":"a0982dbd-62a1-47c5-8510-5045b9ca5785","Type":"ContainerStarted","Data":"a4aeba6501ad544549300fc2c4204c4a5cf7f1d1edc84405ce7d7b1974c966cc"} Mar 13 10:08:57 crc kubenswrapper[4632]: I0313 10:08:57.772225 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" event={"ID":"a0982dbd-62a1-47c5-8510-5045b9ca5785","Type":"ContainerStarted","Data":"e2910cae586a594472f36055ddd013b1c782ed1e61c6b7e0b1e88de17b0d81e9"} Mar 13 10:08:57 crc kubenswrapper[4632]: I0313 10:08:57.774464 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" event={"ID":"0816f595-2f7f-425a-9a6b-1022e2a4ca04","Type":"ContainerStarted","Data":"83de0881072cb52ab7a7fbd2d8ef18cbb3eb4eb7897fd1301bfd2cbf304913b7"} Mar 13 10:08:57 crc kubenswrapper[4632]: I0313 10:08:57.774584 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" event={"ID":"0816f595-2f7f-425a-9a6b-1022e2a4ca04","Type":"ContainerStarted","Data":"650730d84aa4384d18f0228647070b36709f320f05d4d8ef5c14d5e680d6b8ca"} Mar 13 10:08:57 crc kubenswrapper[4632]: I0313 10:08:57.779423 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"e4989d70178427347867288c3fc7b62a339fa6ecdddde954f719a53f3db7fe17"} Mar 13 10:08:58 crc kubenswrapper[4632]: E0313 10:08:58.284074 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 13 10:08:58 crc kubenswrapper[4632]: E0313 10:08:58.284409 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z5jx8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-txp2w_openshift-marketplace(f0cd0b7e-eded-4a51-8b1e-e67b9381bc87): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 10:08:58 crc kubenswrapper[4632]: E0313 10:08:58.285504 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-txp2w" podUID="f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" Mar 13 10:08:58 crc kubenswrapper[4632]: I0313 10:08:58.791393 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" podUID="a0982dbd-62a1-47c5-8510-5045b9ca5785" containerName="route-controller-manager" containerID="cri-o://a4aeba6501ad544549300fc2c4204c4a5cf7f1d1edc84405ce7d7b1974c966cc" gracePeriod=30 Mar 13 10:08:58 crc kubenswrapper[4632]: I0313 10:08:58.792517 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:08:58 crc kubenswrapper[4632]: I0313 10:08:58.792571 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:08:58 crc kubenswrapper[4632]: I0313 10:08:58.792729 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:58 crc kubenswrapper[4632]: I0313 10:08:58.793784 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" podUID="0816f595-2f7f-425a-9a6b-1022e2a4ca04" containerName="controller-manager" containerID="cri-o://83de0881072cb52ab7a7fbd2d8ef18cbb3eb4eb7897fd1301bfd2cbf304913b7" gracePeriod=30 Mar 13 10:08:58 crc kubenswrapper[4632]: E0313 10:08:58.798126 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-txp2w" podUID="f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" Mar 13 10:08:58 crc kubenswrapper[4632]: I0313 10:08:58.807376 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:08:58 crc kubenswrapper[4632]: I0313 10:08:58.859063 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" podStartSLOduration=39.859041263 podStartE2EDuration="39.859041263s" podCreationTimestamp="2026-03-13 10:08:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:08:58.826250362 +0000 UTC m=+312.848780495" watchObservedRunningTime="2026-03-13 10:08:58.859041263 +0000 UTC m=+312.881571396" Mar 13 10:08:58 crc kubenswrapper[4632]: I0313 10:08:58.892651 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=15.892632546 podStartE2EDuration="15.892632546s" podCreationTimestamp="2026-03-13 10:08:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:08:58.86364358 +0000 UTC m=+312.886173713" watchObservedRunningTime="2026-03-13 10:08:58.892632546 +0000 UTC m=+312.915162679" Mar 13 10:08:58 crc kubenswrapper[4632]: I0313 10:08:58.910736 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=19.910718974 podStartE2EDuration="19.910718974s" podCreationTimestamp="2026-03-13 10:08:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:08:58.907849035 +0000 UTC m=+312.930379168" watchObservedRunningTime="2026-03-13 10:08:58.910718974 +0000 UTC m=+312.933249107" Mar 13 10:08:58 crc kubenswrapper[4632]: I0313 10:08:58.938399 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556608-9kzfk" podStartSLOduration=4.828758615 podStartE2EDuration="58.938370624s" podCreationTimestamp="2026-03-13 10:08:00 +0000 UTC" firstStartedPulling="2026-03-13 10:08:02.171472122 +0000 UTC m=+256.194002255" lastFinishedPulling="2026-03-13 10:08:56.281084131 +0000 UTC m=+310.303614264" observedRunningTime="2026-03-13 10:08:58.936844462 +0000 UTC m=+312.959374595" watchObservedRunningTime="2026-03-13 10:08:58.938370624 +0000 UTC m=+312.960900777" Mar 13 10:08:58 crc kubenswrapper[4632]: I0313 10:08:58.968210 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" podStartSLOduration=39.968188593 podStartE2EDuration="39.968188593s" podCreationTimestamp="2026-03-13 10:08:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:08:58.964833421 +0000 UTC m=+312.987363554" watchObservedRunningTime="2026-03-13 10:08:58.968188593 +0000 UTC m=+312.990718736" Mar 13 10:08:59 crc kubenswrapper[4632]: I0313 10:08:59.158660 4632 csr.go:261] certificate signing request csr-drwfx is approved, waiting to be issued Mar 13 10:08:59 crc kubenswrapper[4632]: I0313 10:08:59.178184 4632 csr.go:257] certificate signing request csr-drwfx is issued Mar 13 10:08:59 crc kubenswrapper[4632]: E0313 10:08:59.538092 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 13 10:08:59 crc kubenswrapper[4632]: E0313 10:08:59.538273 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bc6ff,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-t6bkt_openshift-marketplace(668c4640-0e5f-4c98-8b6e-dbdffdbfe14e): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 10:08:59 crc kubenswrapper[4632]: E0313 10:08:59.539471 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-t6bkt" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" Mar 13 10:08:59 crc kubenswrapper[4632]: I0313 10:08:59.824370 4632 generic.go:334] "Generic (PLEG): container finished" podID="7a93bf50-5608-4b34-aea5-2f027d469fe7" containerID="9397ca1f1b655dbac921f1df09ecbfaa16d86c267f0a805db268730c6e1431c8" exitCode=0 Mar 13 10:08:59 crc kubenswrapper[4632]: I0313 10:08:59.824450 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7a93bf50-5608-4b34-aea5-2f027d469fe7","Type":"ContainerDied","Data":"9397ca1f1b655dbac921f1df09ecbfaa16d86c267f0a805db268730c6e1431c8"} Mar 13 10:08:59 crc kubenswrapper[4632]: I0313 10:08:59.829994 4632 generic.go:334] "Generic (PLEG): container finished" podID="37ab6711-478f-4cc7-b9a4-c9baa126b1a3" containerID="24d957ae4862987ed76c21db8796ae914a7d2beca83397bc3f90816dc051c956" exitCode=0 Mar 13 10:08:59 crc kubenswrapper[4632]: I0313 10:08:59.830151 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556608-9kzfk" event={"ID":"37ab6711-478f-4cc7-b9a4-c9baa126b1a3","Type":"ContainerDied","Data":"24d957ae4862987ed76c21db8796ae914a7d2beca83397bc3f90816dc051c956"} Mar 13 10:08:59 crc kubenswrapper[4632]: I0313 10:08:59.849501 4632 generic.go:334] "Generic (PLEG): container finished" podID="0816f595-2f7f-425a-9a6b-1022e2a4ca04" containerID="83de0881072cb52ab7a7fbd2d8ef18cbb3eb4eb7897fd1301bfd2cbf304913b7" exitCode=0 Mar 13 10:08:59 crc kubenswrapper[4632]: I0313 10:08:59.849659 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" event={"ID":"0816f595-2f7f-425a-9a6b-1022e2a4ca04","Type":"ContainerDied","Data":"83de0881072cb52ab7a7fbd2d8ef18cbb3eb4eb7897fd1301bfd2cbf304913b7"} Mar 13 10:08:59 crc kubenswrapper[4632]: I0313 10:08:59.860318 4632 generic.go:334] "Generic (PLEG): container finished" podID="a0982dbd-62a1-47c5-8510-5045b9ca5785" containerID="a4aeba6501ad544549300fc2c4204c4a5cf7f1d1edc84405ce7d7b1974c966cc" exitCode=0 Mar 13 10:08:59 crc kubenswrapper[4632]: I0313 10:08:59.861192 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" event={"ID":"a0982dbd-62a1-47c5-8510-5045b9ca5785","Type":"ContainerDied","Data":"a4aeba6501ad544549300fc2c4204c4a5cf7f1d1edc84405ce7d7b1974c966cc"} Mar 13 10:08:59 crc kubenswrapper[4632]: E0313 10:08:59.865696 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-t6bkt" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" Mar 13 10:08:59 crc kubenswrapper[4632]: I0313 10:08:59.928699 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.101887 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0982dbd-62a1-47c5-8510-5045b9ca5785-config\") pod \"a0982dbd-62a1-47c5-8510-5045b9ca5785\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.101977 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cqqg\" (UniqueName: \"kubernetes.io/projected/a0982dbd-62a1-47c5-8510-5045b9ca5785-kube-api-access-2cqqg\") pod \"a0982dbd-62a1-47c5-8510-5045b9ca5785\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.102126 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0982dbd-62a1-47c5-8510-5045b9ca5785-client-ca\") pod \"a0982dbd-62a1-47c5-8510-5045b9ca5785\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.103157 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0982dbd-62a1-47c5-8510-5045b9ca5785-client-ca" (OuterVolumeSpecName: "client-ca") pod "a0982dbd-62a1-47c5-8510-5045b9ca5785" (UID: "a0982dbd-62a1-47c5-8510-5045b9ca5785"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.103193 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a0982dbd-62a1-47c5-8510-5045b9ca5785-config" (OuterVolumeSpecName: "config") pod "a0982dbd-62a1-47c5-8510-5045b9ca5785" (UID: "a0982dbd-62a1-47c5-8510-5045b9ca5785"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.103220 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0982dbd-62a1-47c5-8510-5045b9ca5785-serving-cert\") pod \"a0982dbd-62a1-47c5-8510-5045b9ca5785\" (UID: \"a0982dbd-62a1-47c5-8510-5045b9ca5785\") " Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.103704 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a0982dbd-62a1-47c5-8510-5045b9ca5785-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.103726 4632 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a0982dbd-62a1-47c5-8510-5045b9ca5785-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.109620 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0982dbd-62a1-47c5-8510-5045b9ca5785-kube-api-access-2cqqg" (OuterVolumeSpecName: "kube-api-access-2cqqg") pod "a0982dbd-62a1-47c5-8510-5045b9ca5785" (UID: "a0982dbd-62a1-47c5-8510-5045b9ca5785"). InnerVolumeSpecName "kube-api-access-2cqqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.121156 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0982dbd-62a1-47c5-8510-5045b9ca5785-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a0982dbd-62a1-47c5-8510-5045b9ca5785" (UID: "a0982dbd-62a1-47c5-8510-5045b9ca5785"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.179300 4632 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-18 01:06:32.656724867 +0000 UTC Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.179360 4632 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6710h57m32.477368851s for next certificate rotation Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.204918 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cqqg\" (UniqueName: \"kubernetes.io/projected/a0982dbd-62a1-47c5-8510-5045b9ca5785-kube-api-access-2cqqg\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.205017 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a0982dbd-62a1-47c5-8510-5045b9ca5785-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.872556 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" event={"ID":"0816f595-2f7f-425a-9a6b-1022e2a4ca04","Type":"ContainerDied","Data":"650730d84aa4384d18f0228647070b36709f320f05d4d8ef5c14d5e680d6b8ca"} Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.873074 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="650730d84aa4384d18f0228647070b36709f320f05d4d8ef5c14d5e680d6b8ca" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.874541 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" event={"ID":"a0982dbd-62a1-47c5-8510-5045b9ca5785","Type":"ContainerDied","Data":"e2910cae586a594472f36055ddd013b1c782ed1e61c6b7e0b1e88de17b0d81e9"} Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.874624 4632 scope.go:117] "RemoveContainer" containerID="a4aeba6501ad544549300fc2c4204c4a5cf7f1d1edc84405ce7d7b1974c966cc" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.874687 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.874837 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.960767 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb"] Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.972761 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5c44777cb6-dkmdb"] Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.978481 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-568f8cc7b8-srcn5"] Mar 13 10:09:00 crc kubenswrapper[4632]: E0313 10:09:00.978850 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a0982dbd-62a1-47c5-8510-5045b9ca5785" containerName="route-controller-manager" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.978981 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a0982dbd-62a1-47c5-8510-5045b9ca5785" containerName="route-controller-manager" Mar 13 10:09:00 crc kubenswrapper[4632]: E0313 10:09:00.979056 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0816f595-2f7f-425a-9a6b-1022e2a4ca04" containerName="controller-manager" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.979199 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="0816f595-2f7f-425a-9a6b-1022e2a4ca04" containerName="controller-manager" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.979349 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="a0982dbd-62a1-47c5-8510-5045b9ca5785" containerName="route-controller-manager" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.979426 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="0816f595-2f7f-425a-9a6b-1022e2a4ca04" containerName="controller-manager" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.979824 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.993232 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc"] Mar 13 10:09:00 crc kubenswrapper[4632]: I0313 10:09:00.998833 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.004620 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-568f8cc7b8-srcn5"] Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.009362 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.009740 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.009509 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.009571 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.009665 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.010455 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.013677 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc"] Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.023744 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-client-ca\") pod \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.024070 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-proxy-ca-bundles\") pod \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.024811 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0816f595-2f7f-425a-9a6b-1022e2a4ca04" (UID: "0816f595-2f7f-425a-9a6b-1022e2a4ca04"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.024826 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-client-ca" (OuterVolumeSpecName: "client-ca") pod "0816f595-2f7f-425a-9a6b-1022e2a4ca04" (UID: "0816f595-2f7f-425a-9a6b-1022e2a4ca04"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.025086 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-config\") pod \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.025205 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0816f595-2f7f-425a-9a6b-1022e2a4ca04-serving-cert\") pod \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.025465 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2prq4\" (UniqueName: \"kubernetes.io/projected/0816f595-2f7f-425a-9a6b-1022e2a4ca04-kube-api-access-2prq4\") pod \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\" (UID: \"0816f595-2f7f-425a-9a6b-1022e2a4ca04\") " Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.025656 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-config" (OuterVolumeSpecName: "config") pod "0816f595-2f7f-425a-9a6b-1022e2a4ca04" (UID: "0816f595-2f7f-425a-9a6b-1022e2a4ca04"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.025860 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzsfh\" (UniqueName: \"kubernetes.io/projected/6c1bb71f-b506-4779-997a-b45aa2d7f99d-kube-api-access-gzsfh\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.025956 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-config\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.026045 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c1bb71f-b506-4779-997a-b45aa2d7f99d-serving-cert\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.026497 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmsjq\" (UniqueName: \"kubernetes.io/projected/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-kube-api-access-mmsjq\") pod \"route-controller-manager-db59c8bd6-cs8jc\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.026618 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-proxy-ca-bundles\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.026723 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-client-ca\") pod \"route-controller-manager-db59c8bd6-cs8jc\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.027478 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-client-ca\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.027643 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-config\") pod \"route-controller-manager-db59c8bd6-cs8jc\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.027825 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-serving-cert\") pod \"route-controller-manager-db59c8bd6-cs8jc\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.028052 4632 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.028144 4632 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.028239 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0816f595-2f7f-425a-9a6b-1022e2a4ca04-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.038230 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0816f595-2f7f-425a-9a6b-1022e2a4ca04-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0816f595-2f7f-425a-9a6b-1022e2a4ca04" (UID: "0816f595-2f7f-425a-9a6b-1022e2a4ca04"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.038246 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0816f595-2f7f-425a-9a6b-1022e2a4ca04-kube-api-access-2prq4" (OuterVolumeSpecName: "kube-api-access-2prq4") pod "0816f595-2f7f-425a-9a6b-1022e2a4ca04" (UID: "0816f595-2f7f-425a-9a6b-1022e2a4ca04"). InnerVolumeSpecName "kube-api-access-2prq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.130673 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c1bb71f-b506-4779-997a-b45aa2d7f99d-serving-cert\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.130754 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmsjq\" (UniqueName: \"kubernetes.io/projected/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-kube-api-access-mmsjq\") pod \"route-controller-manager-db59c8bd6-cs8jc\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.130786 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-proxy-ca-bundles\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.130864 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-client-ca\") pod \"route-controller-manager-db59c8bd6-cs8jc\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.130885 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-client-ca\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.130911 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-config\") pod \"route-controller-manager-db59c8bd6-cs8jc\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.130968 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-serving-cert\") pod \"route-controller-manager-db59c8bd6-cs8jc\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.131042 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzsfh\" (UniqueName: \"kubernetes.io/projected/6c1bb71f-b506-4779-997a-b45aa2d7f99d-kube-api-access-gzsfh\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.131062 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-config\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.131119 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0816f595-2f7f-425a-9a6b-1022e2a4ca04-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.131132 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2prq4\" (UniqueName: \"kubernetes.io/projected/0816f595-2f7f-425a-9a6b-1022e2a4ca04-kube-api-access-2prq4\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.133430 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-config\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.133801 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-client-ca\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.134825 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-client-ca\") pod \"route-controller-manager-db59c8bd6-cs8jc\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.135543 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-proxy-ca-bundles\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.145033 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-serving-cert\") pod \"route-controller-manager-db59c8bd6-cs8jc\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.145580 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c1bb71f-b506-4779-997a-b45aa2d7f99d-serving-cert\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.160673 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmsjq\" (UniqueName: \"kubernetes.io/projected/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-kube-api-access-mmsjq\") pod \"route-controller-manager-db59c8bd6-cs8jc\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.177282 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzsfh\" (UniqueName: \"kubernetes.io/projected/6c1bb71f-b506-4779-997a-b45aa2d7f99d-kube-api-access-gzsfh\") pod \"controller-manager-568f8cc7b8-srcn5\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.179712 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-config\") pod \"route-controller-manager-db59c8bd6-cs8jc\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.179846 4632 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-29 12:43:02.614047778 +0000 UTC Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.179866 4632 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6266h34m1.43418472s for next certificate rotation Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.241045 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556608-9kzfk" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.301302 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.326422 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.334575 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dmxng\" (UniqueName: \"kubernetes.io/projected/37ab6711-478f-4cc7-b9a4-c9baa126b1a3-kube-api-access-dmxng\") pod \"37ab6711-478f-4cc7-b9a4-c9baa126b1a3\" (UID: \"37ab6711-478f-4cc7-b9a4-c9baa126b1a3\") " Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.344143 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.345133 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37ab6711-478f-4cc7-b9a4-c9baa126b1a3-kube-api-access-dmxng" (OuterVolumeSpecName: "kube-api-access-dmxng") pod "37ab6711-478f-4cc7-b9a4-c9baa126b1a3" (UID: "37ab6711-478f-4cc7-b9a4-c9baa126b1a3"). InnerVolumeSpecName "kube-api-access-dmxng". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.436572 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a93bf50-5608-4b34-aea5-2f027d469fe7-kubelet-dir\") pod \"7a93bf50-5608-4b34-aea5-2f027d469fe7\" (UID: \"7a93bf50-5608-4b34-aea5-2f027d469fe7\") " Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.436657 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a93bf50-5608-4b34-aea5-2f027d469fe7-kube-api-access\") pod \"7a93bf50-5608-4b34-aea5-2f027d469fe7\" (UID: \"7a93bf50-5608-4b34-aea5-2f027d469fe7\") " Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.437076 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dmxng\" (UniqueName: \"kubernetes.io/projected/37ab6711-478f-4cc7-b9a4-c9baa126b1a3-kube-api-access-dmxng\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.440521 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a93bf50-5608-4b34-aea5-2f027d469fe7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7a93bf50-5608-4b34-aea5-2f027d469fe7" (UID: "7a93bf50-5608-4b34-aea5-2f027d469fe7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.445207 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a93bf50-5608-4b34-aea5-2f027d469fe7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7a93bf50-5608-4b34-aea5-2f027d469fe7" (UID: "7a93bf50-5608-4b34-aea5-2f027d469fe7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.538529 4632 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7a93bf50-5608-4b34-aea5-2f027d469fe7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.538578 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7a93bf50-5608-4b34-aea5-2f027d469fe7-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.641257 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc"] Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.674263 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-568f8cc7b8-srcn5"] Mar 13 10:09:01 crc kubenswrapper[4632]: W0313 10:09:01.685089 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6c1bb71f_b506_4779_997a_b45aa2d7f99d.slice/crio-cafd0a70d6f6f5cc33c18e32a6f9118b758897e41e807521c630031166429878 WatchSource:0}: Error finding container cafd0a70d6f6f5cc33c18e32a6f9118b758897e41e807521c630031166429878: Status 404 returned error can't find the container with id cafd0a70d6f6f5cc33c18e32a6f9118b758897e41e807521c630031166429878 Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.881993 4632 generic.go:334] "Generic (PLEG): container finished" podID="c822257d-9d2f-4b6f-87de-131de5cd0efe" containerID="481e1788f663e81921b410cd12a9e3666afaa2b706dda68096288fee3498f2fa" exitCode=0 Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.882140 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556606-mkrp2" event={"ID":"c822257d-9d2f-4b6f-87de-131de5cd0efe","Type":"ContainerDied","Data":"481e1788f663e81921b410cd12a9e3666afaa2b706dda68096288fee3498f2fa"} Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.885074 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" event={"ID":"6c1bb71f-b506-4779-997a-b45aa2d7f99d","Type":"ContainerStarted","Data":"cafd0a70d6f6f5cc33c18e32a6f9118b758897e41e807521c630031166429878"} Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.885999 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" event={"ID":"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd","Type":"ContainerStarted","Data":"354150b09f5c97255b0f1b1d13a0e96b39ea150051699e29f1947c848d48cfc7"} Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.887963 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7a93bf50-5608-4b34-aea5-2f027d469fe7","Type":"ContainerDied","Data":"8ca4d20a3432f315af224e18c181a59d62f280063326feb2da6b42b782268a63"} Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.887993 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ca4d20a3432f315af224e18c181a59d62f280063326feb2da6b42b782268a63" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.888024 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.899962 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-747c9765b-lqcx6" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.899972 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556608-9kzfk" event={"ID":"37ab6711-478f-4cc7-b9a4-c9baa126b1a3","Type":"ContainerDied","Data":"ff362806bee1867b720f220a4cde4dbe8551207f73438d3af60407d151505f16"} Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.900065 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ff362806bee1867b720f220a4cde4dbe8551207f73438d3af60407d151505f16" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.900009 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556608-9kzfk" Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.942186 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-747c9765b-lqcx6"] Mar 13 10:09:01 crc kubenswrapper[4632]: I0313 10:09:01.947495 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-747c9765b-lqcx6"] Mar 13 10:09:02 crc kubenswrapper[4632]: I0313 10:09:02.066470 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0816f595-2f7f-425a-9a6b-1022e2a4ca04" path="/var/lib/kubelet/pods/0816f595-2f7f-425a-9a6b-1022e2a4ca04/volumes" Mar 13 10:09:02 crc kubenswrapper[4632]: I0313 10:09:02.067743 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0982dbd-62a1-47c5-8510-5045b9ca5785" path="/var/lib/kubelet/pods/a0982dbd-62a1-47c5-8510-5045b9ca5785/volumes" Mar 13 10:09:02 crc kubenswrapper[4632]: I0313 10:09:02.910289 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" event={"ID":"6c1bb71f-b506-4779-997a-b45aa2d7f99d","Type":"ContainerStarted","Data":"e6647232ecd3206958018db1d05543d27c1d81899737af2d56a6d6a78463f69b"} Mar 13 10:09:02 crc kubenswrapper[4632]: I0313 10:09:02.911177 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:02 crc kubenswrapper[4632]: I0313 10:09:02.915547 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" event={"ID":"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd","Type":"ContainerStarted","Data":"c21dcff3106ebcb8e41bf57ec34ca478155cc655a5099ecf6ed4d6d8ef778c01"} Mar 13 10:09:02 crc kubenswrapper[4632]: I0313 10:09:02.916726 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:02 crc kubenswrapper[4632]: I0313 10:09:02.940168 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" podStartSLOduration=3.940146764 podStartE2EDuration="3.940146764s" podCreationTimestamp="2026-03-13 10:08:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:09:02.937666035 +0000 UTC m=+316.960196198" watchObservedRunningTime="2026-03-13 10:09:02.940146764 +0000 UTC m=+316.962676897" Mar 13 10:09:02 crc kubenswrapper[4632]: I0313 10:09:02.963047 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" podStartSLOduration=3.963018793 podStartE2EDuration="3.963018793s" podCreationTimestamp="2026-03-13 10:08:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:09:02.961427239 +0000 UTC m=+316.983957392" watchObservedRunningTime="2026-03-13 10:09:02.963018793 +0000 UTC m=+316.985548926" Mar 13 10:09:03 crc kubenswrapper[4632]: I0313 10:09:03.924144 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:03 crc kubenswrapper[4632]: I0313 10:09:03.930423 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:04 crc kubenswrapper[4632]: I0313 10:09:04.640361 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556606-mkrp2" Mar 13 10:09:04 crc kubenswrapper[4632]: I0313 10:09:04.710708 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sd756\" (UniqueName: \"kubernetes.io/projected/c822257d-9d2f-4b6f-87de-131de5cd0efe-kube-api-access-sd756\") pod \"c822257d-9d2f-4b6f-87de-131de5cd0efe\" (UID: \"c822257d-9d2f-4b6f-87de-131de5cd0efe\") " Mar 13 10:09:04 crc kubenswrapper[4632]: I0313 10:09:04.726994 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c822257d-9d2f-4b6f-87de-131de5cd0efe-kube-api-access-sd756" (OuterVolumeSpecName: "kube-api-access-sd756") pod "c822257d-9d2f-4b6f-87de-131de5cd0efe" (UID: "c822257d-9d2f-4b6f-87de-131de5cd0efe"). InnerVolumeSpecName "kube-api-access-sd756". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:09:04 crc kubenswrapper[4632]: I0313 10:09:04.814172 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sd756\" (UniqueName: \"kubernetes.io/projected/c822257d-9d2f-4b6f-87de-131de5cd0efe-kube-api-access-sd756\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:04 crc kubenswrapper[4632]: I0313 10:09:04.935181 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556606-mkrp2" event={"ID":"c822257d-9d2f-4b6f-87de-131de5cd0efe","Type":"ContainerDied","Data":"4b486b426e38ba0d310d07052394a9d5bdba25cfa8d2705294f114f94eaedc81"} Mar 13 10:09:04 crc kubenswrapper[4632]: I0313 10:09:04.935219 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4b486b426e38ba0d310d07052394a9d5bdba25cfa8d2705294f114f94eaedc81" Mar 13 10:09:04 crc kubenswrapper[4632]: I0313 10:09:04.935468 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556606-mkrp2" Mar 13 10:09:05 crc kubenswrapper[4632]: I0313 10:09:05.632830 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:09:05 crc kubenswrapper[4632]: I0313 10:09:05.632894 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:09:05 crc kubenswrapper[4632]: I0313 10:09:05.633591 4632 patch_prober.go:28] interesting pod/downloads-7954f5f757-w2hhj container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Mar 13 10:09:05 crc kubenswrapper[4632]: I0313 10:09:05.633751 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-w2hhj" podUID="7d155f24-9bfc-4039-9981-10e7f724fa51" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Mar 13 10:09:07 crc kubenswrapper[4632]: I0313 10:09:07.971791 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvh86" event={"ID":"bd46ae04-0610-4aa5-9385-dd45de66c5dd","Type":"ContainerStarted","Data":"ba15fa8797c3390ead2f6a2f6b5a64ad766bc4a942dfc13cbdc76a3242dd09c0"} Mar 13 10:09:08 crc kubenswrapper[4632]: I0313 10:09:08.980105 4632 generic.go:334] "Generic (PLEG): container finished" podID="cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" containerID="1c7d4d3dbdb9375cd1f14c42f62f344139bab8e0abb1403e1fe655b1b72e40c4" exitCode=0 Mar 13 10:09:08 crc kubenswrapper[4632]: I0313 10:09:08.980175 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xd455" event={"ID":"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8","Type":"ContainerDied","Data":"1c7d4d3dbdb9375cd1f14c42f62f344139bab8e0abb1403e1fe655b1b72e40c4"} Mar 13 10:09:08 crc kubenswrapper[4632]: I0313 10:09:08.981804 4632 generic.go:334] "Generic (PLEG): container finished" podID="bd46ae04-0610-4aa5-9385-dd45de66c5dd" containerID="ba15fa8797c3390ead2f6a2f6b5a64ad766bc4a942dfc13cbdc76a3242dd09c0" exitCode=0 Mar 13 10:09:08 crc kubenswrapper[4632]: I0313 10:09:08.981829 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvh86" event={"ID":"bd46ae04-0610-4aa5-9385-dd45de66c5dd","Type":"ContainerDied","Data":"ba15fa8797c3390ead2f6a2f6b5a64ad766bc4a942dfc13cbdc76a3242dd09c0"} Mar 13 10:09:13 crc kubenswrapper[4632]: I0313 10:09:13.027923 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xd455" event={"ID":"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8","Type":"ContainerStarted","Data":"f1dbecf7ff84705a27018ceaf7e07f776f8da213446108c63db8f788119a4f28"} Mar 13 10:09:13 crc kubenswrapper[4632]: I0313 10:09:13.048721 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2gc7" event={"ID":"a110c276-8516-4f9e-a6af-d6837cd0f387","Type":"ContainerStarted","Data":"ef106624caa843911d5171f0d70f22c07e7e2bd19b6992932276ca1226b858e3"} Mar 13 10:09:13 crc kubenswrapper[4632]: I0313 10:09:13.056481 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8wjg" event={"ID":"b11a7dff-bf08-44c3-b4f4-923119c13717","Type":"ContainerStarted","Data":"06491b70d16bc5a697f5518128f63de5fdeb769cc33d09d9262078f5aa75a5b8"} Mar 13 10:09:13 crc kubenswrapper[4632]: I0313 10:09:13.057671 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xd455" podStartSLOduration=4.379077818 podStartE2EDuration="1m18.05761769s" podCreationTimestamp="2026-03-13 10:07:55 +0000 UTC" firstStartedPulling="2026-03-13 10:07:58.786823407 +0000 UTC m=+252.809353540" lastFinishedPulling="2026-03-13 10:09:12.465363279 +0000 UTC m=+326.487893412" observedRunningTime="2026-03-13 10:09:13.05276817 +0000 UTC m=+327.075298313" watchObservedRunningTime="2026-03-13 10:09:13.05761769 +0000 UTC m=+327.080147833" Mar 13 10:09:13 crc kubenswrapper[4632]: I0313 10:09:13.060856 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvh86" event={"ID":"bd46ae04-0610-4aa5-9385-dd45de66c5dd","Type":"ContainerStarted","Data":"5a10aa8d51646d1f515364874b0426c82d85f03f52a4924f31299cb0395b0607"} Mar 13 10:09:13 crc kubenswrapper[4632]: I0313 10:09:13.133692 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jvh86" podStartSLOduration=5.474108285 podStartE2EDuration="1m19.13366979s" podCreationTimestamp="2026-03-13 10:07:54 +0000 UTC" firstStartedPulling="2026-03-13 10:07:58.814511709 +0000 UTC m=+252.837041842" lastFinishedPulling="2026-03-13 10:09:12.474073224 +0000 UTC m=+326.496603347" observedRunningTime="2026-03-13 10:09:13.130487784 +0000 UTC m=+327.153017917" watchObservedRunningTime="2026-03-13 10:09:13.13366979 +0000 UTC m=+327.156199923" Mar 13 10:09:14 crc kubenswrapper[4632]: I0313 10:09:14.071428 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr5l9" event={"ID":"87965e39-b879-4e26-9c8b-b78068c52aa0","Type":"ContainerStarted","Data":"e3753ece91188912f076d532ed434a938805848842e1fdb1b100800bcaa42763"} Mar 13 10:09:14 crc kubenswrapper[4632]: I0313 10:09:14.073731 4632 generic.go:334] "Generic (PLEG): container finished" podID="b11a7dff-bf08-44c3-b4f4-923119c13717" containerID="06491b70d16bc5a697f5518128f63de5fdeb769cc33d09d9262078f5aa75a5b8" exitCode=0 Mar 13 10:09:14 crc kubenswrapper[4632]: I0313 10:09:14.073803 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8wjg" event={"ID":"b11a7dff-bf08-44c3-b4f4-923119c13717","Type":"ContainerDied","Data":"06491b70d16bc5a697f5518128f63de5fdeb769cc33d09d9262078f5aa75a5b8"} Mar 13 10:09:14 crc kubenswrapper[4632]: I0313 10:09:14.077201 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txp2w" event={"ID":"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87","Type":"ContainerStarted","Data":"643ad1b648678ed35dcc10aaf9a844460c880f38f688c0da6821345eaf872208"} Mar 13 10:09:15 crc kubenswrapper[4632]: I0313 10:09:15.085074 4632 generic.go:334] "Generic (PLEG): container finished" podID="a110c276-8516-4f9e-a6af-d6837cd0f387" containerID="ef106624caa843911d5171f0d70f22c07e7e2bd19b6992932276ca1226b858e3" exitCode=0 Mar 13 10:09:15 crc kubenswrapper[4632]: I0313 10:09:15.085159 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2gc7" event={"ID":"a110c276-8516-4f9e-a6af-d6837cd0f387","Type":"ContainerDied","Data":"ef106624caa843911d5171f0d70f22c07e7e2bd19b6992932276ca1226b858e3"} Mar 13 10:09:15 crc kubenswrapper[4632]: I0313 10:09:15.087768 4632 generic.go:334] "Generic (PLEG): container finished" podID="f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" containerID="643ad1b648678ed35dcc10aaf9a844460c880f38f688c0da6821345eaf872208" exitCode=0 Mar 13 10:09:15 crc kubenswrapper[4632]: I0313 10:09:15.087801 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txp2w" event={"ID":"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87","Type":"ContainerDied","Data":"643ad1b648678ed35dcc10aaf9a844460c880f38f688c0da6821345eaf872208"} Mar 13 10:09:15 crc kubenswrapper[4632]: I0313 10:09:15.334830 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:09:15 crc kubenswrapper[4632]: I0313 10:09:15.334892 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:09:15 crc kubenswrapper[4632]: I0313 10:09:15.639777 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-w2hhj" Mar 13 10:09:15 crc kubenswrapper[4632]: I0313 10:09:15.707657 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xd455" Mar 13 10:09:15 crc kubenswrapper[4632]: I0313 10:09:15.708146 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xd455" Mar 13 10:09:16 crc kubenswrapper[4632]: I0313 10:09:16.055541 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xd455" Mar 13 10:09:16 crc kubenswrapper[4632]: I0313 10:09:16.094567 4632 generic.go:334] "Generic (PLEG): container finished" podID="87965e39-b879-4e26-9c8b-b78068c52aa0" containerID="e3753ece91188912f076d532ed434a938805848842e1fdb1b100800bcaa42763" exitCode=0 Mar 13 10:09:16 crc kubenswrapper[4632]: I0313 10:09:16.094631 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr5l9" event={"ID":"87965e39-b879-4e26-9c8b-b78068c52aa0","Type":"ContainerDied","Data":"e3753ece91188912f076d532ed434a938805848842e1fdb1b100800bcaa42763"} Mar 13 10:09:16 crc kubenswrapper[4632]: I0313 10:09:16.955136 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jvh86" podUID="bd46ae04-0610-4aa5-9385-dd45de66c5dd" containerName="registry-server" probeResult="failure" output=< Mar 13 10:09:16 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:09:16 crc kubenswrapper[4632]: > Mar 13 10:09:19 crc kubenswrapper[4632]: I0313 10:09:19.059462 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-568f8cc7b8-srcn5"] Mar 13 10:09:19 crc kubenswrapper[4632]: I0313 10:09:19.060406 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" podUID="6c1bb71f-b506-4779-997a-b45aa2d7f99d" containerName="controller-manager" containerID="cri-o://e6647232ecd3206958018db1d05543d27c1d81899737af2d56a6d6a78463f69b" gracePeriod=30 Mar 13 10:09:19 crc kubenswrapper[4632]: I0313 10:09:19.111145 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z668" event={"ID":"9845f384-2720-4d6a-aa73-1e66e30f7c2c","Type":"ContainerStarted","Data":"577f89a2c63ffc7af6c8d9a11a12240e1b316c59a7b108fcd47eb2cd9dc3c8ec"} Mar 13 10:09:19 crc kubenswrapper[4632]: I0313 10:09:19.114972 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8wjg" event={"ID":"b11a7dff-bf08-44c3-b4f4-923119c13717","Type":"ContainerStarted","Data":"55bfc00a5732a457ecbee5c7be945027bdb42c0137a6b22125d44dafb5924f59"} Mar 13 10:09:19 crc kubenswrapper[4632]: I0313 10:09:19.130612 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc"] Mar 13 10:09:19 crc kubenswrapper[4632]: I0313 10:09:19.131051 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" podUID="1d9f7553-a7a4-47b3-8898-990eb6d2fdfd" containerName="route-controller-manager" containerID="cri-o://c21dcff3106ebcb8e41bf57ec34ca478155cc655a5099ecf6ed4d6d8ef778c01" gracePeriod=30 Mar 13 10:09:19 crc kubenswrapper[4632]: I0313 10:09:19.185537 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p8wjg" podStartSLOduration=7.37858684 podStartE2EDuration="1m24.185514459s" podCreationTimestamp="2026-03-13 10:07:55 +0000 UTC" firstStartedPulling="2026-03-13 10:08:01.052597086 +0000 UTC m=+255.075127219" lastFinishedPulling="2026-03-13 10:09:17.859524705 +0000 UTC m=+331.882054838" observedRunningTime="2026-03-13 10:09:19.180438962 +0000 UTC m=+333.202969105" watchObservedRunningTime="2026-03-13 10:09:19.185514459 +0000 UTC m=+333.208044602" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.121173 4632 generic.go:334] "Generic (PLEG): container finished" podID="1d9f7553-a7a4-47b3-8898-990eb6d2fdfd" containerID="c21dcff3106ebcb8e41bf57ec34ca478155cc655a5099ecf6ed4d6d8ef778c01" exitCode=0 Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.121211 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" event={"ID":"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd","Type":"ContainerDied","Data":"c21dcff3106ebcb8e41bf57ec34ca478155cc655a5099ecf6ed4d6d8ef778c01"} Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.123375 4632 generic.go:334] "Generic (PLEG): container finished" podID="9845f384-2720-4d6a-aa73-1e66e30f7c2c" containerID="577f89a2c63ffc7af6c8d9a11a12240e1b316c59a7b108fcd47eb2cd9dc3c8ec" exitCode=0 Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.123508 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z668" event={"ID":"9845f384-2720-4d6a-aa73-1e66e30f7c2c","Type":"ContainerDied","Data":"577f89a2c63ffc7af6c8d9a11a12240e1b316c59a7b108fcd47eb2cd9dc3c8ec"} Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.124918 4632 generic.go:334] "Generic (PLEG): container finished" podID="6c1bb71f-b506-4779-997a-b45aa2d7f99d" containerID="e6647232ecd3206958018db1d05543d27c1d81899737af2d56a6d6a78463f69b" exitCode=0 Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.125433 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" event={"ID":"6c1bb71f-b506-4779-997a-b45aa2d7f99d","Type":"ContainerDied","Data":"e6647232ecd3206958018db1d05543d27c1d81899737af2d56a6d6a78463f69b"} Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.171801 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.172760 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.176273 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.177417 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.191600 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.194835 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.275726 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.276337 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.276494 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.279542 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.279849 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.288442 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.290716 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad-metrics-certs\") pod \"network-metrics-daemon-z2vlz\" (UID: \"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad\") " pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.302580 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.309633 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.359275 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.369118 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.378426 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.459385 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 13 10:09:20 crc kubenswrapper[4632]: I0313 10:09:20.467719 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-z2vlz" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.274854 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-z2vlz"] Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.516232 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:21 crc kubenswrapper[4632]: W0313 10:09:21.520432 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-95cf426241c42da749498523009aa75a5678b42076db750b6c0577fe95ada46f WatchSource:0}: Error finding container 95cf426241c42da749498523009aa75a5678b42076db750b6c0577fe95ada46f: Status 404 returned error can't find the container with id 95cf426241c42da749498523009aa75a5678b42076db750b6c0577fe95ada46f Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.549078 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.556139 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2"] Mar 13 10:09:21 crc kubenswrapper[4632]: E0313 10:09:21.556354 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d9f7553-a7a4-47b3-8898-990eb6d2fdfd" containerName="route-controller-manager" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.556367 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d9f7553-a7a4-47b3-8898-990eb6d2fdfd" containerName="route-controller-manager" Mar 13 10:09:21 crc kubenswrapper[4632]: E0313 10:09:21.556377 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c822257d-9d2f-4b6f-87de-131de5cd0efe" containerName="oc" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.556383 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c822257d-9d2f-4b6f-87de-131de5cd0efe" containerName="oc" Mar 13 10:09:21 crc kubenswrapper[4632]: E0313 10:09:21.556404 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c1bb71f-b506-4779-997a-b45aa2d7f99d" containerName="controller-manager" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.556411 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c1bb71f-b506-4779-997a-b45aa2d7f99d" containerName="controller-manager" Mar 13 10:09:21 crc kubenswrapper[4632]: E0313 10:09:21.556419 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37ab6711-478f-4cc7-b9a4-c9baa126b1a3" containerName="oc" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.556425 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="37ab6711-478f-4cc7-b9a4-c9baa126b1a3" containerName="oc" Mar 13 10:09:21 crc kubenswrapper[4632]: E0313 10:09:21.556443 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a93bf50-5608-4b34-aea5-2f027d469fe7" containerName="pruner" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.556448 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a93bf50-5608-4b34-aea5-2f027d469fe7" containerName="pruner" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.556534 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c822257d-9d2f-4b6f-87de-131de5cd0efe" containerName="oc" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.556546 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="37ab6711-478f-4cc7-b9a4-c9baa126b1a3" containerName="oc" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.556556 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a93bf50-5608-4b34-aea5-2f027d469fe7" containerName="pruner" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.556565 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c1bb71f-b506-4779-997a-b45aa2d7f99d" containerName="controller-manager" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.556572 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d9f7553-a7a4-47b3-8898-990eb6d2fdfd" containerName="route-controller-manager" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.556962 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.562603 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2"] Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.603826 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-client-ca\") pod \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.603881 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-serving-cert\") pod \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.603964 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzsfh\" (UniqueName: \"kubernetes.io/projected/6c1bb71f-b506-4779-997a-b45aa2d7f99d-kube-api-access-gzsfh\") pod \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.604023 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmsjq\" (UniqueName: \"kubernetes.io/projected/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-kube-api-access-mmsjq\") pod \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.604053 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c1bb71f-b506-4779-997a-b45aa2d7f99d-serving-cert\") pod \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.604084 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-client-ca\") pod \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.604131 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-config\") pod \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.604165 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-config\") pod \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\" (UID: \"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd\") " Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.604188 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-proxy-ca-bundles\") pod \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\" (UID: \"6c1bb71f-b506-4779-997a-b45aa2d7f99d\") " Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.604378 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mrsj\" (UniqueName: \"kubernetes.io/projected/2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3-kube-api-access-8mrsj\") pod \"route-controller-manager-db6b8fbf8-pllt2\" (UID: \"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3\") " pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.604416 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3-client-ca\") pod \"route-controller-manager-db6b8fbf8-pllt2\" (UID: \"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3\") " pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.604965 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-client-ca" (OuterVolumeSpecName: "client-ca") pod "1d9f7553-a7a4-47b3-8898-990eb6d2fdfd" (UID: "1d9f7553-a7a4-47b3-8898-990eb6d2fdfd"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.605325 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-config" (OuterVolumeSpecName: "config") pod "1d9f7553-a7a4-47b3-8898-990eb6d2fdfd" (UID: "1d9f7553-a7a4-47b3-8898-990eb6d2fdfd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.605544 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6c1bb71f-b506-4779-997a-b45aa2d7f99d" (UID: "6c1bb71f-b506-4779-997a-b45aa2d7f99d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.605555 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-config" (OuterVolumeSpecName: "config") pod "6c1bb71f-b506-4779-997a-b45aa2d7f99d" (UID: "6c1bb71f-b506-4779-997a-b45aa2d7f99d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.604447 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3-serving-cert\") pod \"route-controller-manager-db6b8fbf8-pllt2\" (UID: \"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3\") " pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.605664 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3-config\") pod \"route-controller-manager-db6b8fbf8-pllt2\" (UID: \"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3\") " pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.605747 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.605771 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.605784 4632 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.605795 4632 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.605588 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-client-ca" (OuterVolumeSpecName: "client-ca") pod "6c1bb71f-b506-4779-997a-b45aa2d7f99d" (UID: "6c1bb71f-b506-4779-997a-b45aa2d7f99d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.611385 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c1bb71f-b506-4779-997a-b45aa2d7f99d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6c1bb71f-b506-4779-997a-b45aa2d7f99d" (UID: "6c1bb71f-b506-4779-997a-b45aa2d7f99d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.611827 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c1bb71f-b506-4779-997a-b45aa2d7f99d-kube-api-access-gzsfh" (OuterVolumeSpecName: "kube-api-access-gzsfh") pod "6c1bb71f-b506-4779-997a-b45aa2d7f99d" (UID: "6c1bb71f-b506-4779-997a-b45aa2d7f99d"). InnerVolumeSpecName "kube-api-access-gzsfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.617627 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-kube-api-access-mmsjq" (OuterVolumeSpecName: "kube-api-access-mmsjq") pod "1d9f7553-a7a4-47b3-8898-990eb6d2fdfd" (UID: "1d9f7553-a7a4-47b3-8898-990eb6d2fdfd"). InnerVolumeSpecName "kube-api-access-mmsjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.611380 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1d9f7553-a7a4-47b3-8898-990eb6d2fdfd" (UID: "1d9f7553-a7a4-47b3-8898-990eb6d2fdfd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.706540 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3-config\") pod \"route-controller-manager-db6b8fbf8-pllt2\" (UID: \"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3\") " pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.706631 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mrsj\" (UniqueName: \"kubernetes.io/projected/2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3-kube-api-access-8mrsj\") pod \"route-controller-manager-db6b8fbf8-pllt2\" (UID: \"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3\") " pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.706652 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3-client-ca\") pod \"route-controller-manager-db6b8fbf8-pllt2\" (UID: \"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3\") " pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.706669 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3-serving-cert\") pod \"route-controller-manager-db6b8fbf8-pllt2\" (UID: \"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3\") " pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.706712 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmsjq\" (UniqueName: \"kubernetes.io/projected/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-kube-api-access-mmsjq\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.706724 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6c1bb71f-b506-4779-997a-b45aa2d7f99d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.706732 4632 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6c1bb71f-b506-4779-997a-b45aa2d7f99d-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.706741 4632 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.706751 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzsfh\" (UniqueName: \"kubernetes.io/projected/6c1bb71f-b506-4779-997a-b45aa2d7f99d-kube-api-access-gzsfh\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.710418 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3-config\") pod \"route-controller-manager-db6b8fbf8-pllt2\" (UID: \"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3\") " pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.711501 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3-client-ca\") pod \"route-controller-manager-db6b8fbf8-pllt2\" (UID: \"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3\") " pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.712472 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3-serving-cert\") pod \"route-controller-manager-db6b8fbf8-pllt2\" (UID: \"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3\") " pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.726393 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mrsj\" (UniqueName: \"kubernetes.io/projected/2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3-kube-api-access-8mrsj\") pod \"route-controller-manager-db6b8fbf8-pllt2\" (UID: \"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3\") " pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:21 crc kubenswrapper[4632]: I0313 10:09:21.883224 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.149441 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"95cf426241c42da749498523009aa75a5678b42076db750b6c0577fe95ada46f"} Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.152564 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" event={"ID":"6c1bb71f-b506-4779-997a-b45aa2d7f99d","Type":"ContainerDied","Data":"cafd0a70d6f6f5cc33c18e32a6f9118b758897e41e807521c630031166429878"} Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.152611 4632 scope.go:117] "RemoveContainer" containerID="e6647232ecd3206958018db1d05543d27c1d81899737af2d56a6d6a78463f69b" Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.152723 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.157632 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" event={"ID":"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad","Type":"ContainerStarted","Data":"33797dc2ede59a88fcccf5c3bd0b68134eb179bcb79626eff62306cd5a0425bc"} Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.160197 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.160138 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" event={"ID":"1d9f7553-a7a4-47b3-8898-990eb6d2fdfd","Type":"ContainerDied","Data":"354150b09f5c97255b0f1b1d13a0e96b39ea150051699e29f1947c848d48cfc7"} Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.173931 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-568f8cc7b8-srcn5"] Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.178890 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-568f8cc7b8-srcn5"] Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.187088 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc"] Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.190449 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc"] Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.302099 4632 patch_prober.go:28] interesting pod/controller-manager-568f8cc7b8-srcn5 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.63:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.302181 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-568f8cc7b8-srcn5" podUID="6c1bb71f-b506-4779-997a-b45aa2d7f99d" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.63:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.327670 4632 patch_prober.go:28] interesting pod/route-controller-manager-db59c8bd6-cs8jc container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.64:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.327751 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-db59c8bd6-cs8jc" podUID="1d9f7553-a7a4-47b3-8898-990eb6d2fdfd" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.64:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:09:22 crc kubenswrapper[4632]: I0313 10:09:22.507577 4632 scope.go:117] "RemoveContainer" containerID="c21dcff3106ebcb8e41bf57ec34ca478155cc655a5099ecf6ed4d6d8ef778c01" Mar 13 10:09:22 crc kubenswrapper[4632]: W0313 10:09:22.792970 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-d4757da4c3828cd79b3324b61ff4d762684579f8dd7aad23fbdcd12f80e73a08 WatchSource:0}: Error finding container d4757da4c3828cd79b3324b61ff4d762684579f8dd7aad23fbdcd12f80e73a08: Status 404 returned error can't find the container with id d4757da4c3828cd79b3324b61ff4d762684579f8dd7aad23fbdcd12f80e73a08 Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.041596 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2"] Mar 13 10:09:23 crc kubenswrapper[4632]: W0313 10:09:23.078474 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-7e24d73dd23728b8d14b7681596c8391aadc6bb6f428216ce8897f1024604589 WatchSource:0}: Error finding container 7e24d73dd23728b8d14b7681596c8391aadc6bb6f428216ce8897f1024604589: Status 404 returned error can't find the container with id 7e24d73dd23728b8d14b7681596c8391aadc6bb6f428216ce8897f1024604589 Mar 13 10:09:23 crc kubenswrapper[4632]: W0313 10:09:23.094327 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f5d4f7c_4d7b_4347_bd38_d5fd29fed3f3.slice/crio-271f65d9f3e67553609f1c4c12988dd76116340addb8ca064d609462b229d7b2 WatchSource:0}: Error finding container 271f65d9f3e67553609f1c4c12988dd76116340addb8ca064d609462b229d7b2: Status 404 returned error can't find the container with id 271f65d9f3e67553609f1c4c12988dd76116340addb8ca064d609462b229d7b2 Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.210726 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z668" event={"ID":"9845f384-2720-4d6a-aa73-1e66e30f7c2c","Type":"ContainerStarted","Data":"001fe1eb309384fca387c523307ccbff0d5d514d2d7b29f074cc94a2210761cb"} Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.211932 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" event={"ID":"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3","Type":"ContainerStarted","Data":"271f65d9f3e67553609f1c4c12988dd76116340addb8ca064d609462b229d7b2"} Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.249820 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8z668" podStartSLOduration=6.7169681 podStartE2EDuration="1m28.249784366s" podCreationTimestamp="2026-03-13 10:07:55 +0000 UTC" firstStartedPulling="2026-03-13 10:08:01.200664416 +0000 UTC m=+255.223194549" lastFinishedPulling="2026-03-13 10:09:22.733480672 +0000 UTC m=+336.756010815" observedRunningTime="2026-03-13 10:09:23.239782896 +0000 UTC m=+337.262313039" watchObservedRunningTime="2026-03-13 10:09:23.249784366 +0000 UTC m=+337.272314499" Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.251756 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr5l9" event={"ID":"87965e39-b879-4e26-9c8b-b78068c52aa0","Type":"ContainerStarted","Data":"f3ec0c7ec6e706f7309cf337b9ffdf829c043fe63c66f9db4fb7e26f23ae6edd"} Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.265461 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" event={"ID":"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad","Type":"ContainerStarted","Data":"a38ae589a44a7a92b909e7238ebce17db2b6570b0803136419752878342215ee"} Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.288806 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2gc7" event={"ID":"a110c276-8516-4f9e-a6af-d6837cd0f387","Type":"ContainerStarted","Data":"f7b31d5849d6707802fb373a1fe6f70b7a45ddade6fd6d9f2c7e5319e74f32d3"} Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.297558 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"7e24d73dd23728b8d14b7681596c8391aadc6bb6f428216ce8897f1024604589"} Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.299832 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xr5l9" podStartSLOduration=4.877968654 podStartE2EDuration="1m25.299794323s" podCreationTimestamp="2026-03-13 10:07:58 +0000 UTC" firstStartedPulling="2026-03-13 10:08:02.373301144 +0000 UTC m=+256.395831277" lastFinishedPulling="2026-03-13 10:09:22.795126823 +0000 UTC m=+336.817656946" observedRunningTime="2026-03-13 10:09:23.298375365 +0000 UTC m=+337.320905508" watchObservedRunningTime="2026-03-13 10:09:23.299794323 +0000 UTC m=+337.322324466" Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.314838 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"7d5e1531a06fe80e285365120ae28d66926c7bc01a7be4f92a00c0ec7d705f3c"} Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.325852 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-z2gc7" podStartSLOduration=5.193410114 podStartE2EDuration="1m25.325832166s" podCreationTimestamp="2026-03-13 10:07:58 +0000 UTC" firstStartedPulling="2026-03-13 10:08:02.379147069 +0000 UTC m=+256.401677202" lastFinishedPulling="2026-03-13 10:09:22.511569101 +0000 UTC m=+336.534099254" observedRunningTime="2026-03-13 10:09:23.322210068 +0000 UTC m=+337.344740211" watchObservedRunningTime="2026-03-13 10:09:23.325832166 +0000 UTC m=+337.348362299" Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.326457 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6bkt" event={"ID":"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e","Type":"ContainerStarted","Data":"9fa8755799160674c9b9254d4cc4cd33b06b805f90944344a392116f94021c66"} Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.334617 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txp2w" event={"ID":"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87","Type":"ContainerStarted","Data":"d3932d25c3aaf08a595c2af7ee315a6a0b2efd503369ee7398e6b39ad609dc3c"} Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.377444 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"5956ba337b2cab47b490c63dbc5d8cd7763461bc0270d79a46c262ba948b7af6"} Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.377503 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"d4757da4c3828cd79b3324b61ff4d762684579f8dd7aad23fbdcd12f80e73a08"} Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.378220 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.383550 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-txp2w" podStartSLOduration=7.762531822 podStartE2EDuration="1m26.383535891s" podCreationTimestamp="2026-03-13 10:07:57 +0000 UTC" firstStartedPulling="2026-03-13 10:08:02.395244955 +0000 UTC m=+256.417775088" lastFinishedPulling="2026-03-13 10:09:21.016249024 +0000 UTC m=+335.038779157" observedRunningTime="2026-03-13 10:09:23.383359096 +0000 UTC m=+337.405889229" watchObservedRunningTime="2026-03-13 10:09:23.383535891 +0000 UTC m=+337.406066024" Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.984114 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7469657588-kpf64"] Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.985175 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:23 crc kubenswrapper[4632]: W0313 10:09:23.988584 4632 reflector.go:561] object-"openshift-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Mar 13 10:09:23 crc kubenswrapper[4632]: E0313 10:09:23.988657 4632 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.988674 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.988978 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.989034 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 10:09:23 crc kubenswrapper[4632]: W0313 10:09:23.989049 4632 reflector.go:561] object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c": failed to list *v1.Secret: secrets "openshift-controller-manager-sa-dockercfg-msq4c" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Mar 13 10:09:23 crc kubenswrapper[4632]: E0313 10:09:23.989131 4632 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"openshift-controller-manager-sa-dockercfg-msq4c\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-controller-manager-sa-dockercfg-msq4c\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 13 10:09:23 crc kubenswrapper[4632]: I0313 10:09:23.992968 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.012541 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.051971 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d9f7553-a7a4-47b3-8898-990eb6d2fdfd" path="/var/lib/kubelet/pods/1d9f7553-a7a4-47b3-8898-990eb6d2fdfd/volumes" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.052818 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c1bb71f-b506-4779-997a-b45aa2d7f99d" path="/var/lib/kubelet/pods/6c1bb71f-b506-4779-997a-b45aa2d7f99d/volumes" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.058737 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7469657588-kpf64"] Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.147418 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8ff14f9-e25c-4839-acab-a622f6f70f88-proxy-ca-bundles\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.147524 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8ff14f9-e25c-4839-acab-a622f6f70f88-config\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.147562 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfhrk\" (UniqueName: \"kubernetes.io/projected/a8ff14f9-e25c-4839-acab-a622f6f70f88-kube-api-access-qfhrk\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.147834 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8ff14f9-e25c-4839-acab-a622f6f70f88-serving-cert\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.147898 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8ff14f9-e25c-4839-acab-a622f6f70f88-client-ca\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.249774 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8ff14f9-e25c-4839-acab-a622f6f70f88-config\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.249850 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfhrk\" (UniqueName: \"kubernetes.io/projected/a8ff14f9-e25c-4839-acab-a622f6f70f88-kube-api-access-qfhrk\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.249905 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8ff14f9-e25c-4839-acab-a622f6f70f88-serving-cert\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.250053 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8ff14f9-e25c-4839-acab-a622f6f70f88-client-ca\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.250089 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8ff14f9-e25c-4839-acab-a622f6f70f88-proxy-ca-bundles\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.251542 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/a8ff14f9-e25c-4839-acab-a622f6f70f88-proxy-ca-bundles\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.251587 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a8ff14f9-e25c-4839-acab-a622f6f70f88-client-ca\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.263272 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a8ff14f9-e25c-4839-acab-a622f6f70f88-serving-cert\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.272042 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfhrk\" (UniqueName: \"kubernetes.io/projected/a8ff14f9-e25c-4839-acab-a622f6f70f88-kube-api-access-qfhrk\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.396825 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e030373dea9ffbc53b1f37a054b8fe51529b1e5690243c2f3a5c0c872134f808"} Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.398672 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" event={"ID":"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3","Type":"ContainerStarted","Data":"71808a85287e54b9fb184ad4c73a074a1ff3d6b35824bd6122d42af589681e05"} Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.398837 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.400926 4632 generic.go:334] "Generic (PLEG): container finished" podID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" containerID="9fa8755799160674c9b9254d4cc4cd33b06b805f90944344a392116f94021c66" exitCode=0 Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.400964 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6bkt" event={"ID":"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e","Type":"ContainerDied","Data":"9fa8755799160674c9b9254d4cc4cd33b06b805f90944344a392116f94021c66"} Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.403456 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-z2vlz" event={"ID":"ab3aaffc-bf11-41a1-9a91-3bf97d2be4ad","Type":"ContainerStarted","Data":"048d443b7ebb101ab26472c762a7d62976644eb2a8a52ba9e023d75be07775dd"} Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.440635 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-z2vlz" podStartSLOduration=284.440611377 podStartE2EDuration="4m44.440611377s" podCreationTimestamp="2026-03-13 10:04:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:09:24.43813261 +0000 UTC m=+338.460662763" watchObservedRunningTime="2026-03-13 10:09:24.440611377 +0000 UTC m=+338.463141510" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.468373 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" podStartSLOduration=5.468351565 podStartE2EDuration="5.468351565s" podCreationTimestamp="2026-03-13 10:09:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:09:24.465757075 +0000 UTC m=+338.488287208" watchObservedRunningTime="2026-03-13 10:09:24.468351565 +0000 UTC m=+338.490881698" Mar 13 10:09:24 crc kubenswrapper[4632]: I0313 10:09:24.566886 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 10:09:25 crc kubenswrapper[4632]: I0313 10:09:25.017078 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 10:09:25 crc kubenswrapper[4632]: I0313 10:09:25.125029 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 10:09:25 crc kubenswrapper[4632]: I0313 10:09:25.134202 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a8ff14f9-e25c-4839-acab-a622f6f70f88-config\") pod \"controller-manager-7469657588-kpf64\" (UID: \"a8ff14f9-e25c-4839-acab-a622f6f70f88\") " pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:25 crc kubenswrapper[4632]: I0313 10:09:25.201323 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:25 crc kubenswrapper[4632]: I0313 10:09:25.416146 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6bkt" event={"ID":"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e","Type":"ContainerStarted","Data":"18dc7d50bd20cdcc8bd750dbb90186346da79c165a2b1398fb9fce2a58fd1224"} Mar 13 10:09:25 crc kubenswrapper[4632]: I0313 10:09:25.446168 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:09:25 crc kubenswrapper[4632]: I0313 10:09:25.452105 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-t6bkt" podStartSLOduration=4.839266753 podStartE2EDuration="1m27.452077464s" podCreationTimestamp="2026-03-13 10:07:58 +0000 UTC" firstStartedPulling="2026-03-13 10:08:02.326755102 +0000 UTC m=+256.349285235" lastFinishedPulling="2026-03-13 10:09:24.939565813 +0000 UTC m=+338.962095946" observedRunningTime="2026-03-13 10:09:25.44667907 +0000 UTC m=+339.469209213" watchObservedRunningTime="2026-03-13 10:09:25.452077464 +0000 UTC m=+339.474607607" Mar 13 10:09:25 crc kubenswrapper[4632]: I0313 10:09:25.571403 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:09:25 crc kubenswrapper[4632]: I0313 10:09:25.817729 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xd455" Mar 13 10:09:25 crc kubenswrapper[4632]: I0313 10:09:25.947722 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7469657588-kpf64"] Mar 13 10:09:26 crc kubenswrapper[4632]: I0313 10:09:26.421262 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" event={"ID":"a8ff14f9-e25c-4839-acab-a622f6f70f88","Type":"ContainerStarted","Data":"432a739d763fd09cb52fbc4a7bbe481e0fb4c89b88f7822f73b594d3596d0d39"} Mar 13 10:09:26 crc kubenswrapper[4632]: I0313 10:09:26.422561 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" event={"ID":"a8ff14f9-e25c-4839-acab-a622f6f70f88","Type":"ContainerStarted","Data":"0c9127c8a737402e76c86b1f73dd05b17a90037c7a9a2c7f8c0120f64d74a91d"} Mar 13 10:09:26 crc kubenswrapper[4632]: I0313 10:09:26.866717 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:09:26 crc kubenswrapper[4632]: I0313 10:09:26.866804 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:09:26 crc kubenswrapper[4632]: I0313 10:09:26.932704 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:09:27 crc kubenswrapper[4632]: I0313 10:09:27.095431 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:09:27 crc kubenswrapper[4632]: I0313 10:09:27.095492 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:09:27 crc kubenswrapper[4632]: I0313 10:09:27.149394 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:09:27 crc kubenswrapper[4632]: I0313 10:09:27.430464 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:27 crc kubenswrapper[4632]: I0313 10:09:27.438864 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 10:09:27 crc kubenswrapper[4632]: I0313 10:09:27.454202 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" podStartSLOduration=8.454183168 podStartE2EDuration="8.454183168s" podCreationTimestamp="2026-03-13 10:09:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:09:27.45385313 +0000 UTC m=+341.476383273" watchObservedRunningTime="2026-03-13 10:09:27.454183168 +0000 UTC m=+341.476713291" Mar 13 10:09:27 crc kubenswrapper[4632]: I0313 10:09:27.481384 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:09:27 crc kubenswrapper[4632]: I0313 10:09:27.496085 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:09:28 crc kubenswrapper[4632]: I0313 10:09:28.508437 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:09:28 crc kubenswrapper[4632]: I0313 10:09:28.508505 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:09:28 crc kubenswrapper[4632]: I0313 10:09:28.560348 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:09:28 crc kubenswrapper[4632]: I0313 10:09:28.979334 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:09:28 crc kubenswrapper[4632]: I0313 10:09:28.979674 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:09:28 crc kubenswrapper[4632]: I0313 10:09:28.993982 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:09:28 crc kubenswrapper[4632]: I0313 10:09:28.994028 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.038017 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.051930 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.052308 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.089962 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xd455"] Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.090254 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xd455" podUID="cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" containerName="registry-server" containerID="cri-o://f1dbecf7ff84705a27018ceaf7e07f776f8da213446108c63db8f788119a4f28" gracePeriod=2 Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.445080 4632 generic.go:334] "Generic (PLEG): container finished" podID="cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" containerID="f1dbecf7ff84705a27018ceaf7e07f776f8da213446108c63db8f788119a4f28" exitCode=0 Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.445196 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xd455" event={"ID":"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8","Type":"ContainerDied","Data":"f1dbecf7ff84705a27018ceaf7e07f776f8da213446108c63db8f788119a4f28"} Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.491726 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.584363 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xd455" Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.653029 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-catalog-content\") pod \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\" (UID: \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\") " Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.653148 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-utilities\") pod \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\" (UID: \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\") " Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.653191 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kh6g9\" (UniqueName: \"kubernetes.io/projected/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-kube-api-access-kh6g9\") pod \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\" (UID: \"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8\") " Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.654165 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-utilities" (OuterVolumeSpecName: "utilities") pod "cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" (UID: "cd6e3c73-fbc1-4213-bbef-02dd2b0587f8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.659194 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-kube-api-access-kh6g9" (OuterVolumeSpecName: "kube-api-access-kh6g9") pod "cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" (UID: "cd6e3c73-fbc1-4213-bbef-02dd2b0587f8"). InnerVolumeSpecName "kube-api-access-kh6g9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.691498 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8z668"] Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.691748 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8z668" podUID="9845f384-2720-4d6a-aa73-1e66e30f7c2c" containerName="registry-server" containerID="cri-o://001fe1eb309384fca387c523307ccbff0d5d514d2d7b29f074cc94a2210761cb" gracePeriod=2 Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.716011 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" (UID: "cd6e3c73-fbc1-4213-bbef-02dd2b0587f8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.756202 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.756247 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:29 crc kubenswrapper[4632]: I0313 10:09:29.756260 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kh6g9\" (UniqueName: \"kubernetes.io/projected/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8-kube-api-access-kh6g9\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.015311 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xr5l9" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" containerName="registry-server" probeResult="failure" output=< Mar 13 10:09:30 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:09:30 crc kubenswrapper[4632]: > Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.089445 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.103861 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-z2gc7" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" containerName="registry-server" probeResult="failure" output=< Mar 13 10:09:30 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:09:30 crc kubenswrapper[4632]: > Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.162342 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9845f384-2720-4d6a-aa73-1e66e30f7c2c-utilities\") pod \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\" (UID: \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\") " Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.162424 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9845f384-2720-4d6a-aa73-1e66e30f7c2c-catalog-content\") pod \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\" (UID: \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\") " Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.162517 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdp56\" (UniqueName: \"kubernetes.io/projected/9845f384-2720-4d6a-aa73-1e66e30f7c2c-kube-api-access-sdp56\") pod \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\" (UID: \"9845f384-2720-4d6a-aa73-1e66e30f7c2c\") " Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.163310 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9845f384-2720-4d6a-aa73-1e66e30f7c2c-utilities" (OuterVolumeSpecName: "utilities") pod "9845f384-2720-4d6a-aa73-1e66e30f7c2c" (UID: "9845f384-2720-4d6a-aa73-1e66e30f7c2c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.165736 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9845f384-2720-4d6a-aa73-1e66e30f7c2c-kube-api-access-sdp56" (OuterVolumeSpecName: "kube-api-access-sdp56") pod "9845f384-2720-4d6a-aa73-1e66e30f7c2c" (UID: "9845f384-2720-4d6a-aa73-1e66e30f7c2c"). InnerVolumeSpecName "kube-api-access-sdp56". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.223084 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9845f384-2720-4d6a-aa73-1e66e30f7c2c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9845f384-2720-4d6a-aa73-1e66e30f7c2c" (UID: "9845f384-2720-4d6a-aa73-1e66e30f7c2c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.264117 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9845f384-2720-4d6a-aa73-1e66e30f7c2c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.264162 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdp56\" (UniqueName: \"kubernetes.io/projected/9845f384-2720-4d6a-aa73-1e66e30f7c2c-kube-api-access-sdp56\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.264174 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9845f384-2720-4d6a-aa73-1e66e30f7c2c-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.452500 4632 generic.go:334] "Generic (PLEG): container finished" podID="9845f384-2720-4d6a-aa73-1e66e30f7c2c" containerID="001fe1eb309384fca387c523307ccbff0d5d514d2d7b29f074cc94a2210761cb" exitCode=0 Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.452583 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z668" event={"ID":"9845f384-2720-4d6a-aa73-1e66e30f7c2c","Type":"ContainerDied","Data":"001fe1eb309384fca387c523307ccbff0d5d514d2d7b29f074cc94a2210761cb"} Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.452587 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8z668" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.452624 4632 scope.go:117] "RemoveContainer" containerID="001fe1eb309384fca387c523307ccbff0d5d514d2d7b29f074cc94a2210761cb" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.452614 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8z668" event={"ID":"9845f384-2720-4d6a-aa73-1e66e30f7c2c","Type":"ContainerDied","Data":"eb6537c579cc3249bae831f8164a219c024fbc6e74b0df55017ce52d6b143567"} Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.455556 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xd455" event={"ID":"cd6e3c73-fbc1-4213-bbef-02dd2b0587f8","Type":"ContainerDied","Data":"33259f14f07cee3a1d7261a44a8f74cbd0957ccef81016b9631c3a0a7ccd4085"} Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.455607 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xd455" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.473085 4632 scope.go:117] "RemoveContainer" containerID="577f89a2c63ffc7af6c8d9a11a12240e1b316c59a7b108fcd47eb2cd9dc3c8ec" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.480155 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8z668"] Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.491150 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8z668"] Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.497463 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xd455"] Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.500735 4632 scope.go:117] "RemoveContainer" containerID="a609582fe9641518ec575a14e6a93f5bb1f502cb63d1e38602356c26bed99217" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.502342 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xd455"] Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.524666 4632 scope.go:117] "RemoveContainer" containerID="001fe1eb309384fca387c523307ccbff0d5d514d2d7b29f074cc94a2210761cb" Mar 13 10:09:30 crc kubenswrapper[4632]: E0313 10:09:30.525072 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"001fe1eb309384fca387c523307ccbff0d5d514d2d7b29f074cc94a2210761cb\": container with ID starting with 001fe1eb309384fca387c523307ccbff0d5d514d2d7b29f074cc94a2210761cb not found: ID does not exist" containerID="001fe1eb309384fca387c523307ccbff0d5d514d2d7b29f074cc94a2210761cb" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.525101 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"001fe1eb309384fca387c523307ccbff0d5d514d2d7b29f074cc94a2210761cb"} err="failed to get container status \"001fe1eb309384fca387c523307ccbff0d5d514d2d7b29f074cc94a2210761cb\": rpc error: code = NotFound desc = could not find container \"001fe1eb309384fca387c523307ccbff0d5d514d2d7b29f074cc94a2210761cb\": container with ID starting with 001fe1eb309384fca387c523307ccbff0d5d514d2d7b29f074cc94a2210761cb not found: ID does not exist" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.525120 4632 scope.go:117] "RemoveContainer" containerID="577f89a2c63ffc7af6c8d9a11a12240e1b316c59a7b108fcd47eb2cd9dc3c8ec" Mar 13 10:09:30 crc kubenswrapper[4632]: E0313 10:09:30.525310 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"577f89a2c63ffc7af6c8d9a11a12240e1b316c59a7b108fcd47eb2cd9dc3c8ec\": container with ID starting with 577f89a2c63ffc7af6c8d9a11a12240e1b316c59a7b108fcd47eb2cd9dc3c8ec not found: ID does not exist" containerID="577f89a2c63ffc7af6c8d9a11a12240e1b316c59a7b108fcd47eb2cd9dc3c8ec" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.525326 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"577f89a2c63ffc7af6c8d9a11a12240e1b316c59a7b108fcd47eb2cd9dc3c8ec"} err="failed to get container status \"577f89a2c63ffc7af6c8d9a11a12240e1b316c59a7b108fcd47eb2cd9dc3c8ec\": rpc error: code = NotFound desc = could not find container \"577f89a2c63ffc7af6c8d9a11a12240e1b316c59a7b108fcd47eb2cd9dc3c8ec\": container with ID starting with 577f89a2c63ffc7af6c8d9a11a12240e1b316c59a7b108fcd47eb2cd9dc3c8ec not found: ID does not exist" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.525338 4632 scope.go:117] "RemoveContainer" containerID="a609582fe9641518ec575a14e6a93f5bb1f502cb63d1e38602356c26bed99217" Mar 13 10:09:30 crc kubenswrapper[4632]: E0313 10:09:30.525490 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a609582fe9641518ec575a14e6a93f5bb1f502cb63d1e38602356c26bed99217\": container with ID starting with a609582fe9641518ec575a14e6a93f5bb1f502cb63d1e38602356c26bed99217 not found: ID does not exist" containerID="a609582fe9641518ec575a14e6a93f5bb1f502cb63d1e38602356c26bed99217" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.525503 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a609582fe9641518ec575a14e6a93f5bb1f502cb63d1e38602356c26bed99217"} err="failed to get container status \"a609582fe9641518ec575a14e6a93f5bb1f502cb63d1e38602356c26bed99217\": rpc error: code = NotFound desc = could not find container \"a609582fe9641518ec575a14e6a93f5bb1f502cb63d1e38602356c26bed99217\": container with ID starting with a609582fe9641518ec575a14e6a93f5bb1f502cb63d1e38602356c26bed99217 not found: ID does not exist" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.525517 4632 scope.go:117] "RemoveContainer" containerID="f1dbecf7ff84705a27018ceaf7e07f776f8da213446108c63db8f788119a4f28" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.572495 4632 scope.go:117] "RemoveContainer" containerID="1c7d4d3dbdb9375cd1f14c42f62f344139bab8e0abb1403e1fe655b1b72e40c4" Mar 13 10:09:30 crc kubenswrapper[4632]: I0313 10:09:30.587565 4632 scope.go:117] "RemoveContainer" containerID="39c617653cdae12029a38a740d3aa9e4c08c056d9865caf4f87830fbf0817555" Mar 13 10:09:32 crc kubenswrapper[4632]: I0313 10:09:32.054060 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9845f384-2720-4d6a-aa73-1e66e30f7c2c" path="/var/lib/kubelet/pods/9845f384-2720-4d6a-aa73-1e66e30f7c2c/volumes" Mar 13 10:09:32 crc kubenswrapper[4632]: I0313 10:09:32.055752 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" path="/var/lib/kubelet/pods/cd6e3c73-fbc1-4213-bbef-02dd2b0587f8/volumes" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.668046 4632 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.668636 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54" gracePeriod=15 Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.668791 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94" gracePeriod=15 Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.668832 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc" gracePeriod=15 Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.668862 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf" gracePeriod=15 Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.668890 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe" gracePeriod=15 Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.669758 4632 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670142 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" containerName="extract-utilities" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670171 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" containerName="extract-utilities" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670192 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670224 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670236 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" containerName="extract-content" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670244 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" containerName="extract-content" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670255 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9845f384-2720-4d6a-aa73-1e66e30f7c2c" containerName="registry-server" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670262 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="9845f384-2720-4d6a-aa73-1e66e30f7c2c" containerName="registry-server" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670274 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9845f384-2720-4d6a-aa73-1e66e30f7c2c" containerName="extract-content" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670311 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="9845f384-2720-4d6a-aa73-1e66e30f7c2c" containerName="extract-content" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670323 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670344 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670352 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670385 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670398 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670405 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670416 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670423 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670434 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670441 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670479 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9845f384-2720-4d6a-aa73-1e66e30f7c2c" containerName="extract-utilities" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670486 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="9845f384-2720-4d6a-aa73-1e66e30f7c2c" containerName="extract-utilities" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670500 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670507 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670516 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670523 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670558 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670565 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.670573 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" containerName="registry-server" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670579 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" containerName="registry-server" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670729 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd6e3c73-fbc1-4213-bbef-02dd2b0587f8" containerName="registry-server" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670754 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670764 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670795 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670802 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670812 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670821 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="9845f384-2720-4d6a-aa73-1e66e30f7c2c" containerName="registry-server" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670830 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.670839 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: E0313 10:09:35.671008 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.671039 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.671215 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.671231 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.672502 4632 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.672962 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.694621 4632 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.718841 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.742008 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.742280 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.742395 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.742493 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.742601 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.742737 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.742831 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.742976 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.844538 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.844632 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.844693 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.844713 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.844744 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.844764 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.844725 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.844888 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.844810 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.844924 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.844968 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.844996 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.845007 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.845019 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.845072 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:35 crc kubenswrapper[4632]: I0313 10:09:35.845149 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.014846 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:09:36 crc kubenswrapper[4632]: E0313 10:09:36.047777 4632 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.182:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c5eca412cae33 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:09:36.046984755 +0000 UTC m=+350.069514888,LastTimestamp:2026-03-13 10:09:36.046984755 +0000 UTC m=+350.069514888,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:09:36 crc kubenswrapper[4632]: E0313 10:09:36.083384 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:09:36Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:09:36Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:09:36Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:09:36Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:1295a1f0e74ae87f51a733e28b64c6fdb6b9a5b069a6897b3870fe52cc1c3b0b\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:505eeaa3f051e9f4ea6a622aca92e5c4eae07078ca185d9fecfe8cc9b6dfc899\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1739173859},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:4855408bd0e4d0711383d0c14dcad53c98255ff9f83f6cbefb57e47eacc1f1f1\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:97bdbb5854e4ad7976209a44cff02c8a2b9542f58ad007c06a5c3a5e8266def1\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1284762325},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:898c67bf7fc973e99114f3148976a6c21ae0dbe413051415588fa9b995f5b331\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:a641939d2096609a4cf6eec872a1476b7c671bfd81cffc2edeb6e9f13c9deeba\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1231028434},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:70c85e2aeb7db0a454101307851f490057ab53449c50ad9d86c54a698dd4913a\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:762bdf2da1fce19a4a24a6931f555b482c5c2314895b2f68aed74658266819a7\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1221741278},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-cli@sha256:69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9\\\",\\\"registry.redhat.io/openshift4/ose-cli@sha256:ef83967297f619f45075e7fd1428a1eb981622a6c174c46fb53b158ed24bed85\\\",\\\"registry.redhat.io/openshift4/ose-cli:latest\\\"],\\\"sizeBytes\\\":584351326},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:36 crc kubenswrapper[4632]: E0313 10:09:36.083836 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:36 crc kubenswrapper[4632]: E0313 10:09:36.084108 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:36 crc kubenswrapper[4632]: E0313 10:09:36.084376 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:36 crc kubenswrapper[4632]: E0313 10:09:36.084594 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:36 crc kubenswrapper[4632]: E0313 10:09:36.084616 4632 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.495214 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.496493 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.497173 4632 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94" exitCode=0 Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.497196 4632 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc" exitCode=0 Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.497204 4632 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf" exitCode=0 Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.497212 4632 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe" exitCode=2 Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.497289 4632 scope.go:117] "RemoveContainer" containerID="9898f7d8c644921fdc43a7906faeef577b236192f94a1ed911525b50ba8e68ae" Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.499696 4632 generic.go:334] "Generic (PLEG): container finished" podID="dc39d207-84a2-4a28-9296-bed684aa308d" containerID="c463b62ee1a6928ceb028fe480183c5ca7bb846ec47d4163fa232376e05db524" exitCode=0 Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.499765 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"dc39d207-84a2-4a28-9296-bed684aa308d","Type":"ContainerDied","Data":"c463b62ee1a6928ceb028fe480183c5ca7bb846ec47d4163fa232376e05db524"} Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.501010 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.501459 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.502226 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"80143265ebaee0f3a54053c9c203e48c0b8ae49b675b972452da2268c99bd9ad"} Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.502261 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"7655fa97fbc58dc4d8162ba613e0a9424aef91ec5c3ab185bf572e8fb571eb1a"} Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.502676 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:36 crc kubenswrapper[4632]: I0313 10:09:36.503014 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:37 crc kubenswrapper[4632]: I0313 10:09:37.524096 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 10:09:37 crc kubenswrapper[4632]: E0313 10:09:37.709386 4632 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.182:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c5eca412cae33 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:09:36.046984755 +0000 UTC m=+350.069514888,LastTimestamp:2026-03-13 10:09:36.046984755 +0000 UTC m=+350.069514888,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.046349 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.047152 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.062817 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.063326 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.063729 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.117921 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.118871 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.119677 4632 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.120187 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.120692 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.177850 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.177958 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dc39d207-84a2-4a28-9296-bed684aa308d-var-lock\") pod \"dc39d207-84a2-4a28-9296-bed684aa308d\" (UID: \"dc39d207-84a2-4a28-9296-bed684aa308d\") " Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.178005 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.178047 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.178064 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc39d207-84a2-4a28-9296-bed684aa308d-var-lock" (OuterVolumeSpecName: "var-lock") pod "dc39d207-84a2-4a28-9296-bed684aa308d" (UID: "dc39d207-84a2-4a28-9296-bed684aa308d"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.178094 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.178150 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.178151 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.178188 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc39d207-84a2-4a28-9296-bed684aa308d-kube-api-access\") pod \"dc39d207-84a2-4a28-9296-bed684aa308d\" (UID: \"dc39d207-84a2-4a28-9296-bed684aa308d\") " Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.178235 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc39d207-84a2-4a28-9296-bed684aa308d-kubelet-dir\") pod \"dc39d207-84a2-4a28-9296-bed684aa308d\" (UID: \"dc39d207-84a2-4a28-9296-bed684aa308d\") " Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.178693 4632 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.178720 4632 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/dc39d207-84a2-4a28-9296-bed684aa308d-var-lock\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.178732 4632 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.178743 4632 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.178801 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/dc39d207-84a2-4a28-9296-bed684aa308d-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "dc39d207-84a2-4a28-9296-bed684aa308d" (UID: "dc39d207-84a2-4a28-9296-bed684aa308d"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.183968 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc39d207-84a2-4a28-9296-bed684aa308d-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "dc39d207-84a2-4a28-9296-bed684aa308d" (UID: "dc39d207-84a2-4a28-9296-bed684aa308d"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.280052 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/dc39d207-84a2-4a28-9296-bed684aa308d-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.280594 4632 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/dc39d207-84a2-4a28-9296-bed684aa308d-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.554167 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.555008 4632 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54" exitCode=0 Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.555076 4632 scope.go:117] "RemoveContainer" containerID="21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.555252 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.564098 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"dc39d207-84a2-4a28-9296-bed684aa308d","Type":"ContainerDied","Data":"7017997794d887d37a83222e44995b72d5d076c42028b8e1498fdb1f2cb4d188"} Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.564140 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7017997794d887d37a83222e44995b72d5d076c42028b8e1498fdb1f2cb4d188" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.564220 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.573105 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.574062 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.574309 4632 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.576417 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.576637 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.576911 4632 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.581053 4632 scope.go:117] "RemoveContainer" containerID="706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.597771 4632 scope.go:117] "RemoveContainer" containerID="516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.611314 4632 scope.go:117] "RemoveContainer" containerID="cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.624105 4632 scope.go:117] "RemoveContainer" containerID="5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.640393 4632 scope.go:117] "RemoveContainer" containerID="7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.659119 4632 scope.go:117] "RemoveContainer" containerID="21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94" Mar 13 10:09:38 crc kubenswrapper[4632]: E0313 10:09:38.659882 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\": container with ID starting with 21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94 not found: ID does not exist" containerID="21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.659919 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94"} err="failed to get container status \"21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\": rpc error: code = NotFound desc = could not find container \"21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94\": container with ID starting with 21d33939d9cfb29a666b40f15b0bd1e73ec4c62db28db999433541739f2a1c94 not found: ID does not exist" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.659963 4632 scope.go:117] "RemoveContainer" containerID="706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc" Mar 13 10:09:38 crc kubenswrapper[4632]: E0313 10:09:38.661509 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\": container with ID starting with 706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc not found: ID does not exist" containerID="706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.661547 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc"} err="failed to get container status \"706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\": rpc error: code = NotFound desc = could not find container \"706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc\": container with ID starting with 706e2c8f5823a8e4b5a39a6a6869078c9e5fdf615672ecea030e48b8ab5d13fc not found: ID does not exist" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.661568 4632 scope.go:117] "RemoveContainer" containerID="516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf" Mar 13 10:09:38 crc kubenswrapper[4632]: E0313 10:09:38.661918 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\": container with ID starting with 516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf not found: ID does not exist" containerID="516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.662214 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf"} err="failed to get container status \"516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\": rpc error: code = NotFound desc = could not find container \"516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf\": container with ID starting with 516cedf1e60450f8fef80ef00bf0c4e23dd2a43c5296c49bd4b969c053aa73bf not found: ID does not exist" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.662308 4632 scope.go:117] "RemoveContainer" containerID="cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe" Mar 13 10:09:38 crc kubenswrapper[4632]: E0313 10:09:38.662930 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\": container with ID starting with cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe not found: ID does not exist" containerID="cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.663101 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe"} err="failed to get container status \"cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\": rpc error: code = NotFound desc = could not find container \"cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe\": container with ID starting with cc09fb371b9e509733ffbb2d3e0190320c5ab77049f1e7c01c74eeb4799944fe not found: ID does not exist" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.663373 4632 scope.go:117] "RemoveContainer" containerID="5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54" Mar 13 10:09:38 crc kubenswrapper[4632]: E0313 10:09:38.663874 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\": container with ID starting with 5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54 not found: ID does not exist" containerID="5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.663902 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54"} err="failed to get container status \"5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\": rpc error: code = NotFound desc = could not find container \"5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54\": container with ID starting with 5feec47af6cb6bcb0c36f1d5d8a17f568b30ee5b49c27d96b97d82626d650d54 not found: ID does not exist" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.663918 4632 scope.go:117] "RemoveContainer" containerID="7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990" Mar 13 10:09:38 crc kubenswrapper[4632]: E0313 10:09:38.664549 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\": container with ID starting with 7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990 not found: ID does not exist" containerID="7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990" Mar 13 10:09:38 crc kubenswrapper[4632]: I0313 10:09:38.664581 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990"} err="failed to get container status \"7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\": rpc error: code = NotFound desc = could not find container \"7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990\": container with ID starting with 7deddadce58678bd2e5b5e8f190024255509ceb687563db8f9d8984bba296990 not found: ID does not exist" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.026326 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.027730 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.028051 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.028318 4632 status_manager.go:851] "Failed to get status for pod" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" pod="openshift-marketplace/redhat-operators-xr5l9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xr5l9\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.028662 4632 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.053554 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.054056 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.054331 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.054523 4632 status_manager.go:851] "Failed to get status for pod" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" pod="openshift-marketplace/redhat-operators-xr5l9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xr5l9\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.054747 4632 status_manager.go:851] "Failed to get status for pod" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" pod="openshift-marketplace/redhat-marketplace-t6bkt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-t6bkt\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.055154 4632 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.071773 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.072287 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.072732 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.073022 4632 status_manager.go:851] "Failed to get status for pod" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" pod="openshift-marketplace/redhat-operators-xr5l9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xr5l9\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.073354 4632 status_manager.go:851] "Failed to get status for pod" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" pod="openshift-marketplace/redhat-marketplace-t6bkt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-t6bkt\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.073565 4632 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.094748 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.095361 4632 status_manager.go:851] "Failed to get status for pod" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" pod="openshift-marketplace/redhat-operators-z2gc7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-z2gc7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.095706 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.095996 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.096302 4632 status_manager.go:851] "Failed to get status for pod" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" pod="openshift-marketplace/redhat-operators-xr5l9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xr5l9\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.096729 4632 status_manager.go:851] "Failed to get status for pod" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" pod="openshift-marketplace/redhat-marketplace-t6bkt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-t6bkt\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.097059 4632 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.133198 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.133738 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.134172 4632 status_manager.go:851] "Failed to get status for pod" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" pod="openshift-marketplace/redhat-operators-z2gc7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-z2gc7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.134548 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.134895 4632 status_manager.go:851] "Failed to get status for pod" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" pod="openshift-marketplace/redhat-operators-xr5l9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xr5l9\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.135179 4632 status_manager.go:851] "Failed to get status for pod" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" pod="openshift-marketplace/redhat-marketplace-t6bkt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-t6bkt\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:39 crc kubenswrapper[4632]: I0313 10:09:39.135545 4632 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:40 crc kubenswrapper[4632]: I0313 10:09:40.050279 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 13 10:09:45 crc kubenswrapper[4632]: E0313 10:09:45.331634 4632 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:45 crc kubenswrapper[4632]: E0313 10:09:45.332691 4632 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:45 crc kubenswrapper[4632]: E0313 10:09:45.333188 4632 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:45 crc kubenswrapper[4632]: E0313 10:09:45.333440 4632 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:45 crc kubenswrapper[4632]: E0313 10:09:45.333881 4632 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:45 crc kubenswrapper[4632]: I0313 10:09:45.333914 4632 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 13 10:09:45 crc kubenswrapper[4632]: E0313 10:09:45.334395 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="200ms" Mar 13 10:09:45 crc kubenswrapper[4632]: E0313 10:09:45.535395 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="400ms" Mar 13 10:09:45 crc kubenswrapper[4632]: E0313 10:09:45.936848 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="800ms" Mar 13 10:09:46 crc kubenswrapper[4632]: E0313 10:09:46.301110 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:09:46Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:09:46Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:09:46Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T10:09:46Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:1295a1f0e74ae87f51a733e28b64c6fdb6b9a5b069a6897b3870fe52cc1c3b0b\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:505eeaa3f051e9f4ea6a622aca92e5c4eae07078ca185d9fecfe8cc9b6dfc899\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1739173859},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:4855408bd0e4d0711383d0c14dcad53c98255ff9f83f6cbefb57e47eacc1f1f1\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:97bdbb5854e4ad7976209a44cff02c8a2b9542f58ad007c06a5c3a5e8266def1\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1284762325},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:898c67bf7fc973e99114f3148976a6c21ae0dbe413051415588fa9b995f5b331\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:a641939d2096609a4cf6eec872a1476b7c671bfd81cffc2edeb6e9f13c9deeba\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1231028434},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:70c85e2aeb7db0a454101307851f490057ab53449c50ad9d86c54a698dd4913a\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:762bdf2da1fce19a4a24a6931f555b482c5c2314895b2f68aed74658266819a7\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1221741278},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-cli@sha256:69762925e16053d77685ff3a08b3b45dd2bfa5d68277851bc6969b368bbd0cb9\\\",\\\"registry.redhat.io/openshift4/ose-cli@sha256:ef83967297f619f45075e7fd1428a1eb981622a6c174c46fb53b158ed24bed85\\\",\\\"registry.redhat.io/openshift4/ose-cli:latest\\\"],\\\"sizeBytes\\\":584351326},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:46 crc kubenswrapper[4632]: E0313 10:09:46.301772 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:46 crc kubenswrapper[4632]: E0313 10:09:46.302334 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:46 crc kubenswrapper[4632]: E0313 10:09:46.302609 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:46 crc kubenswrapper[4632]: E0313 10:09:46.302979 4632 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:46 crc kubenswrapper[4632]: E0313 10:09:46.303010 4632 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 10:09:46 crc kubenswrapper[4632]: E0313 10:09:46.738284 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="1.6s" Mar 13 10:09:47 crc kubenswrapper[4632]: E0313 10:09:47.711268 4632 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.182:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c5eca412cae33 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 10:09:36.046984755 +0000 UTC m=+350.069514888,LastTimestamp:2026-03-13 10:09:36.046984755 +0000 UTC m=+350.069514888,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.046070 4632 status_manager.go:851] "Failed to get status for pod" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" pod="openshift-marketplace/redhat-operators-xr5l9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xr5l9\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.047682 4632 status_manager.go:851] "Failed to get status for pod" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" pod="openshift-marketplace/redhat-marketplace-t6bkt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-t6bkt\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.048012 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.048485 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.048818 4632 status_manager.go:851] "Failed to get status for pod" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" pod="openshift-marketplace/redhat-operators-z2gc7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-z2gc7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:48 crc kubenswrapper[4632]: E0313 10:09:48.339446 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="3.2s" Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.638800 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.639470 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.639522 4632 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d" exitCode=1 Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.639558 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d"} Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.640091 4632 scope.go:117] "RemoveContainer" containerID="8207ad7aa524df5af853ad8235c24a6addbc04e168248391629f00124901672d" Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.644525 4632 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.645139 4632 status_manager.go:851] "Failed to get status for pod" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" pod="openshift-marketplace/redhat-operators-z2gc7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-z2gc7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.645724 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.645973 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.646217 4632 status_manager.go:851] "Failed to get status for pod" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" pod="openshift-marketplace/redhat-operators-xr5l9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xr5l9\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:48 crc kubenswrapper[4632]: I0313 10:09:48.646599 4632 status_manager.go:851] "Failed to get status for pod" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" pod="openshift-marketplace/redhat-marketplace-t6bkt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-t6bkt\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:49 crc kubenswrapper[4632]: I0313 10:09:49.464191 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:09:49 crc kubenswrapper[4632]: I0313 10:09:49.647565 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 10:09:49 crc kubenswrapper[4632]: I0313 10:09:49.648181 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 13 10:09:49 crc kubenswrapper[4632]: I0313 10:09:49.648224 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"baae4c96bdfec2410a2abf4602bb303365672a79eb0060c14f3d9416601f60d1"} Mar 13 10:09:49 crc kubenswrapper[4632]: I0313 10:09:49.648932 4632 status_manager.go:851] "Failed to get status for pod" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" pod="openshift-marketplace/redhat-marketplace-t6bkt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-t6bkt\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:49 crc kubenswrapper[4632]: I0313 10:09:49.649286 4632 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:49 crc kubenswrapper[4632]: I0313 10:09:49.649558 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:49 crc kubenswrapper[4632]: I0313 10:09:49.649820 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:49 crc kubenswrapper[4632]: I0313 10:09:49.650190 4632 status_manager.go:851] "Failed to get status for pod" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" pod="openshift-marketplace/redhat-operators-z2gc7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-z2gc7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:49 crc kubenswrapper[4632]: I0313 10:09:49.650431 4632 status_manager.go:851] "Failed to get status for pod" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" pod="openshift-marketplace/redhat-operators-xr5l9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xr5l9\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:50 crc kubenswrapper[4632]: I0313 10:09:50.659264 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.043605 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.045669 4632 status_manager.go:851] "Failed to get status for pod" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" pod="openshift-marketplace/redhat-operators-xr5l9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xr5l9\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.046336 4632 status_manager.go:851] "Failed to get status for pod" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" pod="openshift-marketplace/redhat-marketplace-t6bkt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-t6bkt\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.046735 4632 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.048381 4632 status_manager.go:851] "Failed to get status for pod" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" pod="openshift-marketplace/redhat-operators-z2gc7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-z2gc7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.049144 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.049551 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.063767 4632 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="894cdc70-0747-4975-a22f-0dbd657e91a3" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.063841 4632 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="894cdc70-0747-4975-a22f-0dbd657e91a3" Mar 13 10:09:51 crc kubenswrapper[4632]: E0313 10:09:51.064551 4632 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.065180 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:51 crc kubenswrapper[4632]: W0313 10:09:51.093925 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-feb3d08f78c5327c6b227ef5a9b9e96cb8c32626a27e5135a0eef53c7204e559 WatchSource:0}: Error finding container feb3d08f78c5327c6b227ef5a9b9e96cb8c32626a27e5135a0eef53c7204e559: Status 404 returned error can't find the container with id feb3d08f78c5327c6b227ef5a9b9e96cb8c32626a27e5135a0eef53c7204e559 Mar 13 10:09:51 crc kubenswrapper[4632]: E0313 10:09:51.371892 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-conmon-5b221fb36af28b42296aad6aec56f5d67570fe622107623a4d8c3a607f65ef16.scope\": RecentStats: unable to find data in memory cache]" Mar 13 10:09:51 crc kubenswrapper[4632]: E0313 10:09:51.540755 4632 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.182:6443: connect: connection refused" interval="6.4s" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.664800 4632 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="5b221fb36af28b42296aad6aec56f5d67570fe622107623a4d8c3a607f65ef16" exitCode=0 Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.664866 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"5b221fb36af28b42296aad6aec56f5d67570fe622107623a4d8c3a607f65ef16"} Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.664972 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"feb3d08f78c5327c6b227ef5a9b9e96cb8c32626a27e5135a0eef53c7204e559"} Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.665453 4632 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="894cdc70-0747-4975-a22f-0dbd657e91a3" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.665480 4632 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="894cdc70-0747-4975-a22f-0dbd657e91a3" Mar 13 10:09:51 crc kubenswrapper[4632]: E0313 10:09:51.666351 4632 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.666376 4632 status_manager.go:851] "Failed to get status for pod" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" pod="openshift-marketplace/redhat-operators-xr5l9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-xr5l9\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.667892 4632 status_manager.go:851] "Failed to get status for pod" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" pod="openshift-marketplace/redhat-marketplace-t6bkt" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-t6bkt\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.668493 4632 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.669343 4632 status_manager.go:851] "Failed to get status for pod" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" pod="openshift-marketplace/redhat-operators-z2gc7" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-z2gc7\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.669855 4632 status_manager.go:851] "Failed to get status for pod" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:51 crc kubenswrapper[4632]: I0313 10:09:51.670318 4632 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.182:6443: connect: connection refused" Mar 13 10:09:52 crc kubenswrapper[4632]: I0313 10:09:52.678564 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"016bae0f85c65923e3cb3cd2dcce29a8981231a5ebd3e4c3946a0114414ae9c4"} Mar 13 10:09:52 crc kubenswrapper[4632]: I0313 10:09:52.679055 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"fdd6e9435e3bd2de1b2abf061d353492f0fc229bccd25152f17a70a41c909a82"} Mar 13 10:09:52 crc kubenswrapper[4632]: I0313 10:09:52.679071 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"da8aaa640f92e70999f9e593cd4ac8ad057243816be267692cf6551a0391d4ff"} Mar 13 10:09:52 crc kubenswrapper[4632]: I0313 10:09:52.679087 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"985806ca915f8813ea8bb973aff61c5c2b66f7004966f949041798df4ba45a99"} Mar 13 10:09:53 crc kubenswrapper[4632]: I0313 10:09:53.688004 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"5181b69ae7fe22887bd111820025dd22fb49a6daf2e2154f18ad4508ff0af707"} Mar 13 10:09:53 crc kubenswrapper[4632]: I0313 10:09:53.688309 4632 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="894cdc70-0747-4975-a22f-0dbd657e91a3" Mar 13 10:09:53 crc kubenswrapper[4632]: I0313 10:09:53.688325 4632 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="894cdc70-0747-4975-a22f-0dbd657e91a3" Mar 13 10:09:53 crc kubenswrapper[4632]: I0313 10:09:53.688519 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:56 crc kubenswrapper[4632]: I0313 10:09:56.066139 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:56 crc kubenswrapper[4632]: I0313 10:09:56.066534 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:56 crc kubenswrapper[4632]: I0313 10:09:56.071906 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:58 crc kubenswrapper[4632]: I0313 10:09:58.712665 4632 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:58 crc kubenswrapper[4632]: I0313 10:09:58.965660 4632 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="40937936-798b-414f-b10d-bc9cd5536d78" Mar 13 10:09:59 crc kubenswrapper[4632]: I0313 10:09:59.464129 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:09:59 crc kubenswrapper[4632]: I0313 10:09:59.469175 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:09:59 crc kubenswrapper[4632]: I0313 10:09:59.720554 4632 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="894cdc70-0747-4975-a22f-0dbd657e91a3" Mar 13 10:09:59 crc kubenswrapper[4632]: I0313 10:09:59.720607 4632 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="894cdc70-0747-4975-a22f-0dbd657e91a3" Mar 13 10:09:59 crc kubenswrapper[4632]: I0313 10:09:59.725178 4632 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="40937936-798b-414f-b10d-bc9cd5536d78" Mar 13 10:09:59 crc kubenswrapper[4632]: I0313 10:09:59.727442 4632 status_manager.go:308] "Container readiness changed before pod has synced" pod="openshift-kube-apiserver/kube-apiserver-crc" containerID="cri-o://985806ca915f8813ea8bb973aff61c5c2b66f7004966f949041798df4ba45a99" Mar 13 10:09:59 crc kubenswrapper[4632]: I0313 10:09:59.727475 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:09:59 crc kubenswrapper[4632]: I0313 10:09:59.727638 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:10:00 crc kubenswrapper[4632]: I0313 10:10:00.365712 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 10:10:00 crc kubenswrapper[4632]: I0313 10:10:00.729388 4632 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="894cdc70-0747-4975-a22f-0dbd657e91a3" Mar 13 10:10:00 crc kubenswrapper[4632]: I0313 10:10:00.729744 4632 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="894cdc70-0747-4975-a22f-0dbd657e91a3" Mar 13 10:10:00 crc kubenswrapper[4632]: I0313 10:10:00.733714 4632 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="40937936-798b-414f-b10d-bc9cd5536d78" Mar 13 10:10:08 crc kubenswrapper[4632]: I0313 10:10:08.200040 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 13 10:10:09 crc kubenswrapper[4632]: I0313 10:10:09.212185 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 13 10:10:09 crc kubenswrapper[4632]: I0313 10:10:09.339015 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 13 10:10:09 crc kubenswrapper[4632]: I0313 10:10:09.472930 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 13 10:10:09 crc kubenswrapper[4632]: I0313 10:10:09.503962 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 13 10:10:09 crc kubenswrapper[4632]: I0313 10:10:09.653500 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 13 10:10:09 crc kubenswrapper[4632]: I0313 10:10:09.820255 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 13 10:10:09 crc kubenswrapper[4632]: I0313 10:10:09.944366 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 13 10:10:09 crc kubenswrapper[4632]: I0313 10:10:09.980279 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 13 10:10:10 crc kubenswrapper[4632]: I0313 10:10:10.015309 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 13 10:10:10 crc kubenswrapper[4632]: I0313 10:10:10.301572 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 13 10:10:10 crc kubenswrapper[4632]: I0313 10:10:10.470062 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 13 10:10:10 crc kubenswrapper[4632]: I0313 10:10:10.487167 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 13 10:10:10 crc kubenswrapper[4632]: I0313 10:10:10.563044 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 13 10:10:10 crc kubenswrapper[4632]: I0313 10:10:10.608484 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 13 10:10:10 crc kubenswrapper[4632]: I0313 10:10:10.635830 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 13 10:10:10 crc kubenswrapper[4632]: I0313 10:10:10.826590 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 13 10:10:10 crc kubenswrapper[4632]: I0313 10:10:10.878575 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 13 10:10:10 crc kubenswrapper[4632]: I0313 10:10:10.892154 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 13 10:10:10 crc kubenswrapper[4632]: I0313 10:10:10.970712 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 13 10:10:11 crc kubenswrapper[4632]: I0313 10:10:11.100787 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 13 10:10:11 crc kubenswrapper[4632]: I0313 10:10:11.239978 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 13 10:10:11 crc kubenswrapper[4632]: I0313 10:10:11.392994 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 13 10:10:11 crc kubenswrapper[4632]: I0313 10:10:11.591293 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 13 10:10:11 crc kubenswrapper[4632]: I0313 10:10:11.591673 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 13 10:10:11 crc kubenswrapper[4632]: I0313 10:10:11.604147 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 13 10:10:11 crc kubenswrapper[4632]: I0313 10:10:11.723371 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 13 10:10:11 crc kubenswrapper[4632]: I0313 10:10:11.750175 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 13 10:10:11 crc kubenswrapper[4632]: I0313 10:10:11.756140 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 13 10:10:11 crc kubenswrapper[4632]: I0313 10:10:11.845156 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 13 10:10:11 crc kubenswrapper[4632]: I0313 10:10:11.922550 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 10:10:11 crc kubenswrapper[4632]: I0313 10:10:11.975557 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 13 10:10:11 crc kubenswrapper[4632]: I0313 10:10:11.975713 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.037587 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.188821 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.213862 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.241077 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.294920 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.317515 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.442500 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.623449 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.648497 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.679992 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.712567 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.715710 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.734773 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.750360 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.767491 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.843622 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.941301 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 13 10:10:12 crc kubenswrapper[4632]: I0313 10:10:12.973682 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.002688 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.052984 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.075100 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.312283 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.357416 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.383806 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.468932 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.510032 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.730596 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.738575 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.745712 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.885578 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.896458 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.964067 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.972262 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.973130 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 13 10:10:13 crc kubenswrapper[4632]: I0313 10:10:13.980160 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 13 10:10:14 crc kubenswrapper[4632]: I0313 10:10:14.166801 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 13 10:10:14 crc kubenswrapper[4632]: I0313 10:10:14.231170 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 13 10:10:14 crc kubenswrapper[4632]: I0313 10:10:14.241061 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 13 10:10:14 crc kubenswrapper[4632]: I0313 10:10:14.319101 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 13 10:10:14 crc kubenswrapper[4632]: I0313 10:10:14.352865 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 10:10:14 crc kubenswrapper[4632]: I0313 10:10:14.478395 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 13 10:10:14 crc kubenswrapper[4632]: I0313 10:10:14.506671 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 13 10:10:14 crc kubenswrapper[4632]: I0313 10:10:14.560846 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 13 10:10:14 crc kubenswrapper[4632]: I0313 10:10:14.667909 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 13 10:10:14 crc kubenswrapper[4632]: I0313 10:10:14.704228 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 13 10:10:14 crc kubenswrapper[4632]: I0313 10:10:14.707630 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 13 10:10:14 crc kubenswrapper[4632]: I0313 10:10:14.796474 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 13 10:10:14 crc kubenswrapper[4632]: I0313 10:10:14.816826 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 13 10:10:14 crc kubenswrapper[4632]: I0313 10:10:14.883147 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.083676 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.083796 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.202270 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.234201 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.438044 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.441819 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.458597 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.481630 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.501257 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.520572 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.631884 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.806578 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.874968 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.882891 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.921142 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 13 10:10:15 crc kubenswrapper[4632]: I0313 10:10:15.979962 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.046017 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.049034 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.122288 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.156302 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.178054 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.184667 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.186890 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.193958 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.211818 4632 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.223795 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.265845 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.321023 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.361882 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.392332 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.440647 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.454335 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.463381 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.508922 4632 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.510869 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=41.510848116 podStartE2EDuration="41.510848116s" podCreationTimestamp="2026-03-13 10:09:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:09:58.71875137 +0000 UTC m=+372.741281513" watchObservedRunningTime="2026-03-13 10:10:16.510848116 +0000 UTC m=+390.533378249" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.518330 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.518428 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.527142 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.546479 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=18.546444216 podStartE2EDuration="18.546444216s" podCreationTimestamp="2026-03-13 10:09:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:10:16.542735966 +0000 UTC m=+390.565266119" watchObservedRunningTime="2026-03-13 10:10:16.546444216 +0000 UTC m=+390.568974349" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.666369 4632 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.674508 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.700048 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.718367 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.733528 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.784585 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.797253 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.868227 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.881577 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.930005 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 13 10:10:16 crc kubenswrapper[4632]: I0313 10:10:16.947772 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.019012 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.024496 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.031870 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.060104 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556610-sg5bx"] Mar 13 10:10:17 crc kubenswrapper[4632]: E0313 10:10:17.060816 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" containerName="installer" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.060897 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" containerName="installer" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.061120 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc39d207-84a2-4a28-9296-bed684aa308d" containerName="installer" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.061726 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556610-sg5bx" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.064396 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.064652 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.071732 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.127325 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmrds\" (UniqueName: \"kubernetes.io/projected/795727b7-7a2e-4e97-8707-aecf893fd332-kube-api-access-tmrds\") pod \"auto-csr-approver-29556610-sg5bx\" (UID: \"795727b7-7a2e-4e97-8707-aecf893fd332\") " pod="openshift-infra/auto-csr-approver-29556610-sg5bx" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.168281 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.228295 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmrds\" (UniqueName: \"kubernetes.io/projected/795727b7-7a2e-4e97-8707-aecf893fd332-kube-api-access-tmrds\") pod \"auto-csr-approver-29556610-sg5bx\" (UID: \"795727b7-7a2e-4e97-8707-aecf893fd332\") " pod="openshift-infra/auto-csr-approver-29556610-sg5bx" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.257279 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmrds\" (UniqueName: \"kubernetes.io/projected/795727b7-7a2e-4e97-8707-aecf893fd332-kube-api-access-tmrds\") pod \"auto-csr-approver-29556610-sg5bx\" (UID: \"795727b7-7a2e-4e97-8707-aecf893fd332\") " pod="openshift-infra/auto-csr-approver-29556610-sg5bx" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.409755 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556610-sg5bx" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.459556 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.494829 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.521703 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.578384 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.612816 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.637127 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.756738 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.775605 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.798899 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.962222 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 13 10:10:17 crc kubenswrapper[4632]: I0313 10:10:17.978801 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.015216 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.074077 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.124248 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.153614 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.153679 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.222372 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.259905 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.295932 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.311835 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.339888 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.384585 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.484925 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.492537 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.534564 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.672613 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.893518 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.920136 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.922106 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 13 10:10:18 crc kubenswrapper[4632]: I0313 10:10:18.937442 4632 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.051914 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.123188 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.125324 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.154853 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.483237 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.483630 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.558585 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.561204 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.610295 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.636023 4632 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.638843 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.841070 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.860159 4632 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.892744 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.904540 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 13 10:10:19 crc kubenswrapper[4632]: I0313 10:10:19.917781 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.064226 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.109041 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.142527 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.159710 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.229691 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.302255 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.330760 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.415082 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.458089 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.506721 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.557438 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.576687 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.635631 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.723127 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.725149 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556610-sg5bx"] Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.805033 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.883997 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.886485 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.887254 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 13 10:10:20 crc kubenswrapper[4632]: I0313 10:10:20.893815 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.050872 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.074786 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.081726 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.283737 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.337333 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.423264 4632 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.423633 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://80143265ebaee0f3a54053c9c203e48c0b8ae49b675b972452da2268c99bd9ad" gracePeriod=5 Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.439833 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 13 10:10:21 crc kubenswrapper[4632]: E0313 10:10:21.467291 4632 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 13 10:10:21 crc kubenswrapper[4632]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29556610-sg5bx_openshift-infra_795727b7-7a2e-4e97-8707-aecf893fd332_0(f44c58a66c3c09c74b677cf4ec9f76b205bf4b131d95b02c4a9987dbc9e9bd8a): error adding pod openshift-infra_auto-csr-approver-29556610-sg5bx to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f44c58a66c3c09c74b677cf4ec9f76b205bf4b131d95b02c4a9987dbc9e9bd8a" Netns:"/var/run/netns/731b0fe9-200c-49e0-9ae2-97aa60457e8f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-infra;K8S_POD_NAME=auto-csr-approver-29556610-sg5bx;K8S_POD_INFRA_CONTAINER_ID=f44c58a66c3c09c74b677cf4ec9f76b205bf4b131d95b02c4a9987dbc9e9bd8a;K8S_POD_UID=795727b7-7a2e-4e97-8707-aecf893fd332" Path:"" ERRORED: error configuring pod [openshift-infra/auto-csr-approver-29556610-sg5bx] networking: Multus: [openshift-infra/auto-csr-approver-29556610-sg5bx/795727b7-7a2e-4e97-8707-aecf893fd332]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod auto-csr-approver-29556610-sg5bx in out of cluster comm: pod "auto-csr-approver-29556610-sg5bx" not found Mar 13 10:10:21 crc kubenswrapper[4632]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 10:10:21 crc kubenswrapper[4632]: > Mar 13 10:10:21 crc kubenswrapper[4632]: E0313 10:10:21.467366 4632 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 13 10:10:21 crc kubenswrapper[4632]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29556610-sg5bx_openshift-infra_795727b7-7a2e-4e97-8707-aecf893fd332_0(f44c58a66c3c09c74b677cf4ec9f76b205bf4b131d95b02c4a9987dbc9e9bd8a): error adding pod openshift-infra_auto-csr-approver-29556610-sg5bx to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f44c58a66c3c09c74b677cf4ec9f76b205bf4b131d95b02c4a9987dbc9e9bd8a" Netns:"/var/run/netns/731b0fe9-200c-49e0-9ae2-97aa60457e8f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-infra;K8S_POD_NAME=auto-csr-approver-29556610-sg5bx;K8S_POD_INFRA_CONTAINER_ID=f44c58a66c3c09c74b677cf4ec9f76b205bf4b131d95b02c4a9987dbc9e9bd8a;K8S_POD_UID=795727b7-7a2e-4e97-8707-aecf893fd332" Path:"" ERRORED: error configuring pod [openshift-infra/auto-csr-approver-29556610-sg5bx] networking: Multus: [openshift-infra/auto-csr-approver-29556610-sg5bx/795727b7-7a2e-4e97-8707-aecf893fd332]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod auto-csr-approver-29556610-sg5bx in out of cluster comm: pod "auto-csr-approver-29556610-sg5bx" not found Mar 13 10:10:21 crc kubenswrapper[4632]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 10:10:21 crc kubenswrapper[4632]: > pod="openshift-infra/auto-csr-approver-29556610-sg5bx" Mar 13 10:10:21 crc kubenswrapper[4632]: E0313 10:10:21.467456 4632 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err=< Mar 13 10:10:21 crc kubenswrapper[4632]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29556610-sg5bx_openshift-infra_795727b7-7a2e-4e97-8707-aecf893fd332_0(f44c58a66c3c09c74b677cf4ec9f76b205bf4b131d95b02c4a9987dbc9e9bd8a): error adding pod openshift-infra_auto-csr-approver-29556610-sg5bx to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"f44c58a66c3c09c74b677cf4ec9f76b205bf4b131d95b02c4a9987dbc9e9bd8a" Netns:"/var/run/netns/731b0fe9-200c-49e0-9ae2-97aa60457e8f" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-infra;K8S_POD_NAME=auto-csr-approver-29556610-sg5bx;K8S_POD_INFRA_CONTAINER_ID=f44c58a66c3c09c74b677cf4ec9f76b205bf4b131d95b02c4a9987dbc9e9bd8a;K8S_POD_UID=795727b7-7a2e-4e97-8707-aecf893fd332" Path:"" ERRORED: error configuring pod [openshift-infra/auto-csr-approver-29556610-sg5bx] networking: Multus: [openshift-infra/auto-csr-approver-29556610-sg5bx/795727b7-7a2e-4e97-8707-aecf893fd332]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod auto-csr-approver-29556610-sg5bx in out of cluster comm: pod "auto-csr-approver-29556610-sg5bx" not found Mar 13 10:10:21 crc kubenswrapper[4632]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 10:10:21 crc kubenswrapper[4632]: > pod="openshift-infra/auto-csr-approver-29556610-sg5bx" Mar 13 10:10:21 crc kubenswrapper[4632]: E0313 10:10:21.467552 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29556610-sg5bx_openshift-infra(795727b7-7a2e-4e97-8707-aecf893fd332)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29556610-sg5bx_openshift-infra(795727b7-7a2e-4e97-8707-aecf893fd332)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29556610-sg5bx_openshift-infra_795727b7-7a2e-4e97-8707-aecf893fd332_0(f44c58a66c3c09c74b677cf4ec9f76b205bf4b131d95b02c4a9987dbc9e9bd8a): error adding pod openshift-infra_auto-csr-approver-29556610-sg5bx to CNI network \\\"multus-cni-network\\\": plugin type=\\\"multus-shim\\\" name=\\\"multus-cni-network\\\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\\\"f44c58a66c3c09c74b677cf4ec9f76b205bf4b131d95b02c4a9987dbc9e9bd8a\\\" Netns:\\\"/var/run/netns/731b0fe9-200c-49e0-9ae2-97aa60457e8f\\\" IfName:\\\"eth0\\\" Args:\\\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-infra;K8S_POD_NAME=auto-csr-approver-29556610-sg5bx;K8S_POD_INFRA_CONTAINER_ID=f44c58a66c3c09c74b677cf4ec9f76b205bf4b131d95b02c4a9987dbc9e9bd8a;K8S_POD_UID=795727b7-7a2e-4e97-8707-aecf893fd332\\\" Path:\\\"\\\" ERRORED: error configuring pod [openshift-infra/auto-csr-approver-29556610-sg5bx] networking: Multus: [openshift-infra/auto-csr-approver-29556610-sg5bx/795727b7-7a2e-4e97-8707-aecf893fd332]: error setting the networks status, pod was already deleted: SetPodNetworkStatusAnnotation: failed to query the pod auto-csr-approver-29556610-sg5bx in out of cluster comm: pod \\\"auto-csr-approver-29556610-sg5bx\\\" not found\\n': StdinData: {\\\"binDir\\\":\\\"/var/lib/cni/bin\\\",\\\"clusterNetwork\\\":\\\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\\\",\\\"cniVersion\\\":\\\"0.3.1\\\",\\\"daemonSocketDir\\\":\\\"/run/multus/socket\\\",\\\"globalNamespaces\\\":\\\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\\\",\\\"logLevel\\\":\\\"verbose\\\",\\\"logToStderr\\\":true,\\\"name\\\":\\\"multus-cni-network\\\",\\\"namespaceIsolation\\\":true,\\\"type\\\":\\\"multus-shim\\\"}\"" pod="openshift-infra/auto-csr-approver-29556610-sg5bx" podUID="795727b7-7a2e-4e97-8707-aecf893fd332" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.536794 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.558525 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.567434 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.624440 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.644349 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.690442 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.709858 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.737621 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.791064 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.855916 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.882166 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.893332 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556610-sg5bx" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.894620 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556610-sg5bx" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.932394 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.944588 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 13 10:10:21 crc kubenswrapper[4632]: I0313 10:10:21.949076 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 10:10:22 crc kubenswrapper[4632]: I0313 10:10:22.005231 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 13 10:10:22 crc kubenswrapper[4632]: I0313 10:10:22.058082 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 13 10:10:22 crc kubenswrapper[4632]: I0313 10:10:22.112706 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 13 10:10:22 crc kubenswrapper[4632]: I0313 10:10:22.295215 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 13 10:10:22 crc kubenswrapper[4632]: I0313 10:10:22.432358 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 13 10:10:22 crc kubenswrapper[4632]: I0313 10:10:22.628659 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 13 10:10:22 crc kubenswrapper[4632]: I0313 10:10:22.676903 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 13 10:10:22 crc kubenswrapper[4632]: I0313 10:10:22.687809 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 13 10:10:22 crc kubenswrapper[4632]: I0313 10:10:22.704980 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 13 10:10:22 crc kubenswrapper[4632]: I0313 10:10:22.748458 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 13 10:10:22 crc kubenswrapper[4632]: I0313 10:10:22.898560 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 13 10:10:23 crc kubenswrapper[4632]: I0313 10:10:23.015996 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 13 10:10:23 crc kubenswrapper[4632]: I0313 10:10:23.165505 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 13 10:10:23 crc kubenswrapper[4632]: I0313 10:10:23.268654 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 13 10:10:23 crc kubenswrapper[4632]: I0313 10:10:23.300148 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 13 10:10:23 crc kubenswrapper[4632]: I0313 10:10:23.369241 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 13 10:10:23 crc kubenswrapper[4632]: I0313 10:10:23.669832 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 13 10:10:23 crc kubenswrapper[4632]: I0313 10:10:23.764464 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 13 10:10:23 crc kubenswrapper[4632]: I0313 10:10:23.852190 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 13 10:10:23 crc kubenswrapper[4632]: I0313 10:10:23.876113 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 13 10:10:24 crc kubenswrapper[4632]: I0313 10:10:24.051034 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 13 10:10:24 crc kubenswrapper[4632]: I0313 10:10:24.060739 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 10:10:24 crc kubenswrapper[4632]: I0313 10:10:24.121973 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 13 10:10:24 crc kubenswrapper[4632]: I0313 10:10:24.140001 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 13 10:10:24 crc kubenswrapper[4632]: I0313 10:10:24.144604 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 13 10:10:24 crc kubenswrapper[4632]: I0313 10:10:24.264838 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 13 10:10:24 crc kubenswrapper[4632]: I0313 10:10:24.321206 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556610-sg5bx"] Mar 13 10:10:24 crc kubenswrapper[4632]: I0313 10:10:24.331265 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 13 10:10:24 crc kubenswrapper[4632]: I0313 10:10:24.446900 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 13 10:10:24 crc kubenswrapper[4632]: I0313 10:10:24.539882 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 13 10:10:24 crc kubenswrapper[4632]: I0313 10:10:24.911572 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556610-sg5bx" event={"ID":"795727b7-7a2e-4e97-8707-aecf893fd332","Type":"ContainerStarted","Data":"10a2d90beda673bcf86de10163ece976f67cb0baa2a32060af9801525b02c6e6"} Mar 13 10:10:25 crc kubenswrapper[4632]: I0313 10:10:25.296863 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 13 10:10:25 crc kubenswrapper[4632]: I0313 10:10:25.918183 4632 generic.go:334] "Generic (PLEG): container finished" podID="795727b7-7a2e-4e97-8707-aecf893fd332" containerID="2858228a654d1c5c1b9a9a04d00ea882bfe929e6c810389040bc3c0ba67d7a46" exitCode=0 Mar 13 10:10:25 crc kubenswrapper[4632]: I0313 10:10:25.918379 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556610-sg5bx" event={"ID":"795727b7-7a2e-4e97-8707-aecf893fd332","Type":"ContainerDied","Data":"2858228a654d1c5c1b9a9a04d00ea882bfe929e6c810389040bc3c0ba67d7a46"} Mar 13 10:10:26 crc kubenswrapper[4632]: I0313 10:10:26.927269 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 13 10:10:26 crc kubenswrapper[4632]: I0313 10:10:26.927612 4632 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="80143265ebaee0f3a54053c9c203e48c0b8ae49b675b972452da2268c99bd9ad" exitCode=137 Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.008032 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.008154 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.105504 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.105589 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.105622 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.105753 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.105786 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.105844 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.105897 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.105978 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.105844 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.106240 4632 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.106268 4632 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.106337 4632 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.106349 4632 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.116808 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.167688 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556610-sg5bx" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.206959 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmrds\" (UniqueName: \"kubernetes.io/projected/795727b7-7a2e-4e97-8707-aecf893fd332-kube-api-access-tmrds\") pod \"795727b7-7a2e-4e97-8707-aecf893fd332\" (UID: \"795727b7-7a2e-4e97-8707-aecf893fd332\") " Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.207169 4632 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.211321 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/795727b7-7a2e-4e97-8707-aecf893fd332-kube-api-access-tmrds" (OuterVolumeSpecName: "kube-api-access-tmrds") pod "795727b7-7a2e-4e97-8707-aecf893fd332" (UID: "795727b7-7a2e-4e97-8707-aecf893fd332"). InnerVolumeSpecName "kube-api-access-tmrds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.309313 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmrds\" (UniqueName: \"kubernetes.io/projected/795727b7-7a2e-4e97-8707-aecf893fd332-kube-api-access-tmrds\") on node \"crc\" DevicePath \"\"" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.938427 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.938588 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.938677 4632 scope.go:117] "RemoveContainer" containerID="80143265ebaee0f3a54053c9c203e48c0b8ae49b675b972452da2268c99bd9ad" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.944640 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556610-sg5bx" event={"ID":"795727b7-7a2e-4e97-8707-aecf893fd332","Type":"ContainerDied","Data":"10a2d90beda673bcf86de10163ece976f67cb0baa2a32060af9801525b02c6e6"} Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.944690 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10a2d90beda673bcf86de10163ece976f67cb0baa2a32060af9801525b02c6e6" Mar 13 10:10:27 crc kubenswrapper[4632]: I0313 10:10:27.944750 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556610-sg5bx" Mar 13 10:10:28 crc kubenswrapper[4632]: I0313 10:10:28.053817 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 13 10:10:28 crc kubenswrapper[4632]: I0313 10:10:28.054248 4632 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Mar 13 10:10:28 crc kubenswrapper[4632]: I0313 10:10:28.066070 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 10:10:28 crc kubenswrapper[4632]: I0313 10:10:28.066116 4632 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="79d3accd-e208-422a-b26e-bb6023b74edf" Mar 13 10:10:28 crc kubenswrapper[4632]: I0313 10:10:28.068873 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 10:10:28 crc kubenswrapper[4632]: I0313 10:10:28.068901 4632 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="79d3accd-e208-422a-b26e-bb6023b74edf" Mar 13 10:10:36 crc kubenswrapper[4632]: I0313 10:10:36.595132 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 13 10:10:38 crc kubenswrapper[4632]: I0313 10:10:38.064972 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 13 10:10:41 crc kubenswrapper[4632]: I0313 10:10:41.177826 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 13 10:10:43 crc kubenswrapper[4632]: I0313 10:10:43.222682 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 13 10:10:46 crc kubenswrapper[4632]: I0313 10:10:46.840237 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.080239 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t6bkt"] Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.081912 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-t6bkt" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" containerName="registry-server" containerID="cri-o://18dc7d50bd20cdcc8bd750dbb90186346da79c165a2b1398fb9fce2a58fd1224" gracePeriod=2 Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.280054 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xr5l9"] Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.280374 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xr5l9" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" containerName="registry-server" containerID="cri-o://f3ec0c7ec6e706f7309cf337b9ffdf829c043fe63c66f9db4fb7e26f23ae6edd" gracePeriod=2 Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.535550 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.621419 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-catalog-content\") pod \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\" (UID: \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\") " Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.621486 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-utilities\") pod \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\" (UID: \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\") " Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.621540 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bc6ff\" (UniqueName: \"kubernetes.io/projected/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-kube-api-access-bc6ff\") pod \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\" (UID: \"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e\") " Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.624007 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-utilities" (OuterVolumeSpecName: "utilities") pod "668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" (UID: "668c4640-0e5f-4c98-8b6e-dbdffdbfe14e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.630390 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-kube-api-access-bc6ff" (OuterVolumeSpecName: "kube-api-access-bc6ff") pod "668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" (UID: "668c4640-0e5f-4c98-8b6e-dbdffdbfe14e"). InnerVolumeSpecName "kube-api-access-bc6ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.647838 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" (UID: "668c4640-0e5f-4c98-8b6e-dbdffdbfe14e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.664837 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.722758 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87965e39-b879-4e26-9c8b-b78068c52aa0-utilities\") pod \"87965e39-b879-4e26-9c8b-b78068c52aa0\" (UID: \"87965e39-b879-4e26-9c8b-b78068c52aa0\") " Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.722824 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hfz2\" (UniqueName: \"kubernetes.io/projected/87965e39-b879-4e26-9c8b-b78068c52aa0-kube-api-access-6hfz2\") pod \"87965e39-b879-4e26-9c8b-b78068c52aa0\" (UID: \"87965e39-b879-4e26-9c8b-b78068c52aa0\") " Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.722897 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87965e39-b879-4e26-9c8b-b78068c52aa0-catalog-content\") pod \"87965e39-b879-4e26-9c8b-b78068c52aa0\" (UID: \"87965e39-b879-4e26-9c8b-b78068c52aa0\") " Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.723212 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bc6ff\" (UniqueName: \"kubernetes.io/projected/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-kube-api-access-bc6ff\") on node \"crc\" DevicePath \"\"" Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.723238 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.723249 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.723881 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87965e39-b879-4e26-9c8b-b78068c52aa0-utilities" (OuterVolumeSpecName: "utilities") pod "87965e39-b879-4e26-9c8b-b78068c52aa0" (UID: "87965e39-b879-4e26-9c8b-b78068c52aa0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.728154 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87965e39-b879-4e26-9c8b-b78068c52aa0-kube-api-access-6hfz2" (OuterVolumeSpecName: "kube-api-access-6hfz2") pod "87965e39-b879-4e26-9c8b-b78068c52aa0" (UID: "87965e39-b879-4e26-9c8b-b78068c52aa0"). InnerVolumeSpecName "kube-api-access-6hfz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.824697 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/87965e39-b879-4e26-9c8b-b78068c52aa0-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.824743 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hfz2\" (UniqueName: \"kubernetes.io/projected/87965e39-b879-4e26-9c8b-b78068c52aa0-kube-api-access-6hfz2\") on node \"crc\" DevicePath \"\"" Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.875893 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87965e39-b879-4e26-9c8b-b78068c52aa0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "87965e39-b879-4e26-9c8b-b78068c52aa0" (UID: "87965e39-b879-4e26-9c8b-b78068c52aa0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:10:51 crc kubenswrapper[4632]: I0313 10:10:51.926719 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/87965e39-b879-4e26-9c8b-b78068c52aa0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.126054 4632 generic.go:334] "Generic (PLEG): container finished" podID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" containerID="18dc7d50bd20cdcc8bd750dbb90186346da79c165a2b1398fb9fce2a58fd1224" exitCode=0 Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.126168 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6bkt" event={"ID":"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e","Type":"ContainerDied","Data":"18dc7d50bd20cdcc8bd750dbb90186346da79c165a2b1398fb9fce2a58fd1224"} Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.126217 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-t6bkt" event={"ID":"668c4640-0e5f-4c98-8b6e-dbdffdbfe14e","Type":"ContainerDied","Data":"3e3a79d99a0e6a35edab86938ccf523a35c4606e460775b549d1924f20dc4204"} Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.126237 4632 scope.go:117] "RemoveContainer" containerID="18dc7d50bd20cdcc8bd750dbb90186346da79c165a2b1398fb9fce2a58fd1224" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.126205 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-t6bkt" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.128469 4632 generic.go:334] "Generic (PLEG): container finished" podID="87965e39-b879-4e26-9c8b-b78068c52aa0" containerID="f3ec0c7ec6e706f7309cf337b9ffdf829c043fe63c66f9db4fb7e26f23ae6edd" exitCode=0 Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.128495 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr5l9" event={"ID":"87965e39-b879-4e26-9c8b-b78068c52aa0","Type":"ContainerDied","Data":"f3ec0c7ec6e706f7309cf337b9ffdf829c043fe63c66f9db4fb7e26f23ae6edd"} Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.128513 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xr5l9" event={"ID":"87965e39-b879-4e26-9c8b-b78068c52aa0","Type":"ContainerDied","Data":"9d37b680fdc1d8687e48df9dab9cd8ad8fcee9b7cdb15c920f34a9cbf7bad5ef"} Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.128574 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xr5l9" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.175906 4632 scope.go:117] "RemoveContainer" containerID="9fa8755799160674c9b9254d4cc4cd33b06b805f90944344a392116f94021c66" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.208647 4632 scope.go:117] "RemoveContainer" containerID="01ae56b596391f3b7877c67539058596dbfd086754ed8db1f4f40f76d82a4c4c" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.215199 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xr5l9"] Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.218875 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xr5l9"] Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.233451 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-t6bkt"] Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.238138 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-t6bkt"] Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.253812 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.274675 4632 scope.go:117] "RemoveContainer" containerID="18dc7d50bd20cdcc8bd750dbb90186346da79c165a2b1398fb9fce2a58fd1224" Mar 13 10:10:52 crc kubenswrapper[4632]: E0313 10:10:52.275445 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18dc7d50bd20cdcc8bd750dbb90186346da79c165a2b1398fb9fce2a58fd1224\": container with ID starting with 18dc7d50bd20cdcc8bd750dbb90186346da79c165a2b1398fb9fce2a58fd1224 not found: ID does not exist" containerID="18dc7d50bd20cdcc8bd750dbb90186346da79c165a2b1398fb9fce2a58fd1224" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.275504 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18dc7d50bd20cdcc8bd750dbb90186346da79c165a2b1398fb9fce2a58fd1224"} err="failed to get container status \"18dc7d50bd20cdcc8bd750dbb90186346da79c165a2b1398fb9fce2a58fd1224\": rpc error: code = NotFound desc = could not find container \"18dc7d50bd20cdcc8bd750dbb90186346da79c165a2b1398fb9fce2a58fd1224\": container with ID starting with 18dc7d50bd20cdcc8bd750dbb90186346da79c165a2b1398fb9fce2a58fd1224 not found: ID does not exist" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.275539 4632 scope.go:117] "RemoveContainer" containerID="9fa8755799160674c9b9254d4cc4cd33b06b805f90944344a392116f94021c66" Mar 13 10:10:52 crc kubenswrapper[4632]: E0313 10:10:52.275912 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9fa8755799160674c9b9254d4cc4cd33b06b805f90944344a392116f94021c66\": container with ID starting with 9fa8755799160674c9b9254d4cc4cd33b06b805f90944344a392116f94021c66 not found: ID does not exist" containerID="9fa8755799160674c9b9254d4cc4cd33b06b805f90944344a392116f94021c66" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.276027 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9fa8755799160674c9b9254d4cc4cd33b06b805f90944344a392116f94021c66"} err="failed to get container status \"9fa8755799160674c9b9254d4cc4cd33b06b805f90944344a392116f94021c66\": rpc error: code = NotFound desc = could not find container \"9fa8755799160674c9b9254d4cc4cd33b06b805f90944344a392116f94021c66\": container with ID starting with 9fa8755799160674c9b9254d4cc4cd33b06b805f90944344a392116f94021c66 not found: ID does not exist" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.276059 4632 scope.go:117] "RemoveContainer" containerID="01ae56b596391f3b7877c67539058596dbfd086754ed8db1f4f40f76d82a4c4c" Mar 13 10:10:52 crc kubenswrapper[4632]: E0313 10:10:52.276351 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01ae56b596391f3b7877c67539058596dbfd086754ed8db1f4f40f76d82a4c4c\": container with ID starting with 01ae56b596391f3b7877c67539058596dbfd086754ed8db1f4f40f76d82a4c4c not found: ID does not exist" containerID="01ae56b596391f3b7877c67539058596dbfd086754ed8db1f4f40f76d82a4c4c" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.276378 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01ae56b596391f3b7877c67539058596dbfd086754ed8db1f4f40f76d82a4c4c"} err="failed to get container status \"01ae56b596391f3b7877c67539058596dbfd086754ed8db1f4f40f76d82a4c4c\": rpc error: code = NotFound desc = could not find container \"01ae56b596391f3b7877c67539058596dbfd086754ed8db1f4f40f76d82a4c4c\": container with ID starting with 01ae56b596391f3b7877c67539058596dbfd086754ed8db1f4f40f76d82a4c4c not found: ID does not exist" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.276398 4632 scope.go:117] "RemoveContainer" containerID="f3ec0c7ec6e706f7309cf337b9ffdf829c043fe63c66f9db4fb7e26f23ae6edd" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.293427 4632 scope.go:117] "RemoveContainer" containerID="e3753ece91188912f076d532ed434a938805848842e1fdb1b100800bcaa42763" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.310741 4632 scope.go:117] "RemoveContainer" containerID="bafd48184b1b00528329793cfe1af87f0aa9502582cddad85d784407e60c249d" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.330906 4632 scope.go:117] "RemoveContainer" containerID="f3ec0c7ec6e706f7309cf337b9ffdf829c043fe63c66f9db4fb7e26f23ae6edd" Mar 13 10:10:52 crc kubenswrapper[4632]: E0313 10:10:52.331411 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3ec0c7ec6e706f7309cf337b9ffdf829c043fe63c66f9db4fb7e26f23ae6edd\": container with ID starting with f3ec0c7ec6e706f7309cf337b9ffdf829c043fe63c66f9db4fb7e26f23ae6edd not found: ID does not exist" containerID="f3ec0c7ec6e706f7309cf337b9ffdf829c043fe63c66f9db4fb7e26f23ae6edd" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.331439 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3ec0c7ec6e706f7309cf337b9ffdf829c043fe63c66f9db4fb7e26f23ae6edd"} err="failed to get container status \"f3ec0c7ec6e706f7309cf337b9ffdf829c043fe63c66f9db4fb7e26f23ae6edd\": rpc error: code = NotFound desc = could not find container \"f3ec0c7ec6e706f7309cf337b9ffdf829c043fe63c66f9db4fb7e26f23ae6edd\": container with ID starting with f3ec0c7ec6e706f7309cf337b9ffdf829c043fe63c66f9db4fb7e26f23ae6edd not found: ID does not exist" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.331459 4632 scope.go:117] "RemoveContainer" containerID="e3753ece91188912f076d532ed434a938805848842e1fdb1b100800bcaa42763" Mar 13 10:10:52 crc kubenswrapper[4632]: E0313 10:10:52.331721 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3753ece91188912f076d532ed434a938805848842e1fdb1b100800bcaa42763\": container with ID starting with e3753ece91188912f076d532ed434a938805848842e1fdb1b100800bcaa42763 not found: ID does not exist" containerID="e3753ece91188912f076d532ed434a938805848842e1fdb1b100800bcaa42763" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.331762 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3753ece91188912f076d532ed434a938805848842e1fdb1b100800bcaa42763"} err="failed to get container status \"e3753ece91188912f076d532ed434a938805848842e1fdb1b100800bcaa42763\": rpc error: code = NotFound desc = could not find container \"e3753ece91188912f076d532ed434a938805848842e1fdb1b100800bcaa42763\": container with ID starting with e3753ece91188912f076d532ed434a938805848842e1fdb1b100800bcaa42763 not found: ID does not exist" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.331788 4632 scope.go:117] "RemoveContainer" containerID="bafd48184b1b00528329793cfe1af87f0aa9502582cddad85d784407e60c249d" Mar 13 10:10:52 crc kubenswrapper[4632]: E0313 10:10:52.332097 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bafd48184b1b00528329793cfe1af87f0aa9502582cddad85d784407e60c249d\": container with ID starting with bafd48184b1b00528329793cfe1af87f0aa9502582cddad85d784407e60c249d not found: ID does not exist" containerID="bafd48184b1b00528329793cfe1af87f0aa9502582cddad85d784407e60c249d" Mar 13 10:10:52 crc kubenswrapper[4632]: I0313 10:10:52.332121 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bafd48184b1b00528329793cfe1af87f0aa9502582cddad85d784407e60c249d"} err="failed to get container status \"bafd48184b1b00528329793cfe1af87f0aa9502582cddad85d784407e60c249d\": rpc error: code = NotFound desc = could not find container \"bafd48184b1b00528329793cfe1af87f0aa9502582cddad85d784407e60c249d\": container with ID starting with bafd48184b1b00528329793cfe1af87f0aa9502582cddad85d784407e60c249d not found: ID does not exist" Mar 13 10:10:54 crc kubenswrapper[4632]: I0313 10:10:54.056130 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" path="/var/lib/kubelet/pods/668c4640-0e5f-4c98-8b6e-dbdffdbfe14e/volumes" Mar 13 10:10:54 crc kubenswrapper[4632]: I0313 10:10:54.057820 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" path="/var/lib/kubelet/pods/87965e39-b879-4e26-9c8b-b78068c52aa0/volumes" Mar 13 10:10:54 crc kubenswrapper[4632]: I0313 10:10:54.573634 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8sl88"] Mar 13 10:11:10 crc kubenswrapper[4632]: I0313 10:11:10.461236 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:11:10 crc kubenswrapper[4632]: I0313 10:11:10.463163 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:11:19 crc kubenswrapper[4632]: I0313 10:11:19.599728 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" podUID="f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" containerName="oauth-openshift" containerID="cri-o://7595f698758f3f9ece4af82a45628aeb01bfa58cde4c80fefdeaf746f39aba12" gracePeriod=15 Mar 13 10:11:19 crc kubenswrapper[4632]: I0313 10:11:19.997931 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034168 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x"] Mar 13 10:11:20 crc kubenswrapper[4632]: E0313 10:11:20.034447 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" containerName="extract-content" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034460 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" containerName="extract-content" Mar 13 10:11:20 crc kubenswrapper[4632]: E0313 10:11:20.034473 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" containerName="registry-server" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034480 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" containerName="registry-server" Mar 13 10:11:20 crc kubenswrapper[4632]: E0313 10:11:20.034490 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" containerName="extract-utilities" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034497 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" containerName="extract-utilities" Mar 13 10:11:20 crc kubenswrapper[4632]: E0313 10:11:20.034507 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" containerName="extract-content" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034514 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" containerName="extract-content" Mar 13 10:11:20 crc kubenswrapper[4632]: E0313 10:11:20.034524 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" containerName="oauth-openshift" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034530 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" containerName="oauth-openshift" Mar 13 10:11:20 crc kubenswrapper[4632]: E0313 10:11:20.034541 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="795727b7-7a2e-4e97-8707-aecf893fd332" containerName="oc" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034547 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="795727b7-7a2e-4e97-8707-aecf893fd332" containerName="oc" Mar 13 10:11:20 crc kubenswrapper[4632]: E0313 10:11:20.034558 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" containerName="extract-utilities" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034564 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" containerName="extract-utilities" Mar 13 10:11:20 crc kubenswrapper[4632]: E0313 10:11:20.034571 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" containerName="registry-server" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034593 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" containerName="registry-server" Mar 13 10:11:20 crc kubenswrapper[4632]: E0313 10:11:20.034607 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034614 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034724 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" containerName="oauth-openshift" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034749 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="87965e39-b879-4e26-9c8b-b78068c52aa0" containerName="registry-server" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034759 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="795727b7-7a2e-4e97-8707-aecf893fd332" containerName="oc" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034769 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="668c4640-0e5f-4c98-8b6e-dbdffdbfe14e" containerName="registry-server" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.034782 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.035214 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.064884 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x"] Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.125825 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-ocp-branding-template\") pod \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.125882 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-session\") pod \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.125913 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-provider-selection\") pod \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.125957 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-idp-0-file-data\") pod \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.125988 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-audit-dir\") pod \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.126011 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-error\") pod \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.126053 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-cliconfig\") pod \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.126092 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-router-certs\") pod \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.126115 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nsdx\" (UniqueName: \"kubernetes.io/projected/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-kube-api-access-9nsdx\") pod \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.126136 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-login\") pod \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.126158 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-trusted-ca-bundle\") pod \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.126196 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-service-ca\") pod \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.126216 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-audit-policies\") pod \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.126242 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-serving-cert\") pod \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\" (UID: \"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a\") " Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.127835 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" (UID: "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.127901 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" (UID: "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.127964 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" (UID: "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.128789 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" (UID: "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.132553 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-kube-api-access-9nsdx" (OuterVolumeSpecName: "kube-api-access-9nsdx") pod "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" (UID: "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a"). InnerVolumeSpecName "kube-api-access-9nsdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.132620 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" (UID: "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.133317 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" (UID: "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.136909 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" (UID: "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.137594 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" (UID: "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.140295 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" (UID: "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.140908 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" (UID: "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.143091 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" (UID: "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.145101 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" (UID: "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.145839 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" (UID: "f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227140 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227194 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227213 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9sch\" (UniqueName: \"kubernetes.io/projected/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-kube-api-access-q9sch\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227244 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-session\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227260 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-user-template-login\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227278 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-audit-policies\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227310 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227331 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227349 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-service-ca\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227369 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-router-certs\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227384 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-user-template-error\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227398 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227414 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-audit-dir\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227430 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227464 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227474 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227484 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227494 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227504 4632 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227513 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227521 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227529 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227538 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nsdx\" (UniqueName: \"kubernetes.io/projected/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-kube-api-access-9nsdx\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227546 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227557 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227567 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227577 4632 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.227585 4632 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.271240 4632 generic.go:334] "Generic (PLEG): container finished" podID="f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" containerID="7595f698758f3f9ece4af82a45628aeb01bfa58cde4c80fefdeaf746f39aba12" exitCode=0 Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.271284 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" event={"ID":"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a","Type":"ContainerDied","Data":"7595f698758f3f9ece4af82a45628aeb01bfa58cde4c80fefdeaf746f39aba12"} Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.271310 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" event={"ID":"f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a","Type":"ContainerDied","Data":"46d5cd8b5a8d1e4d5e145a625b40cd39a2bdcba910908f1195bf38b9cf2ad7c8"} Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.271327 4632 scope.go:117] "RemoveContainer" containerID="7595f698758f3f9ece4af82a45628aeb01bfa58cde4c80fefdeaf746f39aba12" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.271420 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8sl88" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.294047 4632 scope.go:117] "RemoveContainer" containerID="7595f698758f3f9ece4af82a45628aeb01bfa58cde4c80fefdeaf746f39aba12" Mar 13 10:11:20 crc kubenswrapper[4632]: E0313 10:11:20.294628 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7595f698758f3f9ece4af82a45628aeb01bfa58cde4c80fefdeaf746f39aba12\": container with ID starting with 7595f698758f3f9ece4af82a45628aeb01bfa58cde4c80fefdeaf746f39aba12 not found: ID does not exist" containerID="7595f698758f3f9ece4af82a45628aeb01bfa58cde4c80fefdeaf746f39aba12" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.294663 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7595f698758f3f9ece4af82a45628aeb01bfa58cde4c80fefdeaf746f39aba12"} err="failed to get container status \"7595f698758f3f9ece4af82a45628aeb01bfa58cde4c80fefdeaf746f39aba12\": rpc error: code = NotFound desc = could not find container \"7595f698758f3f9ece4af82a45628aeb01bfa58cde4c80fefdeaf746f39aba12\": container with ID starting with 7595f698758f3f9ece4af82a45628aeb01bfa58cde4c80fefdeaf746f39aba12 not found: ID does not exist" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.307475 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8sl88"] Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.311449 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8sl88"] Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.329141 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.329212 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.329237 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9sch\" (UniqueName: \"kubernetes.io/projected/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-kube-api-access-q9sch\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.329318 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-session\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.329406 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-user-template-login\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.329450 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-audit-policies\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.329482 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.329526 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.329554 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-service-ca\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.329603 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-router-certs\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.329621 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-user-template-error\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.329638 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.329674 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-audit-dir\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.329694 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.330482 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-audit-dir\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.332183 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-service-ca\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.332197 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-audit-policies\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.332975 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.335057 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-cliconfig\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.336853 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.339237 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-user-template-error\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.339561 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-serving-cert\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.339562 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-router-certs\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.340582 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.342098 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-user-template-login\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.342433 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.343207 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-v4-0-config-system-session\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.347048 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9sch\" (UniqueName: \"kubernetes.io/projected/48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd-kube-api-access-q9sch\") pod \"oauth-openshift-75bb75cfd7-8sh2x\" (UID: \"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd\") " pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.364013 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:20 crc kubenswrapper[4632]: I0313 10:11:20.759205 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x"] Mar 13 10:11:20 crc kubenswrapper[4632]: W0313 10:11:20.767378 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48d2bc7e_c929_42c9_b3f2_9e78c7eac8cd.slice/crio-9668ce6a02f8a0451908f7868d18802f5139657e99d614dabe929ff79634f8e7 WatchSource:0}: Error finding container 9668ce6a02f8a0451908f7868d18802f5139657e99d614dabe929ff79634f8e7: Status 404 returned error can't find the container with id 9668ce6a02f8a0451908f7868d18802f5139657e99d614dabe929ff79634f8e7 Mar 13 10:11:21 crc kubenswrapper[4632]: I0313 10:11:21.276587 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" event={"ID":"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd","Type":"ContainerStarted","Data":"224837f104bcdbc6545d62209161e349a9d07cdcaf5c66e47c1de75b3af4b369"} Mar 13 10:11:21 crc kubenswrapper[4632]: I0313 10:11:21.277436 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:21 crc kubenswrapper[4632]: I0313 10:11:21.277456 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" event={"ID":"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd","Type":"ContainerStarted","Data":"9668ce6a02f8a0451908f7868d18802f5139657e99d614dabe929ff79634f8e7"} Mar 13 10:11:21 crc kubenswrapper[4632]: I0313 10:11:21.299139 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" podStartSLOduration=27.299120215 podStartE2EDuration="27.299120215s" podCreationTimestamp="2026-03-13 10:10:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:11:21.296162827 +0000 UTC m=+455.318692960" watchObservedRunningTime="2026-03-13 10:11:21.299120215 +0000 UTC m=+455.321650388" Mar 13 10:11:21 crc kubenswrapper[4632]: I0313 10:11:21.331693 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 10:11:22 crc kubenswrapper[4632]: I0313 10:11:22.052304 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a" path="/var/lib/kubelet/pods/f6f7bdd0-1aaa-48f1-a3b7-55bd145aec0a/volumes" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.420822 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-sbzcm"] Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.422348 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.442052 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-sbzcm"] Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.534228 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/121318f2-259b-4187-a348-1282d0e63995-trusted-ca\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.534483 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/121318f2-259b-4187-a348-1282d0e63995-ca-trust-extracted\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.534523 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/121318f2-259b-4187-a348-1282d0e63995-installation-pull-secrets\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.534550 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/121318f2-259b-4187-a348-1282d0e63995-registry-tls\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.534586 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.534684 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/121318f2-259b-4187-a348-1282d0e63995-bound-sa-token\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.534749 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/121318f2-259b-4187-a348-1282d0e63995-registry-certificates\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.534784 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zznsv\" (UniqueName: \"kubernetes.io/projected/121318f2-259b-4187-a348-1282d0e63995-kube-api-access-zznsv\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.586345 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.636116 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/121318f2-259b-4187-a348-1282d0e63995-registry-certificates\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.636164 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/121318f2-259b-4187-a348-1282d0e63995-bound-sa-token\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.636517 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zznsv\" (UniqueName: \"kubernetes.io/projected/121318f2-259b-4187-a348-1282d0e63995-kube-api-access-zznsv\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.636734 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/121318f2-259b-4187-a348-1282d0e63995-trusted-ca\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.637781 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/121318f2-259b-4187-a348-1282d0e63995-registry-certificates\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.638196 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/121318f2-259b-4187-a348-1282d0e63995-trusted-ca\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.636767 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/121318f2-259b-4187-a348-1282d0e63995-ca-trust-extracted\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.638302 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/121318f2-259b-4187-a348-1282d0e63995-installation-pull-secrets\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.638335 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/121318f2-259b-4187-a348-1282d0e63995-registry-tls\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.638334 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/121318f2-259b-4187-a348-1282d0e63995-ca-trust-extracted\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.644591 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/121318f2-259b-4187-a348-1282d0e63995-registry-tls\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.646691 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/121318f2-259b-4187-a348-1282d0e63995-installation-pull-secrets\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.654670 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zznsv\" (UniqueName: \"kubernetes.io/projected/121318f2-259b-4187-a348-1282d0e63995-kube-api-access-zznsv\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.658645 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/121318f2-259b-4187-a348-1282d0e63995-bound-sa-token\") pod \"image-registry-66df7c8f76-sbzcm\" (UID: \"121318f2-259b-4187-a348-1282d0e63995\") " pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:36 crc kubenswrapper[4632]: I0313 10:11:36.740197 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:37 crc kubenswrapper[4632]: I0313 10:11:37.237820 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-sbzcm"] Mar 13 10:11:37 crc kubenswrapper[4632]: W0313 10:11:37.249070 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod121318f2_259b_4187_a348_1282d0e63995.slice/crio-194c14a843a5324b5d9fa98672e41df35bc8eea13219d905d3a50858030e3537 WatchSource:0}: Error finding container 194c14a843a5324b5d9fa98672e41df35bc8eea13219d905d3a50858030e3537: Status 404 returned error can't find the container with id 194c14a843a5324b5d9fa98672e41df35bc8eea13219d905d3a50858030e3537 Mar 13 10:11:37 crc kubenswrapper[4632]: I0313 10:11:37.383577 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" event={"ID":"121318f2-259b-4187-a348-1282d0e63995","Type":"ContainerStarted","Data":"194c14a843a5324b5d9fa98672e41df35bc8eea13219d905d3a50858030e3537"} Mar 13 10:11:38 crc kubenswrapper[4632]: I0313 10:11:38.391449 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" event={"ID":"121318f2-259b-4187-a348-1282d0e63995","Type":"ContainerStarted","Data":"d8762342ae6795e17a6078aa65bfe76494086e447cc848fc87a90a6d426b1bdf"} Mar 13 10:11:38 crc kubenswrapper[4632]: I0313 10:11:38.392883 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:38 crc kubenswrapper[4632]: I0313 10:11:38.417629 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" podStartSLOduration=2.417611691 podStartE2EDuration="2.417611691s" podCreationTimestamp="2026-03-13 10:11:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:11:38.41494862 +0000 UTC m=+472.437478753" watchObservedRunningTime="2026-03-13 10:11:38.417611691 +0000 UTC m=+472.440141824" Mar 13 10:11:40 crc kubenswrapper[4632]: I0313 10:11:40.461092 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:11:40 crc kubenswrapper[4632]: I0313 10:11:40.463057 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.028074 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p8wjg"] Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.028587 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p8wjg" podUID="b11a7dff-bf08-44c3-b4f4-923119c13717" containerName="registry-server" containerID="cri-o://55bfc00a5732a457ecbee5c7be945027bdb42c0137a6b22125d44dafb5924f59" gracePeriod=30 Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.041847 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jvh86"] Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.042232 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jvh86" podUID="bd46ae04-0610-4aa5-9385-dd45de66c5dd" containerName="registry-server" containerID="cri-o://5a10aa8d51646d1f515364874b0426c82d85f03f52a4924f31299cb0395b0607" gracePeriod=30 Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.077289 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2n99d"] Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.077697 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" podUID="797176c6-dd56-48d6-8004-ff1dd5353a50" containerName="marketplace-operator" containerID="cri-o://0fd5f07ae3c28f8c24cc66a585de93acc08f170fb621bbeb190cd66596980871" gracePeriod=30 Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.095225 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-txp2w"] Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.095468 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-txp2w" podUID="f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" containerName="registry-server" containerID="cri-o://d3932d25c3aaf08a595c2af7ee315a6a0b2efd503369ee7398e6b39ad609dc3c" gracePeriod=30 Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.117421 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z2gc7"] Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.117780 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-z2gc7" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" containerName="registry-server" containerID="cri-o://f7b31d5849d6707802fb373a1fe6f70b7a45ddade6fd6d9f2c7e5319e74f32d3" gracePeriod=30 Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.130728 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d9n25"] Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.131447 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.174308 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d9n25"] Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.295564 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf5vg\" (UniqueName: \"kubernetes.io/projected/023be687-a773-401c-981b-e3d7136f53b6-kube-api-access-bf5vg\") pod \"marketplace-operator-79b997595-d9n25\" (UID: \"023be687-a773-401c-981b-e3d7136f53b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.295654 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/023be687-a773-401c-981b-e3d7136f53b6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d9n25\" (UID: \"023be687-a773-401c-981b-e3d7136f53b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.295702 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/023be687-a773-401c-981b-e3d7136f53b6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d9n25\" (UID: \"023be687-a773-401c-981b-e3d7136f53b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.397088 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf5vg\" (UniqueName: \"kubernetes.io/projected/023be687-a773-401c-981b-e3d7136f53b6-kube-api-access-bf5vg\") pod \"marketplace-operator-79b997595-d9n25\" (UID: \"023be687-a773-401c-981b-e3d7136f53b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.397181 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/023be687-a773-401c-981b-e3d7136f53b6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d9n25\" (UID: \"023be687-a773-401c-981b-e3d7136f53b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.397207 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/023be687-a773-401c-981b-e3d7136f53b6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d9n25\" (UID: \"023be687-a773-401c-981b-e3d7136f53b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.400120 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/023be687-a773-401c-981b-e3d7136f53b6-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-d9n25\" (UID: \"023be687-a773-401c-981b-e3d7136f53b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.423149 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/023be687-a773-401c-981b-e3d7136f53b6-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-d9n25\" (UID: \"023be687-a773-401c-981b-e3d7136f53b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.443815 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf5vg\" (UniqueName: \"kubernetes.io/projected/023be687-a773-401c-981b-e3d7136f53b6-kube-api-access-bf5vg\") pod \"marketplace-operator-79b997595-d9n25\" (UID: \"023be687-a773-401c-981b-e3d7136f53b6\") " pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.455655 4632 generic.go:334] "Generic (PLEG): container finished" podID="797176c6-dd56-48d6-8004-ff1dd5353a50" containerID="0fd5f07ae3c28f8c24cc66a585de93acc08f170fb621bbeb190cd66596980871" exitCode=0 Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.455851 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" event={"ID":"797176c6-dd56-48d6-8004-ff1dd5353a50","Type":"ContainerDied","Data":"0fd5f07ae3c28f8c24cc66a585de93acc08f170fb621bbeb190cd66596980871"} Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.464335 4632 generic.go:334] "Generic (PLEG): container finished" podID="a110c276-8516-4f9e-a6af-d6837cd0f387" containerID="f7b31d5849d6707802fb373a1fe6f70b7a45ddade6fd6d9f2c7e5319e74f32d3" exitCode=0 Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.464421 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2gc7" event={"ID":"a110c276-8516-4f9e-a6af-d6837cd0f387","Type":"ContainerDied","Data":"f7b31d5849d6707802fb373a1fe6f70b7a45ddade6fd6d9f2c7e5319e74f32d3"} Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.469050 4632 generic.go:334] "Generic (PLEG): container finished" podID="b11a7dff-bf08-44c3-b4f4-923119c13717" containerID="55bfc00a5732a457ecbee5c7be945027bdb42c0137a6b22125d44dafb5924f59" exitCode=0 Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.469124 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8wjg" event={"ID":"b11a7dff-bf08-44c3-b4f4-923119c13717","Type":"ContainerDied","Data":"55bfc00a5732a457ecbee5c7be945027bdb42c0137a6b22125d44dafb5924f59"} Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.472154 4632 generic.go:334] "Generic (PLEG): container finished" podID="f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" containerID="d3932d25c3aaf08a595c2af7ee315a6a0b2efd503369ee7398e6b39ad609dc3c" exitCode=0 Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.472218 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txp2w" event={"ID":"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87","Type":"ContainerDied","Data":"d3932d25c3aaf08a595c2af7ee315a6a0b2efd503369ee7398e6b39ad609dc3c"} Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.493042 4632 generic.go:334] "Generic (PLEG): container finished" podID="bd46ae04-0610-4aa5-9385-dd45de66c5dd" containerID="5a10aa8d51646d1f515364874b0426c82d85f03f52a4924f31299cb0395b0607" exitCode=0 Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.493166 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvh86" event={"ID":"bd46ae04-0610-4aa5-9385-dd45de66c5dd","Type":"ContainerDied","Data":"5a10aa8d51646d1f515364874b0426c82d85f03f52a4924f31299cb0395b0607"} Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.504641 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.592378 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.599753 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11a7dff-bf08-44c3-b4f4-923119c13717-catalog-content\") pod \"b11a7dff-bf08-44c3-b4f4-923119c13717\" (UID: \"b11a7dff-bf08-44c3-b4f4-923119c13717\") " Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.599839 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11a7dff-bf08-44c3-b4f4-923119c13717-utilities\") pod \"b11a7dff-bf08-44c3-b4f4-923119c13717\" (UID: \"b11a7dff-bf08-44c3-b4f4-923119c13717\") " Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.599878 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wlsv8\" (UniqueName: \"kubernetes.io/projected/b11a7dff-bf08-44c3-b4f4-923119c13717-kube-api-access-wlsv8\") pod \"b11a7dff-bf08-44c3-b4f4-923119c13717\" (UID: \"b11a7dff-bf08-44c3-b4f4-923119c13717\") " Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.604426 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11a7dff-bf08-44c3-b4f4-923119c13717-utilities" (OuterVolumeSpecName: "utilities") pod "b11a7dff-bf08-44c3-b4f4-923119c13717" (UID: "b11a7dff-bf08-44c3-b4f4-923119c13717"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.606090 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11a7dff-bf08-44c3-b4f4-923119c13717-kube-api-access-wlsv8" (OuterVolumeSpecName: "kube-api-access-wlsv8") pod "b11a7dff-bf08-44c3-b4f4-923119c13717" (UID: "b11a7dff-bf08-44c3-b4f4-923119c13717"). InnerVolumeSpecName "kube-api-access-wlsv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.689633 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.695282 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11a7dff-bf08-44c3-b4f4-923119c13717-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11a7dff-bf08-44c3-b4f4-923119c13717" (UID: "b11a7dff-bf08-44c3-b4f4-923119c13717"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.701318 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-utilities\") pod \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\" (UID: \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\") " Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.701394 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5jx8\" (UniqueName: \"kubernetes.io/projected/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-kube-api-access-z5jx8\") pod \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\" (UID: \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\") " Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.701494 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-catalog-content\") pod \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\" (UID: \"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87\") " Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.701751 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11a7dff-bf08-44c3-b4f4-923119c13717-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.701772 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11a7dff-bf08-44c3-b4f4-923119c13717-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.701784 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wlsv8\" (UniqueName: \"kubernetes.io/projected/b11a7dff-bf08-44c3-b4f4-923119c13717-kube-api-access-wlsv8\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.703784 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-utilities" (OuterVolumeSpecName: "utilities") pod "f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" (UID: "f0cd0b7e-eded-4a51-8b1e-e67b9381bc87"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.706842 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-kube-api-access-z5jx8" (OuterVolumeSpecName: "kube-api-access-z5jx8") pod "f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" (UID: "f0cd0b7e-eded-4a51-8b1e-e67b9381bc87"). InnerVolumeSpecName "kube-api-access-z5jx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.746001 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" (UID: "f0cd0b7e-eded-4a51-8b1e-e67b9381bc87"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.757133 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.774657 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.803402 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5jx8\" (UniqueName: \"kubernetes.io/projected/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-kube-api-access-z5jx8\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.803443 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.803454 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.905866 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd46ae04-0610-4aa5-9385-dd45de66c5dd-catalog-content\") pod \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\" (UID: \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\") " Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.905951 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd46ae04-0610-4aa5-9385-dd45de66c5dd-utilities\") pod \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\" (UID: \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\") " Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.905994 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a110c276-8516-4f9e-a6af-d6837cd0f387-utilities\") pod \"a110c276-8516-4f9e-a6af-d6837cd0f387\" (UID: \"a110c276-8516-4f9e-a6af-d6837cd0f387\") " Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.906015 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5kzjw\" (UniqueName: \"kubernetes.io/projected/bd46ae04-0610-4aa5-9385-dd45de66c5dd-kube-api-access-5kzjw\") pod \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\" (UID: \"bd46ae04-0610-4aa5-9385-dd45de66c5dd\") " Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.906063 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfdk8\" (UniqueName: \"kubernetes.io/projected/a110c276-8516-4f9e-a6af-d6837cd0f387-kube-api-access-tfdk8\") pod \"a110c276-8516-4f9e-a6af-d6837cd0f387\" (UID: \"a110c276-8516-4f9e-a6af-d6837cd0f387\") " Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.906099 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a110c276-8516-4f9e-a6af-d6837cd0f387-catalog-content\") pod \"a110c276-8516-4f9e-a6af-d6837cd0f387\" (UID: \"a110c276-8516-4f9e-a6af-d6837cd0f387\") " Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.907860 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd46ae04-0610-4aa5-9385-dd45de66c5dd-utilities" (OuterVolumeSpecName: "utilities") pod "bd46ae04-0610-4aa5-9385-dd45de66c5dd" (UID: "bd46ae04-0610-4aa5-9385-dd45de66c5dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.908466 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a110c276-8516-4f9e-a6af-d6837cd0f387-utilities" (OuterVolumeSpecName: "utilities") pod "a110c276-8516-4f9e-a6af-d6837cd0f387" (UID: "a110c276-8516-4f9e-a6af-d6837cd0f387"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.914817 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a110c276-8516-4f9e-a6af-d6837cd0f387-kube-api-access-tfdk8" (OuterVolumeSpecName: "kube-api-access-tfdk8") pod "a110c276-8516-4f9e-a6af-d6837cd0f387" (UID: "a110c276-8516-4f9e-a6af-d6837cd0f387"). InnerVolumeSpecName "kube-api-access-tfdk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:11:41 crc kubenswrapper[4632]: I0313 10:11:41.947681 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd46ae04-0610-4aa5-9385-dd45de66c5dd-kube-api-access-5kzjw" (OuterVolumeSpecName: "kube-api-access-5kzjw") pod "bd46ae04-0610-4aa5-9385-dd45de66c5dd" (UID: "bd46ae04-0610-4aa5-9385-dd45de66c5dd"). InnerVolumeSpecName "kube-api-access-5kzjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.007727 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a110c276-8516-4f9e-a6af-d6837cd0f387-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.007755 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5kzjw\" (UniqueName: \"kubernetes.io/projected/bd46ae04-0610-4aa5-9385-dd45de66c5dd-kube-api-access-5kzjw\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.007766 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfdk8\" (UniqueName: \"kubernetes.io/projected/a110c276-8516-4f9e-a6af-d6837cd0f387-kube-api-access-tfdk8\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.007775 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd46ae04-0610-4aa5-9385-dd45de66c5dd-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.009272 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.009773 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd46ae04-0610-4aa5-9385-dd45de66c5dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bd46ae04-0610-4aa5-9385-dd45de66c5dd" (UID: "bd46ae04-0610-4aa5-9385-dd45de66c5dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.078111 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-d9n25"] Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.108716 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/797176c6-dd56-48d6-8004-ff1dd5353a50-marketplace-operator-metrics\") pod \"797176c6-dd56-48d6-8004-ff1dd5353a50\" (UID: \"797176c6-dd56-48d6-8004-ff1dd5353a50\") " Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.108903 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qtrb\" (UniqueName: \"kubernetes.io/projected/797176c6-dd56-48d6-8004-ff1dd5353a50-kube-api-access-8qtrb\") pod \"797176c6-dd56-48d6-8004-ff1dd5353a50\" (UID: \"797176c6-dd56-48d6-8004-ff1dd5353a50\") " Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.108984 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/797176c6-dd56-48d6-8004-ff1dd5353a50-marketplace-trusted-ca\") pod \"797176c6-dd56-48d6-8004-ff1dd5353a50\" (UID: \"797176c6-dd56-48d6-8004-ff1dd5353a50\") " Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.109396 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd46ae04-0610-4aa5-9385-dd45de66c5dd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.111108 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/797176c6-dd56-48d6-8004-ff1dd5353a50-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "797176c6-dd56-48d6-8004-ff1dd5353a50" (UID: "797176c6-dd56-48d6-8004-ff1dd5353a50"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.115109 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/797176c6-dd56-48d6-8004-ff1dd5353a50-kube-api-access-8qtrb" (OuterVolumeSpecName: "kube-api-access-8qtrb") pod "797176c6-dd56-48d6-8004-ff1dd5353a50" (UID: "797176c6-dd56-48d6-8004-ff1dd5353a50"). InnerVolumeSpecName "kube-api-access-8qtrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.119473 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797176c6-dd56-48d6-8004-ff1dd5353a50-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "797176c6-dd56-48d6-8004-ff1dd5353a50" (UID: "797176c6-dd56-48d6-8004-ff1dd5353a50"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.148757 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a110c276-8516-4f9e-a6af-d6837cd0f387-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a110c276-8516-4f9e-a6af-d6837cd0f387" (UID: "a110c276-8516-4f9e-a6af-d6837cd0f387"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.211125 4632 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/797176c6-dd56-48d6-8004-ff1dd5353a50-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.211173 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a110c276-8516-4f9e-a6af-d6837cd0f387-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.211195 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qtrb\" (UniqueName: \"kubernetes.io/projected/797176c6-dd56-48d6-8004-ff1dd5353a50-kube-api-access-8qtrb\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.211208 4632 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/797176c6-dd56-48d6-8004-ff1dd5353a50-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.500269 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" event={"ID":"023be687-a773-401c-981b-e3d7136f53b6","Type":"ContainerStarted","Data":"077a793783663fc0a0919deecac0d9a526695fee5caabb2c2420a407d92820f6"} Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.500361 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.500424 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" event={"ID":"023be687-a773-401c-981b-e3d7136f53b6","Type":"ContainerStarted","Data":"4a96453b15b7d7e2a7c4a1760c93d848d945f3f4fd15bd89c331bde321bc785c"} Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.503818 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-z2gc7" event={"ID":"a110c276-8516-4f9e-a6af-d6837cd0f387","Type":"ContainerDied","Data":"97491a7f994f5c8dffa29a28fb1914c53f3fb5687971c6cdb3d3b5b636967634"} Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.503884 4632 scope.go:117] "RemoveContainer" containerID="f7b31d5849d6707802fb373a1fe6f70b7a45ddade6fd6d9f2c7e5319e74f32d3" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.504065 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-z2gc7" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.513746 4632 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-d9n25 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.70:8080/healthz\": dial tcp 10.217.0.70:8080: connect: connection refused" start-of-body= Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.513787 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" podUID="023be687-a773-401c-981b-e3d7136f53b6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.70:8080/healthz\": dial tcp 10.217.0.70:8080: connect: connection refused" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.519598 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p8wjg" event={"ID":"b11a7dff-bf08-44c3-b4f4-923119c13717","Type":"ContainerDied","Data":"67d236def43f1634b091443716f5df0abcd64ee4e8ef6768dd906ab3397df097"} Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.519711 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p8wjg" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.526659 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-txp2w" event={"ID":"f0cd0b7e-eded-4a51-8b1e-e67b9381bc87","Type":"ContainerDied","Data":"3442fa414f5a2c2798e2a9a29c903f3acac1f4e2b61c872fefc305318ea1c556"} Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.526746 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-txp2w" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.532568 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" event={"ID":"797176c6-dd56-48d6-8004-ff1dd5353a50","Type":"ContainerDied","Data":"f11fbb0ec92177c2b8cb772cacb63ff7d8a26b02bee6907aaa00dedbedf68d98"} Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.532838 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-2n99d" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.538997 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jvh86" event={"ID":"bd46ae04-0610-4aa5-9385-dd45de66c5dd","Type":"ContainerDied","Data":"c1e7366f3326cfd08308453ff8a94a3f8d3ce8ebc6a33b2bfafadd960643927e"} Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.539290 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jvh86" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.539882 4632 scope.go:117] "RemoveContainer" containerID="ef106624caa843911d5171f0d70f22c07e7e2bd19b6992932276ca1226b858e3" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.544416 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" podStartSLOduration=1.5444010430000001 podStartE2EDuration="1.544401043s" podCreationTimestamp="2026-03-13 10:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:11:42.533229578 +0000 UTC m=+476.555759731" watchObservedRunningTime="2026-03-13 10:11:42.544401043 +0000 UTC m=+476.566931176" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.569248 4632 scope.go:117] "RemoveContainer" containerID="0d073c1adaa82aa87cab8618a50587cfed8b79fe657e3f2878a87c7599c612fb" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.572200 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p8wjg"] Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.581622 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p8wjg"] Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.594023 4632 scope.go:117] "RemoveContainer" containerID="55bfc00a5732a457ecbee5c7be945027bdb42c0137a6b22125d44dafb5924f59" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.598376 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-z2gc7"] Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.616873 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-z2gc7"] Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.618917 4632 scope.go:117] "RemoveContainer" containerID="06491b70d16bc5a697f5518128f63de5fdeb769cc33d09d9262078f5aa75a5b8" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.621364 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2n99d"] Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.641267 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-2n99d"] Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.656038 4632 scope.go:117] "RemoveContainer" containerID="31ed0687958629bbe6ae3de064bae07567e401a6f6f2576bf2e48b7390937742" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.656888 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-txp2w"] Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.661330 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-txp2w"] Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.669501 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jvh86"] Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.675001 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jvh86"] Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.679798 4632 scope.go:117] "RemoveContainer" containerID="d3932d25c3aaf08a595c2af7ee315a6a0b2efd503369ee7398e6b39ad609dc3c" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.702931 4632 scope.go:117] "RemoveContainer" containerID="643ad1b648678ed35dcc10aaf9a844460c880f38f688c0da6821345eaf872208" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.729613 4632 scope.go:117] "RemoveContainer" containerID="d1da7a7847a6ff5346add9e3ed943cdc6232146978e6161d764011992ac73c84" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.747951 4632 scope.go:117] "RemoveContainer" containerID="0fd5f07ae3c28f8c24cc66a585de93acc08f170fb621bbeb190cd66596980871" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.770558 4632 scope.go:117] "RemoveContainer" containerID="5a10aa8d51646d1f515364874b0426c82d85f03f52a4924f31299cb0395b0607" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.792341 4632 scope.go:117] "RemoveContainer" containerID="ba15fa8797c3390ead2f6a2f6b5a64ad766bc4a942dfc13cbdc76a3242dd09c0" Mar 13 10:11:42 crc kubenswrapper[4632]: I0313 10:11:42.823457 4632 scope.go:117] "RemoveContainer" containerID="eabb475f877c5898896f887fa631fab417c1e3579d0424b2b6c06f4278f091af" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245279 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7ksc5"] Mar 13 10:11:43 crc kubenswrapper[4632]: E0313 10:11:43.245515 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd46ae04-0610-4aa5-9385-dd45de66c5dd" containerName="extract-content" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245542 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd46ae04-0610-4aa5-9385-dd45de66c5dd" containerName="extract-content" Mar 13 10:11:43 crc kubenswrapper[4632]: E0313 10:11:43.245556 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" containerName="extract-content" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245564 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" containerName="extract-content" Mar 13 10:11:43 crc kubenswrapper[4632]: E0313 10:11:43.245572 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" containerName="extract-utilities" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245581 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" containerName="extract-utilities" Mar 13 10:11:43 crc kubenswrapper[4632]: E0313 10:11:43.245592 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11a7dff-bf08-44c3-b4f4-923119c13717" containerName="registry-server" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245599 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11a7dff-bf08-44c3-b4f4-923119c13717" containerName="registry-server" Mar 13 10:11:43 crc kubenswrapper[4632]: E0313 10:11:43.245608 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11a7dff-bf08-44c3-b4f4-923119c13717" containerName="extract-utilities" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245615 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11a7dff-bf08-44c3-b4f4-923119c13717" containerName="extract-utilities" Mar 13 10:11:43 crc kubenswrapper[4632]: E0313 10:11:43.245625 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd46ae04-0610-4aa5-9385-dd45de66c5dd" containerName="extract-utilities" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245632 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd46ae04-0610-4aa5-9385-dd45de66c5dd" containerName="extract-utilities" Mar 13 10:11:43 crc kubenswrapper[4632]: E0313 10:11:43.245642 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" containerName="registry-server" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245649 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" containerName="registry-server" Mar 13 10:11:43 crc kubenswrapper[4632]: E0313 10:11:43.245662 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd46ae04-0610-4aa5-9385-dd45de66c5dd" containerName="registry-server" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245669 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd46ae04-0610-4aa5-9385-dd45de66c5dd" containerName="registry-server" Mar 13 10:11:43 crc kubenswrapper[4632]: E0313 10:11:43.245678 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="797176c6-dd56-48d6-8004-ff1dd5353a50" containerName="marketplace-operator" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245685 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="797176c6-dd56-48d6-8004-ff1dd5353a50" containerName="marketplace-operator" Mar 13 10:11:43 crc kubenswrapper[4632]: E0313 10:11:43.245695 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" containerName="extract-content" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245702 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" containerName="extract-content" Mar 13 10:11:43 crc kubenswrapper[4632]: E0313 10:11:43.245713 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" containerName="registry-server" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245721 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" containerName="registry-server" Mar 13 10:11:43 crc kubenswrapper[4632]: E0313 10:11:43.245731 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" containerName="extract-utilities" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245739 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" containerName="extract-utilities" Mar 13 10:11:43 crc kubenswrapper[4632]: E0313 10:11:43.245750 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b11a7dff-bf08-44c3-b4f4-923119c13717" containerName="extract-content" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245758 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="b11a7dff-bf08-44c3-b4f4-923119c13717" containerName="extract-content" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245863 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="797176c6-dd56-48d6-8004-ff1dd5353a50" containerName="marketplace-operator" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245891 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" containerName="registry-server" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245903 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="b11a7dff-bf08-44c3-b4f4-923119c13717" containerName="registry-server" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245918 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd46ae04-0610-4aa5-9385-dd45de66c5dd" containerName="registry-server" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.245931 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" containerName="registry-server" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.246846 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.249392 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.255149 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7ksc5"] Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.325915 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fa3faab-9e82-4fde-afff-3de6939a17d1-utilities\") pod \"certified-operators-7ksc5\" (UID: \"0fa3faab-9e82-4fde-afff-3de6939a17d1\") " pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.326032 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fa3faab-9e82-4fde-afff-3de6939a17d1-catalog-content\") pod \"certified-operators-7ksc5\" (UID: \"0fa3faab-9e82-4fde-afff-3de6939a17d1\") " pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.326120 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72v2p\" (UniqueName: \"kubernetes.io/projected/0fa3faab-9e82-4fde-afff-3de6939a17d1-kube-api-access-72v2p\") pod \"certified-operators-7ksc5\" (UID: \"0fa3faab-9e82-4fde-afff-3de6939a17d1\") " pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.427979 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fa3faab-9e82-4fde-afff-3de6939a17d1-utilities\") pod \"certified-operators-7ksc5\" (UID: \"0fa3faab-9e82-4fde-afff-3de6939a17d1\") " pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.428064 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fa3faab-9e82-4fde-afff-3de6939a17d1-catalog-content\") pod \"certified-operators-7ksc5\" (UID: \"0fa3faab-9e82-4fde-afff-3de6939a17d1\") " pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.428147 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72v2p\" (UniqueName: \"kubernetes.io/projected/0fa3faab-9e82-4fde-afff-3de6939a17d1-kube-api-access-72v2p\") pod \"certified-operators-7ksc5\" (UID: \"0fa3faab-9e82-4fde-afff-3de6939a17d1\") " pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.429376 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0fa3faab-9e82-4fde-afff-3de6939a17d1-catalog-content\") pod \"certified-operators-7ksc5\" (UID: \"0fa3faab-9e82-4fde-afff-3de6939a17d1\") " pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.429471 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0fa3faab-9e82-4fde-afff-3de6939a17d1-utilities\") pod \"certified-operators-7ksc5\" (UID: \"0fa3faab-9e82-4fde-afff-3de6939a17d1\") " pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.448054 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72v2p\" (UniqueName: \"kubernetes.io/projected/0fa3faab-9e82-4fde-afff-3de6939a17d1-kube-api-access-72v2p\") pod \"certified-operators-7ksc5\" (UID: \"0fa3faab-9e82-4fde-afff-3de6939a17d1\") " pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.449038 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-gdt8x"] Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.450400 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.452001 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.472631 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gdt8x"] Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.529647 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6s75\" (UniqueName: \"kubernetes.io/projected/f7f61b75-16bf-4c5a-be30-c88d155c203f-kube-api-access-k6s75\") pod \"redhat-marketplace-gdt8x\" (UID: \"f7f61b75-16bf-4c5a-be30-c88d155c203f\") " pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.529715 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7f61b75-16bf-4c5a-be30-c88d155c203f-utilities\") pod \"redhat-marketplace-gdt8x\" (UID: \"f7f61b75-16bf-4c5a-be30-c88d155c203f\") " pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.529739 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7f61b75-16bf-4c5a-be30-c88d155c203f-catalog-content\") pod \"redhat-marketplace-gdt8x\" (UID: \"f7f61b75-16bf-4c5a-be30-c88d155c203f\") " pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.552250 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.575556 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.631837 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6s75\" (UniqueName: \"kubernetes.io/projected/f7f61b75-16bf-4c5a-be30-c88d155c203f-kube-api-access-k6s75\") pod \"redhat-marketplace-gdt8x\" (UID: \"f7f61b75-16bf-4c5a-be30-c88d155c203f\") " pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.631924 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7f61b75-16bf-4c5a-be30-c88d155c203f-utilities\") pod \"redhat-marketplace-gdt8x\" (UID: \"f7f61b75-16bf-4c5a-be30-c88d155c203f\") " pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.632029 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7f61b75-16bf-4c5a-be30-c88d155c203f-catalog-content\") pod \"redhat-marketplace-gdt8x\" (UID: \"f7f61b75-16bf-4c5a-be30-c88d155c203f\") " pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.632577 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f7f61b75-16bf-4c5a-be30-c88d155c203f-catalog-content\") pod \"redhat-marketplace-gdt8x\" (UID: \"f7f61b75-16bf-4c5a-be30-c88d155c203f\") " pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.633291 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f7f61b75-16bf-4c5a-be30-c88d155c203f-utilities\") pod \"redhat-marketplace-gdt8x\" (UID: \"f7f61b75-16bf-4c5a-be30-c88d155c203f\") " pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.657726 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6s75\" (UniqueName: \"kubernetes.io/projected/f7f61b75-16bf-4c5a-be30-c88d155c203f-kube-api-access-k6s75\") pod \"redhat-marketplace-gdt8x\" (UID: \"f7f61b75-16bf-4c5a-be30-c88d155c203f\") " pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:43 crc kubenswrapper[4632]: I0313 10:11:43.787227 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:44 crc kubenswrapper[4632]: I0313 10:11:44.029081 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7ksc5"] Mar 13 10:11:44 crc kubenswrapper[4632]: I0313 10:11:44.038313 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-gdt8x"] Mar 13 10:11:44 crc kubenswrapper[4632]: W0313 10:11:44.041208 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0fa3faab_9e82_4fde_afff_3de6939a17d1.slice/crio-05d058946411b3aa47f2245f16be7366b96dd2370570fc574b5f1accea9ab081 WatchSource:0}: Error finding container 05d058946411b3aa47f2245f16be7366b96dd2370570fc574b5f1accea9ab081: Status 404 returned error can't find the container with id 05d058946411b3aa47f2245f16be7366b96dd2370570fc574b5f1accea9ab081 Mar 13 10:11:44 crc kubenswrapper[4632]: I0313 10:11:44.054147 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="797176c6-dd56-48d6-8004-ff1dd5353a50" path="/var/lib/kubelet/pods/797176c6-dd56-48d6-8004-ff1dd5353a50/volumes" Mar 13 10:11:44 crc kubenswrapper[4632]: I0313 10:11:44.055710 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a110c276-8516-4f9e-a6af-d6837cd0f387" path="/var/lib/kubelet/pods/a110c276-8516-4f9e-a6af-d6837cd0f387/volumes" Mar 13 10:11:44 crc kubenswrapper[4632]: I0313 10:11:44.058292 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11a7dff-bf08-44c3-b4f4-923119c13717" path="/var/lib/kubelet/pods/b11a7dff-bf08-44c3-b4f4-923119c13717/volumes" Mar 13 10:11:44 crc kubenswrapper[4632]: W0313 10:11:44.058313 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf7f61b75_16bf_4c5a_be30_c88d155c203f.slice/crio-0a9cc2740ac0ee6593eb911e246dbb00bc5e6cd7aaf49b2f0c50b4d82a976c9f WatchSource:0}: Error finding container 0a9cc2740ac0ee6593eb911e246dbb00bc5e6cd7aaf49b2f0c50b4d82a976c9f: Status 404 returned error can't find the container with id 0a9cc2740ac0ee6593eb911e246dbb00bc5e6cd7aaf49b2f0c50b4d82a976c9f Mar 13 10:11:44 crc kubenswrapper[4632]: I0313 10:11:44.062024 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd46ae04-0610-4aa5-9385-dd45de66c5dd" path="/var/lib/kubelet/pods/bd46ae04-0610-4aa5-9385-dd45de66c5dd/volumes" Mar 13 10:11:44 crc kubenswrapper[4632]: I0313 10:11:44.063552 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0cd0b7e-eded-4a51-8b1e-e67b9381bc87" path="/var/lib/kubelet/pods/f0cd0b7e-eded-4a51-8b1e-e67b9381bc87/volumes" Mar 13 10:11:44 crc kubenswrapper[4632]: I0313 10:11:44.559401 4632 generic.go:334] "Generic (PLEG): container finished" podID="f7f61b75-16bf-4c5a-be30-c88d155c203f" containerID="ea39ae0061ed4982483c38e5508ad6a45dafe3af038369c1ee929d7fd6c1f92c" exitCode=0 Mar 13 10:11:44 crc kubenswrapper[4632]: I0313 10:11:44.559777 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gdt8x" event={"ID":"f7f61b75-16bf-4c5a-be30-c88d155c203f","Type":"ContainerDied","Data":"ea39ae0061ed4982483c38e5508ad6a45dafe3af038369c1ee929d7fd6c1f92c"} Mar 13 10:11:44 crc kubenswrapper[4632]: I0313 10:11:44.559814 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gdt8x" event={"ID":"f7f61b75-16bf-4c5a-be30-c88d155c203f","Type":"ContainerStarted","Data":"0a9cc2740ac0ee6593eb911e246dbb00bc5e6cd7aaf49b2f0c50b4d82a976c9f"} Mar 13 10:11:44 crc kubenswrapper[4632]: I0313 10:11:44.562769 4632 generic.go:334] "Generic (PLEG): container finished" podID="0fa3faab-9e82-4fde-afff-3de6939a17d1" containerID="a8c4b44ced612e786b4a2c37400abd9d40040bd937d3c7c8a679c1986d317cb9" exitCode=0 Mar 13 10:11:44 crc kubenswrapper[4632]: I0313 10:11:44.562825 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ksc5" event={"ID":"0fa3faab-9e82-4fde-afff-3de6939a17d1","Type":"ContainerDied","Data":"a8c4b44ced612e786b4a2c37400abd9d40040bd937d3c7c8a679c1986d317cb9"} Mar 13 10:11:44 crc kubenswrapper[4632]: I0313 10:11:44.562900 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ksc5" event={"ID":"0fa3faab-9e82-4fde-afff-3de6939a17d1","Type":"ContainerStarted","Data":"05d058946411b3aa47f2245f16be7366b96dd2370570fc574b5f1accea9ab081"} Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.569832 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gdt8x" event={"ID":"f7f61b75-16bf-4c5a-be30-c88d155c203f","Type":"ContainerStarted","Data":"9d93b23ee83cf056c79c74de74877ea34c2e598d1e294be73d5438be718035cc"} Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.661048 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vwgfr"] Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.662597 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.665296 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.668111 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vwgfr"] Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.762719 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5cc71d2-1901-4778-8e20-93646cfc1a85-utilities\") pod \"redhat-operators-vwgfr\" (UID: \"f5cc71d2-1901-4778-8e20-93646cfc1a85\") " pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.762770 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5cc71d2-1901-4778-8e20-93646cfc1a85-catalog-content\") pod \"redhat-operators-vwgfr\" (UID: \"f5cc71d2-1901-4778-8e20-93646cfc1a85\") " pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.762826 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkq4h\" (UniqueName: \"kubernetes.io/projected/f5cc71d2-1901-4778-8e20-93646cfc1a85-kube-api-access-qkq4h\") pod \"redhat-operators-vwgfr\" (UID: \"f5cc71d2-1901-4778-8e20-93646cfc1a85\") " pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.849575 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lgwff"] Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.852171 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:11:45 crc kubenswrapper[4632]: W0313 10:11:45.857125 4632 reflector.go:561] object-"openshift-marketplace"/"community-operators-dockercfg-dmngl": failed to list *v1.Secret: secrets "community-operators-dockercfg-dmngl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Mar 13 10:11:45 crc kubenswrapper[4632]: E0313 10:11:45.857412 4632 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"community-operators-dockercfg-dmngl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"community-operators-dockercfg-dmngl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.863576 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkq4h\" (UniqueName: \"kubernetes.io/projected/f5cc71d2-1901-4778-8e20-93646cfc1a85-kube-api-access-qkq4h\") pod \"redhat-operators-vwgfr\" (UID: \"f5cc71d2-1901-4778-8e20-93646cfc1a85\") " pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.863652 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5cc71d2-1901-4778-8e20-93646cfc1a85-utilities\") pod \"redhat-operators-vwgfr\" (UID: \"f5cc71d2-1901-4778-8e20-93646cfc1a85\") " pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.863693 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5cc71d2-1901-4778-8e20-93646cfc1a85-catalog-content\") pod \"redhat-operators-vwgfr\" (UID: \"f5cc71d2-1901-4778-8e20-93646cfc1a85\") " pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.864300 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5cc71d2-1901-4778-8e20-93646cfc1a85-catalog-content\") pod \"redhat-operators-vwgfr\" (UID: \"f5cc71d2-1901-4778-8e20-93646cfc1a85\") " pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.864367 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5cc71d2-1901-4778-8e20-93646cfc1a85-utilities\") pod \"redhat-operators-vwgfr\" (UID: \"f5cc71d2-1901-4778-8e20-93646cfc1a85\") " pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.877404 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lgwff"] Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.912072 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkq4h\" (UniqueName: \"kubernetes.io/projected/f5cc71d2-1901-4778-8e20-93646cfc1a85-kube-api-access-qkq4h\") pod \"redhat-operators-vwgfr\" (UID: \"f5cc71d2-1901-4778-8e20-93646cfc1a85\") " pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.965281 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-utilities\") pod \"community-operators-lgwff\" (UID: \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\") " pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.965415 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sw5gz\" (UniqueName: \"kubernetes.io/projected/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-kube-api-access-sw5gz\") pod \"community-operators-lgwff\" (UID: \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\") " pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:11:45 crc kubenswrapper[4632]: I0313 10:11:45.965463 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-catalog-content\") pod \"community-operators-lgwff\" (UID: \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\") " pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.028672 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.066791 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sw5gz\" (UniqueName: \"kubernetes.io/projected/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-kube-api-access-sw5gz\") pod \"community-operators-lgwff\" (UID: \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\") " pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.066861 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-catalog-content\") pod \"community-operators-lgwff\" (UID: \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\") " pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.067326 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-catalog-content\") pod \"community-operators-lgwff\" (UID: \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\") " pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.069056 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-utilities\") pod \"community-operators-lgwff\" (UID: \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\") " pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.068616 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-utilities\") pod \"community-operators-lgwff\" (UID: \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\") " pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.087000 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sw5gz\" (UniqueName: \"kubernetes.io/projected/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-kube-api-access-sw5gz\") pod \"community-operators-lgwff\" (UID: \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\") " pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.236553 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vwgfr"] Mar 13 10:11:46 crc kubenswrapper[4632]: W0313 10:11:46.243669 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5cc71d2_1901_4778_8e20_93646cfc1a85.slice/crio-02f7137cb66e9d8ffc3cef2380010ceb3716ead7cd44ba15fbfbc983daf62896 WatchSource:0}: Error finding container 02f7137cb66e9d8ffc3cef2380010ceb3716ead7cd44ba15fbfbc983daf62896: Status 404 returned error can't find the container with id 02f7137cb66e9d8ffc3cef2380010ceb3716ead7cd44ba15fbfbc983daf62896 Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.576622 4632 generic.go:334] "Generic (PLEG): container finished" podID="f5cc71d2-1901-4778-8e20-93646cfc1a85" containerID="5cd341447acd60a5969c3c5d69de4256985df7ef4a0d5e8e53aa3477a222e75f" exitCode=0 Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.576672 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vwgfr" event={"ID":"f5cc71d2-1901-4778-8e20-93646cfc1a85","Type":"ContainerDied","Data":"5cd341447acd60a5969c3c5d69de4256985df7ef4a0d5e8e53aa3477a222e75f"} Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.576747 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vwgfr" event={"ID":"f5cc71d2-1901-4778-8e20-93646cfc1a85","Type":"ContainerStarted","Data":"02f7137cb66e9d8ffc3cef2380010ceb3716ead7cd44ba15fbfbc983daf62896"} Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.579261 4632 generic.go:334] "Generic (PLEG): container finished" podID="0fa3faab-9e82-4fde-afff-3de6939a17d1" containerID="aad77b665209df43df4f12654ddb920ecda5a086061cc28b2dbcc76554ecf7e4" exitCode=0 Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.579885 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ksc5" event={"ID":"0fa3faab-9e82-4fde-afff-3de6939a17d1","Type":"ContainerDied","Data":"aad77b665209df43df4f12654ddb920ecda5a086061cc28b2dbcc76554ecf7e4"} Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.583558 4632 generic.go:334] "Generic (PLEG): container finished" podID="f7f61b75-16bf-4c5a-be30-c88d155c203f" containerID="9d93b23ee83cf056c79c74de74877ea34c2e598d1e294be73d5438be718035cc" exitCode=0 Mar 13 10:11:46 crc kubenswrapper[4632]: I0313 10:11:46.583595 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gdt8x" event={"ID":"f7f61b75-16bf-4c5a-be30-c88d155c203f","Type":"ContainerDied","Data":"9d93b23ee83cf056c79c74de74877ea34c2e598d1e294be73d5438be718035cc"} Mar 13 10:11:47 crc kubenswrapper[4632]: I0313 10:11:47.117455 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 10:11:47 crc kubenswrapper[4632]: I0313 10:11:47.125571 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:11:47 crc kubenswrapper[4632]: I0313 10:11:47.590707 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-gdt8x" event={"ID":"f7f61b75-16bf-4c5a-be30-c88d155c203f","Type":"ContainerStarted","Data":"7d35b444d1bd4f368370f7c4bf5efbd36d548f0061deecb9c5a6f7055dac2899"} Mar 13 10:11:47 crc kubenswrapper[4632]: I0313 10:11:47.593005 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vwgfr" event={"ID":"f5cc71d2-1901-4778-8e20-93646cfc1a85","Type":"ContainerStarted","Data":"d7f8d7b027c206a1d65a3b139e82c14e6d7189a3bd2c2246d3677e3545220555"} Mar 13 10:11:47 crc kubenswrapper[4632]: W0313 10:11:47.618495 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d0fc567_0682_4bbc_981b_b4d1df62aa4e.slice/crio-4266d075e4d04ea92ccfdc02ec4b3551e54779fe4f2f2c386ac2a209fda18404 WatchSource:0}: Error finding container 4266d075e4d04ea92ccfdc02ec4b3551e54779fe4f2f2c386ac2a209fda18404: Status 404 returned error can't find the container with id 4266d075e4d04ea92ccfdc02ec4b3551e54779fe4f2f2c386ac2a209fda18404 Mar 13 10:11:47 crc kubenswrapper[4632]: I0313 10:11:47.620564 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lgwff"] Mar 13 10:11:47 crc kubenswrapper[4632]: I0313 10:11:47.641513 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-gdt8x" podStartSLOduration=2.187043059 podStartE2EDuration="4.641496544s" podCreationTimestamp="2026-03-13 10:11:43 +0000 UTC" firstStartedPulling="2026-03-13 10:11:44.562136069 +0000 UTC m=+478.584666222" lastFinishedPulling="2026-03-13 10:11:47.016589564 +0000 UTC m=+481.039119707" observedRunningTime="2026-03-13 10:11:47.637925372 +0000 UTC m=+481.660455505" watchObservedRunningTime="2026-03-13 10:11:47.641496544 +0000 UTC m=+481.664026667" Mar 13 10:11:48 crc kubenswrapper[4632]: I0313 10:11:48.602914 4632 generic.go:334] "Generic (PLEG): container finished" podID="f5cc71d2-1901-4778-8e20-93646cfc1a85" containerID="d7f8d7b027c206a1d65a3b139e82c14e6d7189a3bd2c2246d3677e3545220555" exitCode=0 Mar 13 10:11:48 crc kubenswrapper[4632]: I0313 10:11:48.602993 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vwgfr" event={"ID":"f5cc71d2-1901-4778-8e20-93646cfc1a85","Type":"ContainerDied","Data":"d7f8d7b027c206a1d65a3b139e82c14e6d7189a3bd2c2246d3677e3545220555"} Mar 13 10:11:48 crc kubenswrapper[4632]: I0313 10:11:48.604901 4632 generic.go:334] "Generic (PLEG): container finished" podID="0d0fc567-0682-4bbc-981b-b4d1df62aa4e" containerID="647f33468b0e454866917d9beec3f31ad6bc8dca469daccdfcf7e8df5de24312" exitCode=0 Mar 13 10:11:48 crc kubenswrapper[4632]: I0313 10:11:48.604965 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lgwff" event={"ID":"0d0fc567-0682-4bbc-981b-b4d1df62aa4e","Type":"ContainerDied","Data":"647f33468b0e454866917d9beec3f31ad6bc8dca469daccdfcf7e8df5de24312"} Mar 13 10:11:48 crc kubenswrapper[4632]: I0313 10:11:48.605006 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lgwff" event={"ID":"0d0fc567-0682-4bbc-981b-b4d1df62aa4e","Type":"ContainerStarted","Data":"4266d075e4d04ea92ccfdc02ec4b3551e54779fe4f2f2c386ac2a209fda18404"} Mar 13 10:11:49 crc kubenswrapper[4632]: I0313 10:11:49.612496 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vwgfr" event={"ID":"f5cc71d2-1901-4778-8e20-93646cfc1a85","Type":"ContainerStarted","Data":"bb3ba6ce6125af4b8eb6420d2a43fb97d9fe1916c470da9c74f26fdd0873591b"} Mar 13 10:11:49 crc kubenswrapper[4632]: I0313 10:11:49.654756 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vwgfr" podStartSLOduration=2.221006935 podStartE2EDuration="4.654729226s" podCreationTimestamp="2026-03-13 10:11:45 +0000 UTC" firstStartedPulling="2026-03-13 10:11:46.578262367 +0000 UTC m=+480.600792500" lastFinishedPulling="2026-03-13 10:11:49.011984658 +0000 UTC m=+483.034514791" observedRunningTime="2026-03-13 10:11:49.64705096 +0000 UTC m=+483.669581093" watchObservedRunningTime="2026-03-13 10:11:49.654729226 +0000 UTC m=+483.677259359" Mar 13 10:11:50 crc kubenswrapper[4632]: I0313 10:11:50.620303 4632 generic.go:334] "Generic (PLEG): container finished" podID="0d0fc567-0682-4bbc-981b-b4d1df62aa4e" containerID="6ed6c1b1b2793ab4b788dc1723932bf9c4121a7bf0945a697809d4c945eec749" exitCode=0 Mar 13 10:11:50 crc kubenswrapper[4632]: I0313 10:11:50.622737 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lgwff" event={"ID":"0d0fc567-0682-4bbc-981b-b4d1df62aa4e","Type":"ContainerDied","Data":"6ed6c1b1b2793ab4b788dc1723932bf9c4121a7bf0945a697809d4c945eec749"} Mar 13 10:11:51 crc kubenswrapper[4632]: I0313 10:11:51.638210 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7ksc5" event={"ID":"0fa3faab-9e82-4fde-afff-3de6939a17d1","Type":"ContainerStarted","Data":"43406af3f8edd079e1b933465df01dbcc93292bb25548398ea4066e38533524f"} Mar 13 10:11:51 crc kubenswrapper[4632]: I0313 10:11:51.640212 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lgwff" event={"ID":"0d0fc567-0682-4bbc-981b-b4d1df62aa4e","Type":"ContainerStarted","Data":"6fa56a0ef2065ba4287ddc46b227aad0c8d55e685aeeb7889682c05acb775492"} Mar 13 10:11:51 crc kubenswrapper[4632]: I0313 10:11:51.659066 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7ksc5" podStartSLOduration=2.160780737 podStartE2EDuration="8.659050714s" podCreationTimestamp="2026-03-13 10:11:43 +0000 UTC" firstStartedPulling="2026-03-13 10:11:44.566208232 +0000 UTC m=+478.588738365" lastFinishedPulling="2026-03-13 10:11:51.064478209 +0000 UTC m=+485.087008342" observedRunningTime="2026-03-13 10:11:51.656879025 +0000 UTC m=+485.679409168" watchObservedRunningTime="2026-03-13 10:11:51.659050714 +0000 UTC m=+485.681580847" Mar 13 10:11:53 crc kubenswrapper[4632]: I0313 10:11:53.576264 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:11:53 crc kubenswrapper[4632]: I0313 10:11:53.576332 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:11:53 crc kubenswrapper[4632]: I0313 10:11:53.618319 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:11:53 crc kubenswrapper[4632]: I0313 10:11:53.637335 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lgwff" podStartSLOduration=6.078483538 podStartE2EDuration="8.637318265s" podCreationTimestamp="2026-03-13 10:11:45 +0000 UTC" firstStartedPulling="2026-03-13 10:11:48.606607154 +0000 UTC m=+482.629137287" lastFinishedPulling="2026-03-13 10:11:51.165441881 +0000 UTC m=+485.187972014" observedRunningTime="2026-03-13 10:11:51.676715899 +0000 UTC m=+485.699246032" watchObservedRunningTime="2026-03-13 10:11:53.637318265 +0000 UTC m=+487.659848388" Mar 13 10:11:53 crc kubenswrapper[4632]: I0313 10:11:53.787988 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:53 crc kubenswrapper[4632]: I0313 10:11:53.788169 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:53 crc kubenswrapper[4632]: I0313 10:11:53.830169 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:54 crc kubenswrapper[4632]: I0313 10:11:54.711761 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-gdt8x" Mar 13 10:11:56 crc kubenswrapper[4632]: I0313 10:11:56.030467 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:56 crc kubenswrapper[4632]: I0313 10:11:56.030533 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:56 crc kubenswrapper[4632]: I0313 10:11:56.073766 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:56 crc kubenswrapper[4632]: I0313 10:11:56.720582 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vwgfr" Mar 13 10:11:56 crc kubenswrapper[4632]: I0313 10:11:56.751532 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-sbzcm" Mar 13 10:11:56 crc kubenswrapper[4632]: I0313 10:11:56.803711 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxs5z"] Mar 13 10:11:57 crc kubenswrapper[4632]: I0313 10:11:57.126315 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:11:57 crc kubenswrapper[4632]: I0313 10:11:57.126396 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:11:57 crc kubenswrapper[4632]: I0313 10:11:57.166084 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:11:57 crc kubenswrapper[4632]: I0313 10:11:57.722544 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:12:00 crc kubenswrapper[4632]: I0313 10:12:00.134685 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556612-5t5ct"] Mar 13 10:12:00 crc kubenswrapper[4632]: I0313 10:12:00.135538 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556612-5t5ct" Mar 13 10:12:00 crc kubenswrapper[4632]: I0313 10:12:00.139370 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:12:00 crc kubenswrapper[4632]: I0313 10:12:00.140034 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:12:00 crc kubenswrapper[4632]: I0313 10:12:00.140197 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:12:00 crc kubenswrapper[4632]: I0313 10:12:00.145341 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556612-5t5ct"] Mar 13 10:12:00 crc kubenswrapper[4632]: I0313 10:12:00.277366 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6pkz\" (UniqueName: \"kubernetes.io/projected/050ee655-a62f-4991-b493-d98493762823-kube-api-access-v6pkz\") pod \"auto-csr-approver-29556612-5t5ct\" (UID: \"050ee655-a62f-4991-b493-d98493762823\") " pod="openshift-infra/auto-csr-approver-29556612-5t5ct" Mar 13 10:12:00 crc kubenswrapper[4632]: I0313 10:12:00.378852 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6pkz\" (UniqueName: \"kubernetes.io/projected/050ee655-a62f-4991-b493-d98493762823-kube-api-access-v6pkz\") pod \"auto-csr-approver-29556612-5t5ct\" (UID: \"050ee655-a62f-4991-b493-d98493762823\") " pod="openshift-infra/auto-csr-approver-29556612-5t5ct" Mar 13 10:12:00 crc kubenswrapper[4632]: I0313 10:12:00.400390 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6pkz\" (UniqueName: \"kubernetes.io/projected/050ee655-a62f-4991-b493-d98493762823-kube-api-access-v6pkz\") pod \"auto-csr-approver-29556612-5t5ct\" (UID: \"050ee655-a62f-4991-b493-d98493762823\") " pod="openshift-infra/auto-csr-approver-29556612-5t5ct" Mar 13 10:12:00 crc kubenswrapper[4632]: I0313 10:12:00.459535 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556612-5t5ct" Mar 13 10:12:00 crc kubenswrapper[4632]: I0313 10:12:00.685043 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556612-5t5ct"] Mar 13 10:12:01 crc kubenswrapper[4632]: I0313 10:12:01.714183 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556612-5t5ct" event={"ID":"050ee655-a62f-4991-b493-d98493762823","Type":"ContainerStarted","Data":"72656c6d5a8d1a8b63e7d9f1bae015750738acdded0ca8534eb3a2e0a68316b7"} Mar 13 10:12:02 crc kubenswrapper[4632]: I0313 10:12:02.723204 4632 generic.go:334] "Generic (PLEG): container finished" podID="050ee655-a62f-4991-b493-d98493762823" containerID="3025e6a57984dbcc7f1272476cb4a6a1339dea799f52af43239e5a72f7479138" exitCode=0 Mar 13 10:12:02 crc kubenswrapper[4632]: I0313 10:12:02.723324 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556612-5t5ct" event={"ID":"050ee655-a62f-4991-b493-d98493762823","Type":"ContainerDied","Data":"3025e6a57984dbcc7f1272476cb4a6a1339dea799f52af43239e5a72f7479138"} Mar 13 10:12:03 crc kubenswrapper[4632]: I0313 10:12:03.697144 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7ksc5" Mar 13 10:12:04 crc kubenswrapper[4632]: I0313 10:12:04.030348 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556612-5t5ct" Mar 13 10:12:04 crc kubenswrapper[4632]: I0313 10:12:04.153389 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6pkz\" (UniqueName: \"kubernetes.io/projected/050ee655-a62f-4991-b493-d98493762823-kube-api-access-v6pkz\") pod \"050ee655-a62f-4991-b493-d98493762823\" (UID: \"050ee655-a62f-4991-b493-d98493762823\") " Mar 13 10:12:04 crc kubenswrapper[4632]: I0313 10:12:04.160662 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/050ee655-a62f-4991-b493-d98493762823-kube-api-access-v6pkz" (OuterVolumeSpecName: "kube-api-access-v6pkz") pod "050ee655-a62f-4991-b493-d98493762823" (UID: "050ee655-a62f-4991-b493-d98493762823"). InnerVolumeSpecName "kube-api-access-v6pkz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:12:04 crc kubenswrapper[4632]: I0313 10:12:04.255682 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6pkz\" (UniqueName: \"kubernetes.io/projected/050ee655-a62f-4991-b493-d98493762823-kube-api-access-v6pkz\") on node \"crc\" DevicePath \"\"" Mar 13 10:12:04 crc kubenswrapper[4632]: I0313 10:12:04.740458 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556612-5t5ct" event={"ID":"050ee655-a62f-4991-b493-d98493762823","Type":"ContainerDied","Data":"72656c6d5a8d1a8b63e7d9f1bae015750738acdded0ca8534eb3a2e0a68316b7"} Mar 13 10:12:04 crc kubenswrapper[4632]: I0313 10:12:04.741166 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72656c6d5a8d1a8b63e7d9f1bae015750738acdded0ca8534eb3a2e0a68316b7" Mar 13 10:12:04 crc kubenswrapper[4632]: I0313 10:12:04.740567 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556612-5t5ct" Mar 13 10:12:05 crc kubenswrapper[4632]: I0313 10:12:05.082234 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556606-mkrp2"] Mar 13 10:12:05 crc kubenswrapper[4632]: I0313 10:12:05.087474 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556606-mkrp2"] Mar 13 10:12:06 crc kubenswrapper[4632]: I0313 10:12:06.052502 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c822257d-9d2f-4b6f-87de-131de5cd0efe" path="/var/lib/kubelet/pods/c822257d-9d2f-4b6f-87de-131de5cd0efe/volumes" Mar 13 10:12:10 crc kubenswrapper[4632]: I0313 10:12:10.461195 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:12:10 crc kubenswrapper[4632]: I0313 10:12:10.461720 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:12:10 crc kubenswrapper[4632]: I0313 10:12:10.461786 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:12:10 crc kubenswrapper[4632]: I0313 10:12:10.463279 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e4989d70178427347867288c3fc7b62a339fa6ecdddde954f719a53f3db7fe17"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:12:10 crc kubenswrapper[4632]: I0313 10:12:10.463523 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://e4989d70178427347867288c3fc7b62a339fa6ecdddde954f719a53f3db7fe17" gracePeriod=600 Mar 13 10:12:10 crc kubenswrapper[4632]: I0313 10:12:10.786701 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="e4989d70178427347867288c3fc7b62a339fa6ecdddde954f719a53f3db7fe17" exitCode=0 Mar 13 10:12:10 crc kubenswrapper[4632]: I0313 10:12:10.786998 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"e4989d70178427347867288c3fc7b62a339fa6ecdddde954f719a53f3db7fe17"} Mar 13 10:12:10 crc kubenswrapper[4632]: I0313 10:12:10.787681 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"313e3b067f9ea051953ab56cbddeb09cc8cceb68240f33ca492d13584077681c"} Mar 13 10:12:10 crc kubenswrapper[4632]: I0313 10:12:10.787850 4632 scope.go:117] "RemoveContainer" containerID="8d890c96abcbe37c4f2e487a63e4f0d5f48c462a6fee6b8b1930384bdbfebee7" Mar 13 10:12:21 crc kubenswrapper[4632]: I0313 10:12:21.852207 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" podUID="f56fc09a-e2b7-46db-b938-f276df3f033e" containerName="registry" containerID="cri-o://1081e88c7001e64d6f95133ef3938fcdaf6163c9ecf6555e86dc52149387161f" gracePeriod=30 Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.197567 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.305730 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mprmj\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-kube-api-access-mprmj\") pod \"f56fc09a-e2b7-46db-b938-f276df3f033e\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.305785 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-bound-sa-token\") pod \"f56fc09a-e2b7-46db-b938-f276df3f033e\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.306013 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"f56fc09a-e2b7-46db-b938-f276df3f033e\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.306057 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-registry-tls\") pod \"f56fc09a-e2b7-46db-b938-f276df3f033e\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.306091 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f56fc09a-e2b7-46db-b938-f276df3f033e-trusted-ca\") pod \"f56fc09a-e2b7-46db-b938-f276df3f033e\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.306113 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f56fc09a-e2b7-46db-b938-f276df3f033e-registry-certificates\") pod \"f56fc09a-e2b7-46db-b938-f276df3f033e\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.306152 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f56fc09a-e2b7-46db-b938-f276df3f033e-installation-pull-secrets\") pod \"f56fc09a-e2b7-46db-b938-f276df3f033e\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.306193 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f56fc09a-e2b7-46db-b938-f276df3f033e-ca-trust-extracted\") pod \"f56fc09a-e2b7-46db-b938-f276df3f033e\" (UID: \"f56fc09a-e2b7-46db-b938-f276df3f033e\") " Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.307117 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56fc09a-e2b7-46db-b938-f276df3f033e-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "f56fc09a-e2b7-46db-b938-f276df3f033e" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.307288 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f56fc09a-e2b7-46db-b938-f276df3f033e-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "f56fc09a-e2b7-46db-b938-f276df3f033e" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.314904 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "f56fc09a-e2b7-46db-b938-f276df3f033e" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.315431 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "f56fc09a-e2b7-46db-b938-f276df3f033e" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.315633 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-kube-api-access-mprmj" (OuterVolumeSpecName: "kube-api-access-mprmj") pod "f56fc09a-e2b7-46db-b938-f276df3f033e" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e"). InnerVolumeSpecName "kube-api-access-mprmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.317417 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "f56fc09a-e2b7-46db-b938-f276df3f033e" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.317816 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f56fc09a-e2b7-46db-b938-f276df3f033e-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "f56fc09a-e2b7-46db-b938-f276df3f033e" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.334917 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f56fc09a-e2b7-46db-b938-f276df3f033e-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "f56fc09a-e2b7-46db-b938-f276df3f033e" (UID: "f56fc09a-e2b7-46db-b938-f276df3f033e"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.408009 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mprmj\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-kube-api-access-mprmj\") on node \"crc\" DevicePath \"\"" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.408050 4632 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.408065 4632 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/f56fc09a-e2b7-46db-b938-f276df3f033e-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.408078 4632 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f56fc09a-e2b7-46db-b938-f276df3f033e-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.408090 4632 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/f56fc09a-e2b7-46db-b938-f276df3f033e-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.408101 4632 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/f56fc09a-e2b7-46db-b938-f276df3f033e-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.408112 4632 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/f56fc09a-e2b7-46db-b938-f276df3f033e-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.860732 4632 generic.go:334] "Generic (PLEG): container finished" podID="f56fc09a-e2b7-46db-b938-f276df3f033e" containerID="1081e88c7001e64d6f95133ef3938fcdaf6163c9ecf6555e86dc52149387161f" exitCode=0 Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.860782 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.860776 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" event={"ID":"f56fc09a-e2b7-46db-b938-f276df3f033e","Type":"ContainerDied","Data":"1081e88c7001e64d6f95133ef3938fcdaf6163c9ecf6555e86dc52149387161f"} Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.861006 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-fxs5z" event={"ID":"f56fc09a-e2b7-46db-b938-f276df3f033e","Type":"ContainerDied","Data":"6db13fe4cd83b1210971879bf1313cee58732376958e857687de7da1568c6519"} Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.861058 4632 scope.go:117] "RemoveContainer" containerID="1081e88c7001e64d6f95133ef3938fcdaf6163c9ecf6555e86dc52149387161f" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.894049 4632 scope.go:117] "RemoveContainer" containerID="1081e88c7001e64d6f95133ef3938fcdaf6163c9ecf6555e86dc52149387161f" Mar 13 10:12:22 crc kubenswrapper[4632]: E0313 10:12:22.894672 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1081e88c7001e64d6f95133ef3938fcdaf6163c9ecf6555e86dc52149387161f\": container with ID starting with 1081e88c7001e64d6f95133ef3938fcdaf6163c9ecf6555e86dc52149387161f not found: ID does not exist" containerID="1081e88c7001e64d6f95133ef3938fcdaf6163c9ecf6555e86dc52149387161f" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.894694 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1081e88c7001e64d6f95133ef3938fcdaf6163c9ecf6555e86dc52149387161f"} err="failed to get container status \"1081e88c7001e64d6f95133ef3938fcdaf6163c9ecf6555e86dc52149387161f\": rpc error: code = NotFound desc = could not find container \"1081e88c7001e64d6f95133ef3938fcdaf6163c9ecf6555e86dc52149387161f\": container with ID starting with 1081e88c7001e64d6f95133ef3938fcdaf6163c9ecf6555e86dc52149387161f not found: ID does not exist" Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.895982 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxs5z"] Mar 13 10:12:22 crc kubenswrapper[4632]: I0313 10:12:22.899735 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-fxs5z"] Mar 13 10:12:24 crc kubenswrapper[4632]: I0313 10:12:24.050601 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f56fc09a-e2b7-46db-b938-f276df3f033e" path="/var/lib/kubelet/pods/f56fc09a-e2b7-46db-b938-f276df3f033e/volumes" Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.142225 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556614-7pzwt"] Mar 13 10:14:00 crc kubenswrapper[4632]: E0313 10:14:00.143645 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f56fc09a-e2b7-46db-b938-f276df3f033e" containerName="registry" Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.143672 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f56fc09a-e2b7-46db-b938-f276df3f033e" containerName="registry" Mar 13 10:14:00 crc kubenswrapper[4632]: E0313 10:14:00.143694 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="050ee655-a62f-4991-b493-d98493762823" containerName="oc" Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.143701 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="050ee655-a62f-4991-b493-d98493762823" containerName="oc" Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.143808 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="050ee655-a62f-4991-b493-d98493762823" containerName="oc" Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.143834 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f56fc09a-e2b7-46db-b938-f276df3f033e" containerName="registry" Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.144386 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556614-7pzwt" Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.147555 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.147759 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.151639 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.157680 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556614-7pzwt"] Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.247288 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z79d7\" (UniqueName: \"kubernetes.io/projected/f80bfe67-be24-45e3-9e57-b67389f8cc63-kube-api-access-z79d7\") pod \"auto-csr-approver-29556614-7pzwt\" (UID: \"f80bfe67-be24-45e3-9e57-b67389f8cc63\") " pod="openshift-infra/auto-csr-approver-29556614-7pzwt" Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.348102 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z79d7\" (UniqueName: \"kubernetes.io/projected/f80bfe67-be24-45e3-9e57-b67389f8cc63-kube-api-access-z79d7\") pod \"auto-csr-approver-29556614-7pzwt\" (UID: \"f80bfe67-be24-45e3-9e57-b67389f8cc63\") " pod="openshift-infra/auto-csr-approver-29556614-7pzwt" Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.371011 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z79d7\" (UniqueName: \"kubernetes.io/projected/f80bfe67-be24-45e3-9e57-b67389f8cc63-kube-api-access-z79d7\") pod \"auto-csr-approver-29556614-7pzwt\" (UID: \"f80bfe67-be24-45e3-9e57-b67389f8cc63\") " pod="openshift-infra/auto-csr-approver-29556614-7pzwt" Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.460902 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556614-7pzwt" Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.674674 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556614-7pzwt"] Mar 13 10:14:00 crc kubenswrapper[4632]: I0313 10:14:00.685619 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:14:01 crc kubenswrapper[4632]: I0313 10:14:01.439893 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556614-7pzwt" event={"ID":"f80bfe67-be24-45e3-9e57-b67389f8cc63","Type":"ContainerStarted","Data":"9e73f3e8b4f5dc8b5a5a916940efe609dd694e9b9269560be1517d95ab449710"} Mar 13 10:14:02 crc kubenswrapper[4632]: I0313 10:14:02.447321 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556614-7pzwt" event={"ID":"f80bfe67-be24-45e3-9e57-b67389f8cc63","Type":"ContainerDied","Data":"fcf5d9f69f7435b287086bfcb908c42e9330ebc2ef407226d11b60f145efd8de"} Mar 13 10:14:02 crc kubenswrapper[4632]: I0313 10:14:02.447799 4632 generic.go:334] "Generic (PLEG): container finished" podID="f80bfe67-be24-45e3-9e57-b67389f8cc63" containerID="fcf5d9f69f7435b287086bfcb908c42e9330ebc2ef407226d11b60f145efd8de" exitCode=0 Mar 13 10:14:03 crc kubenswrapper[4632]: I0313 10:14:03.657284 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556614-7pzwt" Mar 13 10:14:03 crc kubenswrapper[4632]: I0313 10:14:03.795467 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z79d7\" (UniqueName: \"kubernetes.io/projected/f80bfe67-be24-45e3-9e57-b67389f8cc63-kube-api-access-z79d7\") pod \"f80bfe67-be24-45e3-9e57-b67389f8cc63\" (UID: \"f80bfe67-be24-45e3-9e57-b67389f8cc63\") " Mar 13 10:14:03 crc kubenswrapper[4632]: I0313 10:14:03.805831 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f80bfe67-be24-45e3-9e57-b67389f8cc63-kube-api-access-z79d7" (OuterVolumeSpecName: "kube-api-access-z79d7") pod "f80bfe67-be24-45e3-9e57-b67389f8cc63" (UID: "f80bfe67-be24-45e3-9e57-b67389f8cc63"). InnerVolumeSpecName "kube-api-access-z79d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:14:03 crc kubenswrapper[4632]: I0313 10:14:03.896852 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z79d7\" (UniqueName: \"kubernetes.io/projected/f80bfe67-be24-45e3-9e57-b67389f8cc63-kube-api-access-z79d7\") on node \"crc\" DevicePath \"\"" Mar 13 10:14:04 crc kubenswrapper[4632]: I0313 10:14:04.461277 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556614-7pzwt" event={"ID":"f80bfe67-be24-45e3-9e57-b67389f8cc63","Type":"ContainerDied","Data":"9e73f3e8b4f5dc8b5a5a916940efe609dd694e9b9269560be1517d95ab449710"} Mar 13 10:14:04 crc kubenswrapper[4632]: I0313 10:14:04.461598 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e73f3e8b4f5dc8b5a5a916940efe609dd694e9b9269560be1517d95ab449710" Mar 13 10:14:04 crc kubenswrapper[4632]: I0313 10:14:04.461491 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556614-7pzwt" Mar 13 10:14:04 crc kubenswrapper[4632]: I0313 10:14:04.722464 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556608-9kzfk"] Mar 13 10:14:04 crc kubenswrapper[4632]: I0313 10:14:04.730589 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556608-9kzfk"] Mar 13 10:14:06 crc kubenswrapper[4632]: I0313 10:14:06.056068 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37ab6711-478f-4cc7-b9a4-c9baa126b1a3" path="/var/lib/kubelet/pods/37ab6711-478f-4cc7-b9a4-c9baa126b1a3/volumes" Mar 13 10:14:10 crc kubenswrapper[4632]: I0313 10:14:10.461215 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:14:10 crc kubenswrapper[4632]: I0313 10:14:10.461820 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:14:40 crc kubenswrapper[4632]: I0313 10:14:40.461386 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:14:40 crc kubenswrapper[4632]: I0313 10:14:40.462082 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:14:56 crc kubenswrapper[4632]: I0313 10:14:56.573798 4632 scope.go:117] "RemoveContainer" containerID="24d957ae4862987ed76c21db8796ae914a7d2beca83397bc3f90816dc051c956" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.139474 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm"] Mar 13 10:15:00 crc kubenswrapper[4632]: E0313 10:15:00.140696 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f80bfe67-be24-45e3-9e57-b67389f8cc63" containerName="oc" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.140800 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f80bfe67-be24-45e3-9e57-b67389f8cc63" containerName="oc" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.141072 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f80bfe67-be24-45e3-9e57-b67389f8cc63" containerName="oc" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.141647 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.144234 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.144852 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.149841 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm"] Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.220306 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c3c6392-454c-4131-90a0-6584565cef4c-config-volume\") pod \"collect-profiles-29556615-zj2fm\" (UID: \"3c3c6392-454c-4131-90a0-6584565cef4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.220352 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdggs\" (UniqueName: \"kubernetes.io/projected/3c3c6392-454c-4131-90a0-6584565cef4c-kube-api-access-mdggs\") pod \"collect-profiles-29556615-zj2fm\" (UID: \"3c3c6392-454c-4131-90a0-6584565cef4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.220410 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c3c6392-454c-4131-90a0-6584565cef4c-secret-volume\") pod \"collect-profiles-29556615-zj2fm\" (UID: \"3c3c6392-454c-4131-90a0-6584565cef4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.321158 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c3c6392-454c-4131-90a0-6584565cef4c-config-volume\") pod \"collect-profiles-29556615-zj2fm\" (UID: \"3c3c6392-454c-4131-90a0-6584565cef4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.321221 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdggs\" (UniqueName: \"kubernetes.io/projected/3c3c6392-454c-4131-90a0-6584565cef4c-kube-api-access-mdggs\") pod \"collect-profiles-29556615-zj2fm\" (UID: \"3c3c6392-454c-4131-90a0-6584565cef4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.321256 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c3c6392-454c-4131-90a0-6584565cef4c-secret-volume\") pod \"collect-profiles-29556615-zj2fm\" (UID: \"3c3c6392-454c-4131-90a0-6584565cef4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.322052 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c3c6392-454c-4131-90a0-6584565cef4c-config-volume\") pod \"collect-profiles-29556615-zj2fm\" (UID: \"3c3c6392-454c-4131-90a0-6584565cef4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.327262 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c3c6392-454c-4131-90a0-6584565cef4c-secret-volume\") pod \"collect-profiles-29556615-zj2fm\" (UID: \"3c3c6392-454c-4131-90a0-6584565cef4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.338478 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdggs\" (UniqueName: \"kubernetes.io/projected/3c3c6392-454c-4131-90a0-6584565cef4c-kube-api-access-mdggs\") pod \"collect-profiles-29556615-zj2fm\" (UID: \"3c3c6392-454c-4131-90a0-6584565cef4c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.465441 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.685614 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm"] Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.881854 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" event={"ID":"3c3c6392-454c-4131-90a0-6584565cef4c","Type":"ContainerStarted","Data":"6acdfd407705651773e15ca9493f2efdac886dce6f04123c798b57f93aa775b6"} Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.897628 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" event={"ID":"3c3c6392-454c-4131-90a0-6584565cef4c","Type":"ContainerStarted","Data":"d111a5a6d8b3ddefc63155935fa715861914a40029da3f7348682af126904c65"} Mar 13 10:15:00 crc kubenswrapper[4632]: I0313 10:15:00.916139 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" podStartSLOduration=0.916109748 podStartE2EDuration="916.109748ms" podCreationTimestamp="2026-03-13 10:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:15:00.912088434 +0000 UTC m=+674.934618587" watchObservedRunningTime="2026-03-13 10:15:00.916109748 +0000 UTC m=+674.938639881" Mar 13 10:15:01 crc kubenswrapper[4632]: I0313 10:15:01.897244 4632 generic.go:334] "Generic (PLEG): container finished" podID="3c3c6392-454c-4131-90a0-6584565cef4c" containerID="6acdfd407705651773e15ca9493f2efdac886dce6f04123c798b57f93aa775b6" exitCode=0 Mar 13 10:15:01 crc kubenswrapper[4632]: I0313 10:15:01.897290 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" event={"ID":"3c3c6392-454c-4131-90a0-6584565cef4c","Type":"ContainerDied","Data":"6acdfd407705651773e15ca9493f2efdac886dce6f04123c798b57f93aa775b6"} Mar 13 10:15:03 crc kubenswrapper[4632]: I0313 10:15:03.084997 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" Mar 13 10:15:03 crc kubenswrapper[4632]: I0313 10:15:03.262486 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c3c6392-454c-4131-90a0-6584565cef4c-secret-volume\") pod \"3c3c6392-454c-4131-90a0-6584565cef4c\" (UID: \"3c3c6392-454c-4131-90a0-6584565cef4c\") " Mar 13 10:15:03 crc kubenswrapper[4632]: I0313 10:15:03.262872 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdggs\" (UniqueName: \"kubernetes.io/projected/3c3c6392-454c-4131-90a0-6584565cef4c-kube-api-access-mdggs\") pod \"3c3c6392-454c-4131-90a0-6584565cef4c\" (UID: \"3c3c6392-454c-4131-90a0-6584565cef4c\") " Mar 13 10:15:03 crc kubenswrapper[4632]: I0313 10:15:03.263827 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c3c6392-454c-4131-90a0-6584565cef4c-config-volume\") pod \"3c3c6392-454c-4131-90a0-6584565cef4c\" (UID: \"3c3c6392-454c-4131-90a0-6584565cef4c\") " Mar 13 10:15:03 crc kubenswrapper[4632]: I0313 10:15:03.264583 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3c3c6392-454c-4131-90a0-6584565cef4c-config-volume" (OuterVolumeSpecName: "config-volume") pod "3c3c6392-454c-4131-90a0-6584565cef4c" (UID: "3c3c6392-454c-4131-90a0-6584565cef4c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:15:03 crc kubenswrapper[4632]: I0313 10:15:03.269231 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3c3c6392-454c-4131-90a0-6584565cef4c-kube-api-access-mdggs" (OuterVolumeSpecName: "kube-api-access-mdggs") pod "3c3c6392-454c-4131-90a0-6584565cef4c" (UID: "3c3c6392-454c-4131-90a0-6584565cef4c"). InnerVolumeSpecName "kube-api-access-mdggs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:15:03 crc kubenswrapper[4632]: I0313 10:15:03.269616 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3c3c6392-454c-4131-90a0-6584565cef4c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3c3c6392-454c-4131-90a0-6584565cef4c" (UID: "3c3c6392-454c-4131-90a0-6584565cef4c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:15:03 crc kubenswrapper[4632]: I0313 10:15:03.364793 4632 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3c3c6392-454c-4131-90a0-6584565cef4c-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:15:03 crc kubenswrapper[4632]: I0313 10:15:03.365068 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdggs\" (UniqueName: \"kubernetes.io/projected/3c3c6392-454c-4131-90a0-6584565cef4c-kube-api-access-mdggs\") on node \"crc\" DevicePath \"\"" Mar 13 10:15:03 crc kubenswrapper[4632]: I0313 10:15:03.365186 4632 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c3c6392-454c-4131-90a0-6584565cef4c-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:15:03 crc kubenswrapper[4632]: I0313 10:15:03.907670 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" event={"ID":"3c3c6392-454c-4131-90a0-6584565cef4c","Type":"ContainerDied","Data":"d111a5a6d8b3ddefc63155935fa715861914a40029da3f7348682af126904c65"} Mar 13 10:15:03 crc kubenswrapper[4632]: I0313 10:15:03.907700 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm" Mar 13 10:15:03 crc kubenswrapper[4632]: I0313 10:15:03.907708 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d111a5a6d8b3ddefc63155935fa715861914a40029da3f7348682af126904c65" Mar 13 10:15:10 crc kubenswrapper[4632]: I0313 10:15:10.460725 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:15:10 crc kubenswrapper[4632]: I0313 10:15:10.461163 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:15:10 crc kubenswrapper[4632]: I0313 10:15:10.461209 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:15:10 crc kubenswrapper[4632]: I0313 10:15:10.947362 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"313e3b067f9ea051953ab56cbddeb09cc8cceb68240f33ca492d13584077681c"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:15:10 crc kubenswrapper[4632]: I0313 10:15:10.947470 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://313e3b067f9ea051953ab56cbddeb09cc8cceb68240f33ca492d13584077681c" gracePeriod=600 Mar 13 10:15:11 crc kubenswrapper[4632]: I0313 10:15:11.956053 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="313e3b067f9ea051953ab56cbddeb09cc8cceb68240f33ca492d13584077681c" exitCode=0 Mar 13 10:15:11 crc kubenswrapper[4632]: I0313 10:15:11.956123 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"313e3b067f9ea051953ab56cbddeb09cc8cceb68240f33ca492d13584077681c"} Mar 13 10:15:11 crc kubenswrapper[4632]: I0313 10:15:11.956427 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"7fcd863f1a2b3af4768aa1d32979163bc846d3d472acea1e8c27ffcf3dfe0ffc"} Mar 13 10:15:11 crc kubenswrapper[4632]: I0313 10:15:11.956451 4632 scope.go:117] "RemoveContainer" containerID="e4989d70178427347867288c3fc7b62a339fa6ecdddde954f719a53f3db7fe17" Mar 13 10:15:56 crc kubenswrapper[4632]: I0313 10:15:56.617478 4632 scope.go:117] "RemoveContainer" containerID="481e1788f663e81921b410cd12a9e3666afaa2b706dda68096288fee3498f2fa" Mar 13 10:15:56 crc kubenswrapper[4632]: I0313 10:15:56.656861 4632 scope.go:117] "RemoveContainer" containerID="83de0881072cb52ab7a7fbd2d8ef18cbb3eb4eb7897fd1301bfd2cbf304913b7" Mar 13 10:16:00 crc kubenswrapper[4632]: I0313 10:16:00.147295 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556616-8xbbs"] Mar 13 10:16:00 crc kubenswrapper[4632]: E0313 10:16:00.149556 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3c3c6392-454c-4131-90a0-6584565cef4c" containerName="collect-profiles" Mar 13 10:16:00 crc kubenswrapper[4632]: I0313 10:16:00.149686 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3c3c6392-454c-4131-90a0-6584565cef4c" containerName="collect-profiles" Mar 13 10:16:00 crc kubenswrapper[4632]: I0313 10:16:00.149864 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3c3c6392-454c-4131-90a0-6584565cef4c" containerName="collect-profiles" Mar 13 10:16:00 crc kubenswrapper[4632]: I0313 10:16:00.150411 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556616-8xbbs" Mar 13 10:16:00 crc kubenswrapper[4632]: I0313 10:16:00.155290 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:16:00 crc kubenswrapper[4632]: I0313 10:16:00.155680 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:16:00 crc kubenswrapper[4632]: I0313 10:16:00.158234 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:16:00 crc kubenswrapper[4632]: I0313 10:16:00.158880 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556616-8xbbs"] Mar 13 10:16:00 crc kubenswrapper[4632]: I0313 10:16:00.221195 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72j5z\" (UniqueName: \"kubernetes.io/projected/c21d462b-89d1-4844-9bfc-3f0cdf7727e9-kube-api-access-72j5z\") pod \"auto-csr-approver-29556616-8xbbs\" (UID: \"c21d462b-89d1-4844-9bfc-3f0cdf7727e9\") " pod="openshift-infra/auto-csr-approver-29556616-8xbbs" Mar 13 10:16:00 crc kubenswrapper[4632]: I0313 10:16:00.323427 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72j5z\" (UniqueName: \"kubernetes.io/projected/c21d462b-89d1-4844-9bfc-3f0cdf7727e9-kube-api-access-72j5z\") pod \"auto-csr-approver-29556616-8xbbs\" (UID: \"c21d462b-89d1-4844-9bfc-3f0cdf7727e9\") " pod="openshift-infra/auto-csr-approver-29556616-8xbbs" Mar 13 10:16:00 crc kubenswrapper[4632]: I0313 10:16:00.351091 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72j5z\" (UniqueName: \"kubernetes.io/projected/c21d462b-89d1-4844-9bfc-3f0cdf7727e9-kube-api-access-72j5z\") pod \"auto-csr-approver-29556616-8xbbs\" (UID: \"c21d462b-89d1-4844-9bfc-3f0cdf7727e9\") " pod="openshift-infra/auto-csr-approver-29556616-8xbbs" Mar 13 10:16:00 crc kubenswrapper[4632]: I0313 10:16:00.468376 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556616-8xbbs" Mar 13 10:16:00 crc kubenswrapper[4632]: I0313 10:16:00.685610 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556616-8xbbs"] Mar 13 10:16:01 crc kubenswrapper[4632]: I0313 10:16:01.221805 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556616-8xbbs" event={"ID":"c21d462b-89d1-4844-9bfc-3f0cdf7727e9","Type":"ContainerStarted","Data":"43de2e95e7518fc84e89cd9a199fc2193c16fb5445fb6d9085f1feeba22ffbc3"} Mar 13 10:16:02 crc kubenswrapper[4632]: I0313 10:16:02.229320 4632 generic.go:334] "Generic (PLEG): container finished" podID="c21d462b-89d1-4844-9bfc-3f0cdf7727e9" containerID="6a34c241348123944aa499915ed71c016789c868e3e563c2a1cb71763ed56ad8" exitCode=0 Mar 13 10:16:02 crc kubenswrapper[4632]: I0313 10:16:02.229364 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556616-8xbbs" event={"ID":"c21d462b-89d1-4844-9bfc-3f0cdf7727e9","Type":"ContainerDied","Data":"6a34c241348123944aa499915ed71c016789c868e3e563c2a1cb71763ed56ad8"} Mar 13 10:16:03 crc kubenswrapper[4632]: I0313 10:16:03.436206 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556616-8xbbs" Mar 13 10:16:03 crc kubenswrapper[4632]: I0313 10:16:03.568117 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72j5z\" (UniqueName: \"kubernetes.io/projected/c21d462b-89d1-4844-9bfc-3f0cdf7727e9-kube-api-access-72j5z\") pod \"c21d462b-89d1-4844-9bfc-3f0cdf7727e9\" (UID: \"c21d462b-89d1-4844-9bfc-3f0cdf7727e9\") " Mar 13 10:16:03 crc kubenswrapper[4632]: I0313 10:16:03.575265 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c21d462b-89d1-4844-9bfc-3f0cdf7727e9-kube-api-access-72j5z" (OuterVolumeSpecName: "kube-api-access-72j5z") pod "c21d462b-89d1-4844-9bfc-3f0cdf7727e9" (UID: "c21d462b-89d1-4844-9bfc-3f0cdf7727e9"). InnerVolumeSpecName "kube-api-access-72j5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:16:03 crc kubenswrapper[4632]: I0313 10:16:03.669236 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72j5z\" (UniqueName: \"kubernetes.io/projected/c21d462b-89d1-4844-9bfc-3f0cdf7727e9-kube-api-access-72j5z\") on node \"crc\" DevicePath \"\"" Mar 13 10:16:04 crc kubenswrapper[4632]: I0313 10:16:04.241309 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556616-8xbbs" event={"ID":"c21d462b-89d1-4844-9bfc-3f0cdf7727e9","Type":"ContainerDied","Data":"43de2e95e7518fc84e89cd9a199fc2193c16fb5445fb6d9085f1feeba22ffbc3"} Mar 13 10:16:04 crc kubenswrapper[4632]: I0313 10:16:04.241361 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556616-8xbbs" Mar 13 10:16:04 crc kubenswrapper[4632]: I0313 10:16:04.241360 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43de2e95e7518fc84e89cd9a199fc2193c16fb5445fb6d9085f1feeba22ffbc3" Mar 13 10:16:04 crc kubenswrapper[4632]: I0313 10:16:04.498647 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556610-sg5bx"] Mar 13 10:16:04 crc kubenswrapper[4632]: I0313 10:16:04.502072 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556610-sg5bx"] Mar 13 10:16:06 crc kubenswrapper[4632]: I0313 10:16:06.051526 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="795727b7-7a2e-4e97-8707-aecf893fd332" path="/var/lib/kubelet/pods/795727b7-7a2e-4e97-8707-aecf893fd332/volumes" Mar 13 10:16:56 crc kubenswrapper[4632]: I0313 10:16:56.703586 4632 scope.go:117] "RemoveContainer" containerID="2858228a654d1c5c1b9a9a04d00ea882bfe929e6c810389040bc3c0ba67d7a46" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.424047 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-xg2df"] Mar 13 10:17:40 crc kubenswrapper[4632]: E0313 10:17:40.425058 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c21d462b-89d1-4844-9bfc-3f0cdf7727e9" containerName="oc" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.425084 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c21d462b-89d1-4844-9bfc-3f0cdf7727e9" containerName="oc" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.425252 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c21d462b-89d1-4844-9bfc-3f0cdf7727e9" containerName="oc" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.425743 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xg2df" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.434661 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.439017 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.442777 4632 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-l29dz" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.461514 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.461598 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.469676 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-xg2df"] Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.475466 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-kh4n9"] Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.476323 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-kh4n9" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.495384 4632 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-vbrx7" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.534051 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-kh4n9"] Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.545741 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-tjkbb"] Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.546649 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.554500 4632 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-4ndzq" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.573224 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-tjkbb"] Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.598568 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlwxc\" (UniqueName: \"kubernetes.io/projected/43729a96-008f-4af6-ba0d-d52f2f179c0b-kube-api-access-dlwxc\") pod \"cert-manager-858654f9db-kh4n9\" (UID: \"43729a96-008f-4af6-ba0d-d52f2f179c0b\") " pod="cert-manager/cert-manager-858654f9db-kh4n9" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.598841 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvlcv\" (UniqueName: \"kubernetes.io/projected/348f2814-4e97-4ec5-bcbb-35a868955687-kube-api-access-vvlcv\") pod \"cert-manager-cainjector-cf98fcc89-xg2df\" (UID: \"348f2814-4e97-4ec5-bcbb-35a868955687\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-xg2df" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.699876 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvlcv\" (UniqueName: \"kubernetes.io/projected/348f2814-4e97-4ec5-bcbb-35a868955687-kube-api-access-vvlcv\") pod \"cert-manager-cainjector-cf98fcc89-xg2df\" (UID: \"348f2814-4e97-4ec5-bcbb-35a868955687\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-xg2df" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.700176 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldz2c\" (UniqueName: \"kubernetes.io/projected/a0d52d98-fe87-4bc8-890e-5c5efb1f30d6-kube-api-access-ldz2c\") pod \"cert-manager-webhook-687f57d79b-tjkbb\" (UID: \"a0d52d98-fe87-4bc8-890e-5c5efb1f30d6\") " pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.700338 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlwxc\" (UniqueName: \"kubernetes.io/projected/43729a96-008f-4af6-ba0d-d52f2f179c0b-kube-api-access-dlwxc\") pod \"cert-manager-858654f9db-kh4n9\" (UID: \"43729a96-008f-4af6-ba0d-d52f2f179c0b\") " pod="cert-manager/cert-manager-858654f9db-kh4n9" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.721638 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvlcv\" (UniqueName: \"kubernetes.io/projected/348f2814-4e97-4ec5-bcbb-35a868955687-kube-api-access-vvlcv\") pod \"cert-manager-cainjector-cf98fcc89-xg2df\" (UID: \"348f2814-4e97-4ec5-bcbb-35a868955687\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-xg2df" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.721861 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlwxc\" (UniqueName: \"kubernetes.io/projected/43729a96-008f-4af6-ba0d-d52f2f179c0b-kube-api-access-dlwxc\") pod \"cert-manager-858654f9db-kh4n9\" (UID: \"43729a96-008f-4af6-ba0d-d52f2f179c0b\") " pod="cert-manager/cert-manager-858654f9db-kh4n9" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.756635 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xg2df" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.792190 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-kh4n9" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.801298 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldz2c\" (UniqueName: \"kubernetes.io/projected/a0d52d98-fe87-4bc8-890e-5c5efb1f30d6-kube-api-access-ldz2c\") pod \"cert-manager-webhook-687f57d79b-tjkbb\" (UID: \"a0d52d98-fe87-4bc8-890e-5c5efb1f30d6\") " pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.819916 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldz2c\" (UniqueName: \"kubernetes.io/projected/a0d52d98-fe87-4bc8-890e-5c5efb1f30d6-kube-api-access-ldz2c\") pod \"cert-manager-webhook-687f57d79b-tjkbb\" (UID: \"a0d52d98-fe87-4bc8-890e-5c5efb1f30d6\") " pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" Mar 13 10:17:40 crc kubenswrapper[4632]: I0313 10:17:40.883240 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" Mar 13 10:17:41 crc kubenswrapper[4632]: I0313 10:17:41.044516 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-xg2df"] Mar 13 10:17:41 crc kubenswrapper[4632]: I0313 10:17:41.131417 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-kh4n9"] Mar 13 10:17:41 crc kubenswrapper[4632]: I0313 10:17:41.168486 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-tjkbb"] Mar 13 10:17:41 crc kubenswrapper[4632]: I0313 10:17:41.985569 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" event={"ID":"a0d52d98-fe87-4bc8-890e-5c5efb1f30d6","Type":"ContainerStarted","Data":"82b21e36ecdc4d34750785c32c17a383e36d23b315fdc56d3caf042751d9a95f"} Mar 13 10:17:41 crc kubenswrapper[4632]: I0313 10:17:41.986874 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-kh4n9" event={"ID":"43729a96-008f-4af6-ba0d-d52f2f179c0b","Type":"ContainerStarted","Data":"a0c0f987f5c2a930ae5107733f0d58d6f5e01e7b219af529245d17a68a3343f0"} Mar 13 10:17:41 crc kubenswrapper[4632]: I0313 10:17:41.987668 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xg2df" event={"ID":"348f2814-4e97-4ec5-bcbb-35a868955687","Type":"ContainerStarted","Data":"6a7153edc397d786cbf41e71a3296cd574a4cfc60a16e139cb563a8ccd14fc5c"} Mar 13 10:17:46 crc kubenswrapper[4632]: I0313 10:17:46.014593 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-kh4n9" event={"ID":"43729a96-008f-4af6-ba0d-d52f2f179c0b","Type":"ContainerStarted","Data":"8229d816898988c328ae2fb2bab8f7b34737337cfe0afd10b5293704807d3da1"} Mar 13 10:17:46 crc kubenswrapper[4632]: I0313 10:17:46.017739 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xg2df" event={"ID":"348f2814-4e97-4ec5-bcbb-35a868955687","Type":"ContainerStarted","Data":"6665295b03f0a6068d3b75c34593a0f084b92ac8a39e54bf31273838d5112efd"} Mar 13 10:17:46 crc kubenswrapper[4632]: I0313 10:17:46.018722 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" event={"ID":"a0d52d98-fe87-4bc8-890e-5c5efb1f30d6","Type":"ContainerStarted","Data":"f1eb968bf0483e67b1bec8db44be891387eb9d4776a59eb6ccb079e224198d23"} Mar 13 10:17:46 crc kubenswrapper[4632]: I0313 10:17:46.019125 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" Mar 13 10:17:46 crc kubenswrapper[4632]: I0313 10:17:46.046549 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-kh4n9" podStartSLOduration=2.187180402 podStartE2EDuration="6.046529157s" podCreationTimestamp="2026-03-13 10:17:40 +0000 UTC" firstStartedPulling="2026-03-13 10:17:41.134428136 +0000 UTC m=+835.156958269" lastFinishedPulling="2026-03-13 10:17:44.993776891 +0000 UTC m=+839.016307024" observedRunningTime="2026-03-13 10:17:46.044114435 +0000 UTC m=+840.066644568" watchObservedRunningTime="2026-03-13 10:17:46.046529157 +0000 UTC m=+840.069059300" Mar 13 10:17:46 crc kubenswrapper[4632]: I0313 10:17:46.063078 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" podStartSLOduration=2.251709628 podStartE2EDuration="6.063057779s" podCreationTimestamp="2026-03-13 10:17:40 +0000 UTC" firstStartedPulling="2026-03-13 10:17:41.184753269 +0000 UTC m=+835.207283402" lastFinishedPulling="2026-03-13 10:17:44.99610142 +0000 UTC m=+839.018631553" observedRunningTime="2026-03-13 10:17:46.061563141 +0000 UTC m=+840.084093274" watchObservedRunningTime="2026-03-13 10:17:46.063057779 +0000 UTC m=+840.085587912" Mar 13 10:17:46 crc kubenswrapper[4632]: I0313 10:17:46.085800 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-xg2df" podStartSLOduration=2.160168593 podStartE2EDuration="6.085781709s" podCreationTimestamp="2026-03-13 10:17:40 +0000 UTC" firstStartedPulling="2026-03-13 10:17:41.062564882 +0000 UTC m=+835.085095025" lastFinishedPulling="2026-03-13 10:17:44.988177988 +0000 UTC m=+839.010708141" observedRunningTime="2026-03-13 10:17:46.081456258 +0000 UTC m=+840.103986391" watchObservedRunningTime="2026-03-13 10:17:46.085781709 +0000 UTC m=+840.108311862" Mar 13 10:17:49 crc kubenswrapper[4632]: I0313 10:17:49.260443 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qb725"] Mar 13 10:17:49 crc kubenswrapper[4632]: I0313 10:17:49.262534 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovn-controller" containerID="cri-o://7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705" gracePeriod=30 Mar 13 10:17:49 crc kubenswrapper[4632]: I0313 10:17:49.262647 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="sbdb" containerID="cri-o://e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d" gracePeriod=30 Mar 13 10:17:49 crc kubenswrapper[4632]: I0313 10:17:49.262570 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="nbdb" containerID="cri-o://af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f" gracePeriod=30 Mar 13 10:17:49 crc kubenswrapper[4632]: I0313 10:17:49.262703 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5" gracePeriod=30 Mar 13 10:17:49 crc kubenswrapper[4632]: I0313 10:17:49.262745 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="northd" containerID="cri-o://1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da" gracePeriod=30 Mar 13 10:17:49 crc kubenswrapper[4632]: I0313 10:17:49.262805 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovn-acl-logging" containerID="cri-o://32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719" gracePeriod=30 Mar 13 10:17:49 crc kubenswrapper[4632]: I0313 10:17:49.262821 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="kube-rbac-proxy-node" containerID="cri-o://cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b" gracePeriod=30 Mar 13 10:17:49 crc kubenswrapper[4632]: I0313 10:17:49.316107 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" containerID="cri-o://166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776" gracePeriod=30 Mar 13 10:17:49 crc kubenswrapper[4632]: I0313 10:17:49.989098 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/3.log" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.001616 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovn-acl-logging/0.log" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.002423 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovn-controller/0.log" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.003211 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045074 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-cni-netd\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045124 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-kubelet\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045140 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-run-ovn-kubernetes\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045172 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-var-lib-cni-networks-ovn-kubernetes\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045190 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-openvswitch\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045206 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-etc-openvswitch\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045224 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-run-netns\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045241 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-cni-bin\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045257 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-slash\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045283 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-var-lib-openvswitch\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045309 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-ovnkube-config\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045323 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-log-socket\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045353 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-node-log\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045372 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3b40c6b3-0061-4224-82d5-3ccf67998722-ovn-node-metrics-cert\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045386 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-systemd-units\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045401 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-env-overrides\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045418 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-ovnkube-script-lib\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045436 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dj6cl\" (UniqueName: \"kubernetes.io/projected/3b40c6b3-0061-4224-82d5-3ccf67998722-kube-api-access-dj6cl\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.045464 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-systemd\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.046286 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-slash" (OuterVolumeSpecName: "host-slash") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.046326 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.046355 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.046374 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.046395 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.046419 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.046436 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.046452 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.046468 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.047248 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-ovn\") pod \"3b40c6b3-0061-4224-82d5-3ccf67998722\" (UID: \"3b40c6b3-0061-4224-82d5-3ccf67998722\") " Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.047415 4632 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.047428 4632 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.047437 4632 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.047447 4632 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.047456 4632 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.047465 4632 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.047474 4632 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.047482 4632 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.047490 4632 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-host-slash\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.047515 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.047536 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.047960 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.048287 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.049398 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.049540 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.050442 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-log-socket" (OuterVolumeSpecName: "log-socket") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.050539 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-node-log" (OuterVolumeSpecName: "node-log") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.066390 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b40c6b3-0061-4224-82d5-3ccf67998722-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.066673 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b40c6b3-0061-4224-82d5-3ccf67998722-kube-api-access-dj6cl" (OuterVolumeSpecName: "kube-api-access-dj6cl") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "kube-api-access-dj6cl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.077256 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovnkube-controller/3.log" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.078852 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "3b40c6b3-0061-4224-82d5-3ccf67998722" (UID: "3b40c6b3-0061-4224-82d5-3ccf67998722"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.081712 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-khrch"] Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.082008 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="sbdb" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082027 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="sbdb" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.082040 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="kubecfg-setup" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082047 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="kubecfg-setup" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.082055 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovn-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082062 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovn-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.082071 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovn-acl-logging" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082079 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovn-acl-logging" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.082093 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="nbdb" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082100 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="nbdb" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.082111 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082117 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.082128 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082135 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.082143 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="kube-rbac-proxy-node" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082179 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="kube-rbac-proxy-node" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.082207 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="northd" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082213 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="northd" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.082222 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082229 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.082238 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082246 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.082258 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082267 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082326 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovn-acl-logging/0.log" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082379 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082396 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="nbdb" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082407 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovn-acl-logging" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082464 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082481 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082493 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082503 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="kube-rbac-proxy-node" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082514 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovn-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082524 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082535 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="sbdb" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082543 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="northd" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.082661 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082671 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.082790 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerName="ovnkube-controller" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.083311 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-qb725_3b40c6b3-0061-4224-82d5-3ccf67998722/ovn-controller/0.log" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084197 4632 generic.go:334] "Generic (PLEG): container finished" podID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerID="166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776" exitCode=0 Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084225 4632 generic.go:334] "Generic (PLEG): container finished" podID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerID="e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d" exitCode=0 Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084234 4632 generic.go:334] "Generic (PLEG): container finished" podID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerID="af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f" exitCode=0 Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084245 4632 generic.go:334] "Generic (PLEG): container finished" podID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerID="1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da" exitCode=0 Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084253 4632 generic.go:334] "Generic (PLEG): container finished" podID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerID="a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5" exitCode=0 Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084259 4632 generic.go:334] "Generic (PLEG): container finished" podID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerID="cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b" exitCode=0 Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084266 4632 generic.go:334] "Generic (PLEG): container finished" podID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerID="32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719" exitCode=143 Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084273 4632 generic.go:334] "Generic (PLEG): container finished" podID="3b40c6b3-0061-4224-82d5-3ccf67998722" containerID="7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705" exitCode=143 Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084412 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084702 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerDied","Data":"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084758 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerDied","Data":"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084772 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerDied","Data":"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084783 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerDied","Data":"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084797 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerDied","Data":"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084811 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerDied","Data":"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084824 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084837 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084845 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084853 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084860 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084868 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084874 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084880 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084885 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084894 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerDied","Data":"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084903 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084910 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084915 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084921 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084926 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084932 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084971 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084980 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084987 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.084993 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085002 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerDied","Data":"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085013 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085020 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085026 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085032 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085039 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085045 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085052 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085059 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085065 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085072 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085082 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qb725" event={"ID":"3b40c6b3-0061-4224-82d5-3ccf67998722","Type":"ContainerDied","Data":"5f99589c0e329dc2bea211f1582fe2ff509c48ed7460521bac851a5b63796f30"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085093 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085102 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085108 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085113 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085119 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085125 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085131 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085138 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085144 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085150 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085169 4632 scope.go:117] "RemoveContainer" containerID="166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.085414 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.089038 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.092476 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.092853 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.098901 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gqf22_4ec8e301-3037-4de0-94d2-32c49709660e/kube-multus/2.log" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.102517 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gqf22_4ec8e301-3037-4de0-94d2-32c49709660e/kube-multus/1.log" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.102588 4632 generic.go:334] "Generic (PLEG): container finished" podID="4ec8e301-3037-4de0-94d2-32c49709660e" containerID="5fd2699ddbdedbd54069c44af8e38bc058b347d99af772939ae6ec1d10220723" exitCode=2 Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.102644 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gqf22" event={"ID":"4ec8e301-3037-4de0-94d2-32c49709660e","Type":"ContainerDied","Data":"5fd2699ddbdedbd54069c44af8e38bc058b347d99af772939ae6ec1d10220723"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.102694 4632 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e48bcc5861bda7a15e45c892fa67ba73299d99e896f36f2cb68274a659ec5d34"} Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.103544 4632 scope.go:117] "RemoveContainer" containerID="5fd2699ddbdedbd54069c44af8e38bc058b347d99af772939ae6ec1d10220723" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.131102 4632 scope.go:117] "RemoveContainer" containerID="8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.150232 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b6e936db-ec1c-447a-894d-49bd7c74c315-ovnkube-script-lib\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.150643 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-slash\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.150667 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhfsk\" (UniqueName: \"kubernetes.io/projected/b6e936db-ec1c-447a-894d-49bd7c74c315-kube-api-access-jhfsk\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.150699 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b6e936db-ec1c-447a-894d-49bd7c74c315-ovnkube-config\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.150745 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b6e936db-ec1c-447a-894d-49bd7c74c315-env-overrides\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.150766 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-run-netns\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.150794 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-run-ovn\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.150820 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-run-ovn-kubernetes\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.150840 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-log-socket\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.150878 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-node-log\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.150899 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-cni-bin\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.150933 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-etc-openvswitch\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.150996 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b6e936db-ec1c-447a-894d-49bd7c74c315-ovn-node-metrics-cert\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151022 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-kubelet\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151057 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-var-lib-openvswitch\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151080 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-cni-netd\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151105 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-run-openvswitch\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151127 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151153 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-systemd-units\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151173 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-run-systemd\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151239 4632 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-node-log\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151257 4632 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/3b40c6b3-0061-4224-82d5-3ccf67998722-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151270 4632 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151280 4632 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151291 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dj6cl\" (UniqueName: \"kubernetes.io/projected/3b40c6b3-0061-4224-82d5-3ccf67998722-kube-api-access-dj6cl\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151301 4632 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151311 4632 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151322 4632 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151333 4632 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151344 4632 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/3b40c6b3-0061-4224-82d5-3ccf67998722-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.151353 4632 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/3b40c6b3-0061-4224-82d5-3ccf67998722-log-socket\") on node \"crc\" DevicePath \"\"" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.179207 4632 scope.go:117] "RemoveContainer" containerID="e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.197365 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qb725"] Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.208157 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qb725"] Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.223100 4632 scope.go:117] "RemoveContainer" containerID="af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252133 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b6e936db-ec1c-447a-894d-49bd7c74c315-ovnkube-script-lib\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252209 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-slash\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252238 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhfsk\" (UniqueName: \"kubernetes.io/projected/b6e936db-ec1c-447a-894d-49bd7c74c315-kube-api-access-jhfsk\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252286 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b6e936db-ec1c-447a-894d-49bd7c74c315-ovnkube-config\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252359 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-run-netns\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252384 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b6e936db-ec1c-447a-894d-49bd7c74c315-env-overrides\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252407 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-run-ovn\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252452 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-run-ovn-kubernetes\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252473 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-log-socket\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252490 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-node-log\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252521 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-cni-bin\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252536 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b6e936db-ec1c-447a-894d-49bd7c74c315-ovn-node-metrics-cert\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252550 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-etc-openvswitch\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252565 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-kubelet\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252597 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-var-lib-openvswitch\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252614 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-cni-netd\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252632 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-run-openvswitch\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252669 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252688 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-run-systemd\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252705 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-systemd-units\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.252798 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-systemd-units\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.253903 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/b6e936db-ec1c-447a-894d-49bd7c74c315-ovnkube-script-lib\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.253978 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-slash\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.254215 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-cni-bin\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.254976 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/b6e936db-ec1c-447a-894d-49bd7c74c315-ovnkube-config\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.255030 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-run-netns\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.255393 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/b6e936db-ec1c-447a-894d-49bd7c74c315-env-overrides\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.255445 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-run-ovn\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.255476 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-run-ovn-kubernetes\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.255511 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-log-socket\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.255545 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-node-log\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.255577 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-cni-netd\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.255606 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-etc-openvswitch\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.255638 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-kubelet\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.255668 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-var-lib-openvswitch\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.255699 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.255731 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-run-openvswitch\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.255760 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/b6e936db-ec1c-447a-894d-49bd7c74c315-run-systemd\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.257885 4632 scope.go:117] "RemoveContainer" containerID="1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.260462 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/b6e936db-ec1c-447a-894d-49bd7c74c315-ovn-node-metrics-cert\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.277483 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhfsk\" (UniqueName: \"kubernetes.io/projected/b6e936db-ec1c-447a-894d-49bd7c74c315-kube-api-access-jhfsk\") pod \"ovnkube-node-khrch\" (UID: \"b6e936db-ec1c-447a-894d-49bd7c74c315\") " pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.292135 4632 scope.go:117] "RemoveContainer" containerID="a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.335070 4632 scope.go:117] "RemoveContainer" containerID="cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.362342 4632 scope.go:117] "RemoveContainer" containerID="32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.383608 4632 scope.go:117] "RemoveContainer" containerID="7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.412379 4632 scope.go:117] "RemoveContainer" containerID="fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.432612 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.444103 4632 scope.go:117] "RemoveContainer" containerID="166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.445377 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776\": container with ID starting with 166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776 not found: ID does not exist" containerID="166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.445460 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776"} err="failed to get container status \"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776\": rpc error: code = NotFound desc = could not find container \"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776\": container with ID starting with 166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.445527 4632 scope.go:117] "RemoveContainer" containerID="8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.446852 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b\": container with ID starting with 8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b not found: ID does not exist" containerID="8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.446927 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b"} err="failed to get container status \"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b\": rpc error: code = NotFound desc = could not find container \"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b\": container with ID starting with 8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.446974 4632 scope.go:117] "RemoveContainer" containerID="e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.447551 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\": container with ID starting with e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d not found: ID does not exist" containerID="e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.447611 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d"} err="failed to get container status \"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\": rpc error: code = NotFound desc = could not find container \"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\": container with ID starting with e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.447645 4632 scope.go:117] "RemoveContainer" containerID="af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.448120 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\": container with ID starting with af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f not found: ID does not exist" containerID="af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.448173 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f"} err="failed to get container status \"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\": rpc error: code = NotFound desc = could not find container \"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\": container with ID starting with af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.449430 4632 scope.go:117] "RemoveContainer" containerID="1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.449836 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\": container with ID starting with 1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da not found: ID does not exist" containerID="1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.449869 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da"} err="failed to get container status \"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\": rpc error: code = NotFound desc = could not find container \"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\": container with ID starting with 1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.449891 4632 scope.go:117] "RemoveContainer" containerID="a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.450308 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\": container with ID starting with a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5 not found: ID does not exist" containerID="a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.450343 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5"} err="failed to get container status \"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\": rpc error: code = NotFound desc = could not find container \"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\": container with ID starting with a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.450367 4632 scope.go:117] "RemoveContainer" containerID="cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.450670 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\": container with ID starting with cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b not found: ID does not exist" containerID="cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.450718 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b"} err="failed to get container status \"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\": rpc error: code = NotFound desc = could not find container \"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\": container with ID starting with cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.450741 4632 scope.go:117] "RemoveContainer" containerID="32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.451276 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\": container with ID starting with 32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719 not found: ID does not exist" containerID="32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.451328 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719"} err="failed to get container status \"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\": rpc error: code = NotFound desc = could not find container \"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\": container with ID starting with 32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.451349 4632 scope.go:117] "RemoveContainer" containerID="7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.452280 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\": container with ID starting with 7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705 not found: ID does not exist" containerID="7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.452318 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705"} err="failed to get container status \"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\": rpc error: code = NotFound desc = could not find container \"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\": container with ID starting with 7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.452344 4632 scope.go:117] "RemoveContainer" containerID="fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50" Mar 13 10:17:50 crc kubenswrapper[4632]: E0313 10:17:50.458900 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\": container with ID starting with fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50 not found: ID does not exist" containerID="fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.459003 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50"} err="failed to get container status \"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\": rpc error: code = NotFound desc = could not find container \"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\": container with ID starting with fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.459044 4632 scope.go:117] "RemoveContainer" containerID="166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.460039 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776"} err="failed to get container status \"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776\": rpc error: code = NotFound desc = could not find container \"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776\": container with ID starting with 166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.460134 4632 scope.go:117] "RemoveContainer" containerID="8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.461141 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b"} err="failed to get container status \"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b\": rpc error: code = NotFound desc = could not find container \"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b\": container with ID starting with 8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.461166 4632 scope.go:117] "RemoveContainer" containerID="e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.461502 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d"} err="failed to get container status \"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\": rpc error: code = NotFound desc = could not find container \"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\": container with ID starting with e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.461547 4632 scope.go:117] "RemoveContainer" containerID="af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.462357 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f"} err="failed to get container status \"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\": rpc error: code = NotFound desc = could not find container \"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\": container with ID starting with af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.462383 4632 scope.go:117] "RemoveContainer" containerID="1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.462856 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da"} err="failed to get container status \"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\": rpc error: code = NotFound desc = could not find container \"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\": container with ID starting with 1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.462982 4632 scope.go:117] "RemoveContainer" containerID="a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.464153 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5"} err="failed to get container status \"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\": rpc error: code = NotFound desc = could not find container \"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\": container with ID starting with a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.464267 4632 scope.go:117] "RemoveContainer" containerID="cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.464706 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b"} err="failed to get container status \"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\": rpc error: code = NotFound desc = could not find container \"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\": container with ID starting with cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.464754 4632 scope.go:117] "RemoveContainer" containerID="32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.465276 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719"} err="failed to get container status \"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\": rpc error: code = NotFound desc = could not find container \"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\": container with ID starting with 32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.465353 4632 scope.go:117] "RemoveContainer" containerID="7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.466167 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705"} err="failed to get container status \"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\": rpc error: code = NotFound desc = could not find container \"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\": container with ID starting with 7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.466207 4632 scope.go:117] "RemoveContainer" containerID="fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.466807 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50"} err="failed to get container status \"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\": rpc error: code = NotFound desc = could not find container \"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\": container with ID starting with fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.466860 4632 scope.go:117] "RemoveContainer" containerID="166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.467458 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776"} err="failed to get container status \"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776\": rpc error: code = NotFound desc = could not find container \"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776\": container with ID starting with 166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.467490 4632 scope.go:117] "RemoveContainer" containerID="8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.469885 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b"} err="failed to get container status \"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b\": rpc error: code = NotFound desc = could not find container \"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b\": container with ID starting with 8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.469911 4632 scope.go:117] "RemoveContainer" containerID="e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.470287 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d"} err="failed to get container status \"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\": rpc error: code = NotFound desc = could not find container \"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\": container with ID starting with e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.470311 4632 scope.go:117] "RemoveContainer" containerID="af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.470637 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f"} err="failed to get container status \"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\": rpc error: code = NotFound desc = could not find container \"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\": container with ID starting with af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.470657 4632 scope.go:117] "RemoveContainer" containerID="1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.470880 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da"} err="failed to get container status \"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\": rpc error: code = NotFound desc = could not find container \"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\": container with ID starting with 1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.470908 4632 scope.go:117] "RemoveContainer" containerID="a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.471219 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5"} err="failed to get container status \"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\": rpc error: code = NotFound desc = could not find container \"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\": container with ID starting with a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.471246 4632 scope.go:117] "RemoveContainer" containerID="cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.471597 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b"} err="failed to get container status \"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\": rpc error: code = NotFound desc = could not find container \"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\": container with ID starting with cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.471628 4632 scope.go:117] "RemoveContainer" containerID="32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.472007 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719"} err="failed to get container status \"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\": rpc error: code = NotFound desc = could not find container \"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\": container with ID starting with 32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.472031 4632 scope.go:117] "RemoveContainer" containerID="7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.472648 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705"} err="failed to get container status \"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\": rpc error: code = NotFound desc = could not find container \"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\": container with ID starting with 7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.472689 4632 scope.go:117] "RemoveContainer" containerID="fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.473095 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50"} err="failed to get container status \"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\": rpc error: code = NotFound desc = could not find container \"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\": container with ID starting with fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.473116 4632 scope.go:117] "RemoveContainer" containerID="166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.473862 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776"} err="failed to get container status \"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776\": rpc error: code = NotFound desc = could not find container \"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776\": container with ID starting with 166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.473892 4632 scope.go:117] "RemoveContainer" containerID="8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.474514 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b"} err="failed to get container status \"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b\": rpc error: code = NotFound desc = could not find container \"8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b\": container with ID starting with 8f88d0230bc4958132c1c8a67c55a1e41f6e91534d69eb1a8c5061d17838098b not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.474594 4632 scope.go:117] "RemoveContainer" containerID="e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.475152 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d"} err="failed to get container status \"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\": rpc error: code = NotFound desc = could not find container \"e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d\": container with ID starting with e9f07f221ba02566b33ff83d55a32048ca4b9f412109f1d5b261e5e8576a9e2d not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.475219 4632 scope.go:117] "RemoveContainer" containerID="af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.476717 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f"} err="failed to get container status \"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\": rpc error: code = NotFound desc = could not find container \"af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f\": container with ID starting with af0c4e2f04409c8ddc0f9bc84c72dc88b6475220d060124a714ad4ad78e6101f not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.476747 4632 scope.go:117] "RemoveContainer" containerID="1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da" Mar 13 10:17:50 crc kubenswrapper[4632]: W0313 10:17:50.477304 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6e936db_ec1c_447a_894d_49bd7c74c315.slice/crio-aba0854f8bfee4eec9401b0ece87c9abbe0d79c52c67dba182c74b2eb059ffdd WatchSource:0}: Error finding container aba0854f8bfee4eec9401b0ece87c9abbe0d79c52c67dba182c74b2eb059ffdd: Status 404 returned error can't find the container with id aba0854f8bfee4eec9401b0ece87c9abbe0d79c52c67dba182c74b2eb059ffdd Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.477872 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da"} err="failed to get container status \"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\": rpc error: code = NotFound desc = could not find container \"1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da\": container with ID starting with 1ce7c67dfef6bd36183b1d9a902c15016d24ebf1dc92b5970391aeeaa321e8da not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.477983 4632 scope.go:117] "RemoveContainer" containerID="a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.479375 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5"} err="failed to get container status \"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\": rpc error: code = NotFound desc = could not find container \"a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5\": container with ID starting with a748b574435e8ce2de0ee0e311c4bc983ea5afe25f8a4873e16e13eebb8709b5 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.479429 4632 scope.go:117] "RemoveContainer" containerID="cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.479811 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b"} err="failed to get container status \"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\": rpc error: code = NotFound desc = could not find container \"cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b\": container with ID starting with cc7ffe60a2cb2de18612dc1db2a6002c06b442e1c0d689b556636671fd22c83b not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.479848 4632 scope.go:117] "RemoveContainer" containerID="32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.482507 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719"} err="failed to get container status \"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\": rpc error: code = NotFound desc = could not find container \"32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719\": container with ID starting with 32c437b2fe87b28335f5e5cc4f6c1921c9d7bc2f08ea8369e495cd6f4ab5b719 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.482548 4632 scope.go:117] "RemoveContainer" containerID="7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.483012 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705"} err="failed to get container status \"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\": rpc error: code = NotFound desc = could not find container \"7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705\": container with ID starting with 7de4dee0b604ec5005f3041256a1c8fba71b4dfb59d28a20f0577bfa987d5705 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.483038 4632 scope.go:117] "RemoveContainer" containerID="fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.485199 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50"} err="failed to get container status \"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\": rpc error: code = NotFound desc = could not find container \"fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50\": container with ID starting with fde8ebd48ee574e63496a8cd11e135524ab5dbabff0982f33ef6aa64eb350b50 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.485244 4632 scope.go:117] "RemoveContainer" containerID="166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.487253 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776"} err="failed to get container status \"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776\": rpc error: code = NotFound desc = could not find container \"166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776\": container with ID starting with 166459cac16ec3b496c247f07645c5ed749fa575d48ccabd0d32f2daa03d6776 not found: ID does not exist" Mar 13 10:17:50 crc kubenswrapper[4632]: I0313 10:17:50.887545 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" Mar 13 10:17:51 crc kubenswrapper[4632]: I0313 10:17:51.109344 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gqf22_4ec8e301-3037-4de0-94d2-32c49709660e/kube-multus/2.log" Mar 13 10:17:51 crc kubenswrapper[4632]: I0313 10:17:51.110028 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gqf22_4ec8e301-3037-4de0-94d2-32c49709660e/kube-multus/1.log" Mar 13 10:17:51 crc kubenswrapper[4632]: I0313 10:17:51.110178 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-gqf22" event={"ID":"4ec8e301-3037-4de0-94d2-32c49709660e","Type":"ContainerStarted","Data":"d8c124819539cd51aa3ecc51a7287d0ffc182af68ef9ff0ef3cca9bb296cb657"} Mar 13 10:17:51 crc kubenswrapper[4632]: I0313 10:17:51.113022 4632 generic.go:334] "Generic (PLEG): container finished" podID="b6e936db-ec1c-447a-894d-49bd7c74c315" containerID="7b1f6fd6cd656162231b57d1ed903f8ec0b1fd203839b9b2b1e97ee1cabd2305" exitCode=0 Mar 13 10:17:51 crc kubenswrapper[4632]: I0313 10:17:51.113086 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" event={"ID":"b6e936db-ec1c-447a-894d-49bd7c74c315","Type":"ContainerDied","Data":"7b1f6fd6cd656162231b57d1ed903f8ec0b1fd203839b9b2b1e97ee1cabd2305"} Mar 13 10:17:51 crc kubenswrapper[4632]: I0313 10:17:51.113145 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" event={"ID":"b6e936db-ec1c-447a-894d-49bd7c74c315","Type":"ContainerStarted","Data":"aba0854f8bfee4eec9401b0ece87c9abbe0d79c52c67dba182c74b2eb059ffdd"} Mar 13 10:17:52 crc kubenswrapper[4632]: I0313 10:17:52.053147 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b40c6b3-0061-4224-82d5-3ccf67998722" path="/var/lib/kubelet/pods/3b40c6b3-0061-4224-82d5-3ccf67998722/volumes" Mar 13 10:17:52 crc kubenswrapper[4632]: I0313 10:17:52.122024 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" event={"ID":"b6e936db-ec1c-447a-894d-49bd7c74c315","Type":"ContainerStarted","Data":"fd070db907a81c4129276f7963fb5d55f5374434eab025c5259df83494266b8d"} Mar 13 10:17:52 crc kubenswrapper[4632]: I0313 10:17:52.122092 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" event={"ID":"b6e936db-ec1c-447a-894d-49bd7c74c315","Type":"ContainerStarted","Data":"4c697aed6d5d0065ea2c0087f575a0767350e0241674f5e77e589fc4b744cf98"} Mar 13 10:17:52 crc kubenswrapper[4632]: I0313 10:17:52.122109 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" event={"ID":"b6e936db-ec1c-447a-894d-49bd7c74c315","Type":"ContainerStarted","Data":"89a52384fcc24d9b7edfc959adad0e5850f03abc686b20cf4fb1fc0ea6f9af72"} Mar 13 10:17:52 crc kubenswrapper[4632]: I0313 10:17:52.122123 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" event={"ID":"b6e936db-ec1c-447a-894d-49bd7c74c315","Type":"ContainerStarted","Data":"eb3431c358e931d65a798cec572556aa0c7f665d68e134af66a4f107f8282ba7"} Mar 13 10:17:52 crc kubenswrapper[4632]: I0313 10:17:52.122134 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" event={"ID":"b6e936db-ec1c-447a-894d-49bd7c74c315","Type":"ContainerStarted","Data":"79c0fd42d034938f1cc371005b0f7d56f7b78849587c385549ef8cebd5507bc2"} Mar 13 10:17:52 crc kubenswrapper[4632]: I0313 10:17:52.122144 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" event={"ID":"b6e936db-ec1c-447a-894d-49bd7c74c315","Type":"ContainerStarted","Data":"0ff1db46158b8bcf27a8f6bc8992c67300f851d3bae818c4446435bd5055d4cb"} Mar 13 10:17:54 crc kubenswrapper[4632]: I0313 10:17:54.136072 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" event={"ID":"b6e936db-ec1c-447a-894d-49bd7c74c315","Type":"ContainerStarted","Data":"aa7a10b4e72a51b406ec7be26b6bb8550c44700fdb3440d11638049a2c645726"} Mar 13 10:17:56 crc kubenswrapper[4632]: I0313 10:17:56.770160 4632 scope.go:117] "RemoveContainer" containerID="e48bcc5861bda7a15e45c892fa67ba73299d99e896f36f2cb68274a659ec5d34" Mar 13 10:17:57 crc kubenswrapper[4632]: I0313 10:17:57.156049 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" event={"ID":"b6e936db-ec1c-447a-894d-49bd7c74c315","Type":"ContainerStarted","Data":"c2db77d29f859566314a4b1b0340d322b3adfa73ec2e42760da934777cb7a840"} Mar 13 10:17:57 crc kubenswrapper[4632]: I0313 10:17:57.157418 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:57 crc kubenswrapper[4632]: I0313 10:17:57.157448 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:57 crc kubenswrapper[4632]: I0313 10:17:57.157542 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:57 crc kubenswrapper[4632]: I0313 10:17:57.162146 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-gqf22_4ec8e301-3037-4de0-94d2-32c49709660e/kube-multus/2.log" Mar 13 10:17:57 crc kubenswrapper[4632]: I0313 10:17:57.191477 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:17:57 crc kubenswrapper[4632]: I0313 10:17:57.196549 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" podStartSLOduration=7.196532192 podStartE2EDuration="7.196532192s" podCreationTimestamp="2026-03-13 10:17:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:17:57.192053188 +0000 UTC m=+851.214583321" watchObservedRunningTime="2026-03-13 10:17:57.196532192 +0000 UTC m=+851.219062325" Mar 13 10:17:57 crc kubenswrapper[4632]: I0313 10:17:57.214201 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:18:00 crc kubenswrapper[4632]: I0313 10:18:00.129736 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556618-ngbmk"] Mar 13 10:18:00 crc kubenswrapper[4632]: I0313 10:18:00.130911 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556618-ngbmk" Mar 13 10:18:00 crc kubenswrapper[4632]: I0313 10:18:00.133611 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:18:00 crc kubenswrapper[4632]: I0313 10:18:00.133704 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:18:00 crc kubenswrapper[4632]: I0313 10:18:00.133628 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:18:00 crc kubenswrapper[4632]: I0313 10:18:00.140786 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556618-ngbmk"] Mar 13 10:18:00 crc kubenswrapper[4632]: I0313 10:18:00.283086 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9cs7\" (UniqueName: \"kubernetes.io/projected/b93f1106-edf9-4cde-9acb-e265d8e07191-kube-api-access-v9cs7\") pod \"auto-csr-approver-29556618-ngbmk\" (UID: \"b93f1106-edf9-4cde-9acb-e265d8e07191\") " pod="openshift-infra/auto-csr-approver-29556618-ngbmk" Mar 13 10:18:00 crc kubenswrapper[4632]: I0313 10:18:00.384240 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9cs7\" (UniqueName: \"kubernetes.io/projected/b93f1106-edf9-4cde-9acb-e265d8e07191-kube-api-access-v9cs7\") pod \"auto-csr-approver-29556618-ngbmk\" (UID: \"b93f1106-edf9-4cde-9acb-e265d8e07191\") " pod="openshift-infra/auto-csr-approver-29556618-ngbmk" Mar 13 10:18:00 crc kubenswrapper[4632]: I0313 10:18:00.409664 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9cs7\" (UniqueName: \"kubernetes.io/projected/b93f1106-edf9-4cde-9acb-e265d8e07191-kube-api-access-v9cs7\") pod \"auto-csr-approver-29556618-ngbmk\" (UID: \"b93f1106-edf9-4cde-9acb-e265d8e07191\") " pod="openshift-infra/auto-csr-approver-29556618-ngbmk" Mar 13 10:18:00 crc kubenswrapper[4632]: I0313 10:18:00.447233 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556618-ngbmk" Mar 13 10:18:00 crc kubenswrapper[4632]: I0313 10:18:00.641267 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556618-ngbmk"] Mar 13 10:18:00 crc kubenswrapper[4632]: W0313 10:18:00.645126 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb93f1106_edf9_4cde_9acb_e265d8e07191.slice/crio-cfa2ab88d74bc1da1b618e81508d8824c824e518ede12962da0681c9ac4b54a8 WatchSource:0}: Error finding container cfa2ab88d74bc1da1b618e81508d8824c824e518ede12962da0681c9ac4b54a8: Status 404 returned error can't find the container with id cfa2ab88d74bc1da1b618e81508d8824c824e518ede12962da0681c9ac4b54a8 Mar 13 10:18:01 crc kubenswrapper[4632]: I0313 10:18:01.195584 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556618-ngbmk" event={"ID":"b93f1106-edf9-4cde-9acb-e265d8e07191","Type":"ContainerStarted","Data":"cfa2ab88d74bc1da1b618e81508d8824c824e518ede12962da0681c9ac4b54a8"} Mar 13 10:18:02 crc kubenswrapper[4632]: I0313 10:18:02.202289 4632 generic.go:334] "Generic (PLEG): container finished" podID="b93f1106-edf9-4cde-9acb-e265d8e07191" containerID="971cfa2ec11ce234b8c8c574daddb17b130773fddba410f62dd84c800e0f4023" exitCode=0 Mar 13 10:18:02 crc kubenswrapper[4632]: I0313 10:18:02.202399 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556618-ngbmk" event={"ID":"b93f1106-edf9-4cde-9acb-e265d8e07191","Type":"ContainerDied","Data":"971cfa2ec11ce234b8c8c574daddb17b130773fddba410f62dd84c800e0f4023"} Mar 13 10:18:03 crc kubenswrapper[4632]: I0313 10:18:03.439876 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556618-ngbmk" Mar 13 10:18:03 crc kubenswrapper[4632]: I0313 10:18:03.622321 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9cs7\" (UniqueName: \"kubernetes.io/projected/b93f1106-edf9-4cde-9acb-e265d8e07191-kube-api-access-v9cs7\") pod \"b93f1106-edf9-4cde-9acb-e265d8e07191\" (UID: \"b93f1106-edf9-4cde-9acb-e265d8e07191\") " Mar 13 10:18:03 crc kubenswrapper[4632]: I0313 10:18:03.630557 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b93f1106-edf9-4cde-9acb-e265d8e07191-kube-api-access-v9cs7" (OuterVolumeSpecName: "kube-api-access-v9cs7") pod "b93f1106-edf9-4cde-9acb-e265d8e07191" (UID: "b93f1106-edf9-4cde-9acb-e265d8e07191"). InnerVolumeSpecName "kube-api-access-v9cs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:18:03 crc kubenswrapper[4632]: I0313 10:18:03.723580 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9cs7\" (UniqueName: \"kubernetes.io/projected/b93f1106-edf9-4cde-9acb-e265d8e07191-kube-api-access-v9cs7\") on node \"crc\" DevicePath \"\"" Mar 13 10:18:04 crc kubenswrapper[4632]: I0313 10:18:04.216363 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556618-ngbmk" event={"ID":"b93f1106-edf9-4cde-9acb-e265d8e07191","Type":"ContainerDied","Data":"cfa2ab88d74bc1da1b618e81508d8824c824e518ede12962da0681c9ac4b54a8"} Mar 13 10:18:04 crc kubenswrapper[4632]: I0313 10:18:04.216430 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556618-ngbmk" Mar 13 10:18:04 crc kubenswrapper[4632]: I0313 10:18:04.216462 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfa2ab88d74bc1da1b618e81508d8824c824e518ede12962da0681c9ac4b54a8" Mar 13 10:18:04 crc kubenswrapper[4632]: I0313 10:18:04.485688 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556612-5t5ct"] Mar 13 10:18:04 crc kubenswrapper[4632]: I0313 10:18:04.489677 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556612-5t5ct"] Mar 13 10:18:06 crc kubenswrapper[4632]: I0313 10:18:06.050758 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="050ee655-a62f-4991-b493-d98493762823" path="/var/lib/kubelet/pods/050ee655-a62f-4991-b493-d98493762823/volumes" Mar 13 10:18:06 crc kubenswrapper[4632]: I0313 10:18:06.518499 4632 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 13 10:18:10 crc kubenswrapper[4632]: I0313 10:18:10.461191 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:18:10 crc kubenswrapper[4632]: I0313 10:18:10.462146 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:18:20 crc kubenswrapper[4632]: I0313 10:18:20.458462 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" Mar 13 10:18:32 crc kubenswrapper[4632]: I0313 10:18:32.978434 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg"] Mar 13 10:18:32 crc kubenswrapper[4632]: E0313 10:18:32.979230 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b93f1106-edf9-4cde-9acb-e265d8e07191" containerName="oc" Mar 13 10:18:32 crc kubenswrapper[4632]: I0313 10:18:32.979246 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="b93f1106-edf9-4cde-9acb-e265d8e07191" containerName="oc" Mar 13 10:18:32 crc kubenswrapper[4632]: I0313 10:18:32.979377 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="b93f1106-edf9-4cde-9acb-e265d8e07191" containerName="oc" Mar 13 10:18:32 crc kubenswrapper[4632]: I0313 10:18:32.980315 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" Mar 13 10:18:32 crc kubenswrapper[4632]: I0313 10:18:32.985308 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 10:18:32 crc kubenswrapper[4632]: I0313 10:18:32.986461 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg"] Mar 13 10:18:33 crc kubenswrapper[4632]: I0313 10:18:33.007199 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4x2pj\" (UniqueName: \"kubernetes.io/projected/2e270cfe-55fc-4855-87ff-4313a0ad319c-kube-api-access-4x2pj\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg\" (UID: \"2e270cfe-55fc-4855-87ff-4313a0ad319c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" Mar 13 10:18:33 crc kubenswrapper[4632]: I0313 10:18:33.007278 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2e270cfe-55fc-4855-87ff-4313a0ad319c-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg\" (UID: \"2e270cfe-55fc-4855-87ff-4313a0ad319c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" Mar 13 10:18:33 crc kubenswrapper[4632]: I0313 10:18:33.007306 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2e270cfe-55fc-4855-87ff-4313a0ad319c-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg\" (UID: \"2e270cfe-55fc-4855-87ff-4313a0ad319c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" Mar 13 10:18:33 crc kubenswrapper[4632]: I0313 10:18:33.108882 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4x2pj\" (UniqueName: \"kubernetes.io/projected/2e270cfe-55fc-4855-87ff-4313a0ad319c-kube-api-access-4x2pj\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg\" (UID: \"2e270cfe-55fc-4855-87ff-4313a0ad319c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" Mar 13 10:18:33 crc kubenswrapper[4632]: I0313 10:18:33.109034 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2e270cfe-55fc-4855-87ff-4313a0ad319c-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg\" (UID: \"2e270cfe-55fc-4855-87ff-4313a0ad319c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" Mar 13 10:18:33 crc kubenswrapper[4632]: I0313 10:18:33.109073 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2e270cfe-55fc-4855-87ff-4313a0ad319c-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg\" (UID: \"2e270cfe-55fc-4855-87ff-4313a0ad319c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" Mar 13 10:18:33 crc kubenswrapper[4632]: I0313 10:18:33.109511 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2e270cfe-55fc-4855-87ff-4313a0ad319c-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg\" (UID: \"2e270cfe-55fc-4855-87ff-4313a0ad319c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" Mar 13 10:18:33 crc kubenswrapper[4632]: I0313 10:18:33.109537 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2e270cfe-55fc-4855-87ff-4313a0ad319c-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg\" (UID: \"2e270cfe-55fc-4855-87ff-4313a0ad319c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" Mar 13 10:18:33 crc kubenswrapper[4632]: I0313 10:18:33.135588 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4x2pj\" (UniqueName: \"kubernetes.io/projected/2e270cfe-55fc-4855-87ff-4313a0ad319c-kube-api-access-4x2pj\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg\" (UID: \"2e270cfe-55fc-4855-87ff-4313a0ad319c\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" Mar 13 10:18:33 crc kubenswrapper[4632]: I0313 10:18:33.307505 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" Mar 13 10:18:33 crc kubenswrapper[4632]: I0313 10:18:33.496996 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg"] Mar 13 10:18:33 crc kubenswrapper[4632]: I0313 10:18:33.746551 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" event={"ID":"2e270cfe-55fc-4855-87ff-4313a0ad319c","Type":"ContainerStarted","Data":"70fccb3799fbfc67ff1cb5305ca58be21b5a5c6c0871d3079f2ed0fc9ec195dc"} Mar 13 10:18:33 crc kubenswrapper[4632]: I0313 10:18:33.746873 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" event={"ID":"2e270cfe-55fc-4855-87ff-4313a0ad319c","Type":"ContainerStarted","Data":"30f776f8c54e2a7f9a372e8f33dd1f58da40deb74ed5a948205ed020d75644c7"} Mar 13 10:18:34 crc kubenswrapper[4632]: I0313 10:18:34.752752 4632 generic.go:334] "Generic (PLEG): container finished" podID="2e270cfe-55fc-4855-87ff-4313a0ad319c" containerID="70fccb3799fbfc67ff1cb5305ca58be21b5a5c6c0871d3079f2ed0fc9ec195dc" exitCode=0 Mar 13 10:18:34 crc kubenswrapper[4632]: I0313 10:18:34.752800 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" event={"ID":"2e270cfe-55fc-4855-87ff-4313a0ad319c","Type":"ContainerDied","Data":"70fccb3799fbfc67ff1cb5305ca58be21b5a5c6c0871d3079f2ed0fc9ec195dc"} Mar 13 10:18:35 crc kubenswrapper[4632]: I0313 10:18:35.340074 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bww79"] Mar 13 10:18:35 crc kubenswrapper[4632]: I0313 10:18:35.341339 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:35 crc kubenswrapper[4632]: I0313 10:18:35.366972 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bww79"] Mar 13 10:18:35 crc kubenswrapper[4632]: I0313 10:18:35.539423 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adab5e58-1b8e-4170-b244-d45be51beccb-catalog-content\") pod \"redhat-operators-bww79\" (UID: \"adab5e58-1b8e-4170-b244-d45be51beccb\") " pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:35 crc kubenswrapper[4632]: I0313 10:18:35.539516 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adab5e58-1b8e-4170-b244-d45be51beccb-utilities\") pod \"redhat-operators-bww79\" (UID: \"adab5e58-1b8e-4170-b244-d45be51beccb\") " pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:35 crc kubenswrapper[4632]: I0313 10:18:35.540465 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqhvx\" (UniqueName: \"kubernetes.io/projected/adab5e58-1b8e-4170-b244-d45be51beccb-kube-api-access-rqhvx\") pod \"redhat-operators-bww79\" (UID: \"adab5e58-1b8e-4170-b244-d45be51beccb\") " pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:35 crc kubenswrapper[4632]: I0313 10:18:35.642078 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqhvx\" (UniqueName: \"kubernetes.io/projected/adab5e58-1b8e-4170-b244-d45be51beccb-kube-api-access-rqhvx\") pod \"redhat-operators-bww79\" (UID: \"adab5e58-1b8e-4170-b244-d45be51beccb\") " pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:35 crc kubenswrapper[4632]: I0313 10:18:35.642145 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adab5e58-1b8e-4170-b244-d45be51beccb-catalog-content\") pod \"redhat-operators-bww79\" (UID: \"adab5e58-1b8e-4170-b244-d45be51beccb\") " pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:35 crc kubenswrapper[4632]: I0313 10:18:35.642193 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adab5e58-1b8e-4170-b244-d45be51beccb-utilities\") pod \"redhat-operators-bww79\" (UID: \"adab5e58-1b8e-4170-b244-d45be51beccb\") " pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:35 crc kubenswrapper[4632]: I0313 10:18:35.642855 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adab5e58-1b8e-4170-b244-d45be51beccb-catalog-content\") pod \"redhat-operators-bww79\" (UID: \"adab5e58-1b8e-4170-b244-d45be51beccb\") " pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:35 crc kubenswrapper[4632]: I0313 10:18:35.643122 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adab5e58-1b8e-4170-b244-d45be51beccb-utilities\") pod \"redhat-operators-bww79\" (UID: \"adab5e58-1b8e-4170-b244-d45be51beccb\") " pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:35 crc kubenswrapper[4632]: I0313 10:18:35.663194 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqhvx\" (UniqueName: \"kubernetes.io/projected/adab5e58-1b8e-4170-b244-d45be51beccb-kube-api-access-rqhvx\") pod \"redhat-operators-bww79\" (UID: \"adab5e58-1b8e-4170-b244-d45be51beccb\") " pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:35 crc kubenswrapper[4632]: I0313 10:18:35.959688 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:36 crc kubenswrapper[4632]: I0313 10:18:36.218096 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bww79"] Mar 13 10:18:36 crc kubenswrapper[4632]: W0313 10:18:36.239298 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podadab5e58_1b8e_4170_b244_d45be51beccb.slice/crio-0479633a9d80d98716b8676d2d46358f35e2abf71d04127168560bd5df09cbf0 WatchSource:0}: Error finding container 0479633a9d80d98716b8676d2d46358f35e2abf71d04127168560bd5df09cbf0: Status 404 returned error can't find the container with id 0479633a9d80d98716b8676d2d46358f35e2abf71d04127168560bd5df09cbf0 Mar 13 10:18:36 crc kubenswrapper[4632]: I0313 10:18:36.767159 4632 generic.go:334] "Generic (PLEG): container finished" podID="adab5e58-1b8e-4170-b244-d45be51beccb" containerID="61d923bcef83757034506bbf7a8f076d3bc9f8e9e2edb45de9372bf6e235a420" exitCode=0 Mar 13 10:18:36 crc kubenswrapper[4632]: I0313 10:18:36.767230 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bww79" event={"ID":"adab5e58-1b8e-4170-b244-d45be51beccb","Type":"ContainerDied","Data":"61d923bcef83757034506bbf7a8f076d3bc9f8e9e2edb45de9372bf6e235a420"} Mar 13 10:18:36 crc kubenswrapper[4632]: I0313 10:18:36.767257 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bww79" event={"ID":"adab5e58-1b8e-4170-b244-d45be51beccb","Type":"ContainerStarted","Data":"0479633a9d80d98716b8676d2d46358f35e2abf71d04127168560bd5df09cbf0"} Mar 13 10:18:36 crc kubenswrapper[4632]: I0313 10:18:36.771037 4632 generic.go:334] "Generic (PLEG): container finished" podID="2e270cfe-55fc-4855-87ff-4313a0ad319c" containerID="74e3de3bdd958d58cfbec4868a514778aadc959af0ac02fef0f61e787540e630" exitCode=0 Mar 13 10:18:36 crc kubenswrapper[4632]: I0313 10:18:36.771076 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" event={"ID":"2e270cfe-55fc-4855-87ff-4313a0ad319c","Type":"ContainerDied","Data":"74e3de3bdd958d58cfbec4868a514778aadc959af0ac02fef0f61e787540e630"} Mar 13 10:18:37 crc kubenswrapper[4632]: I0313 10:18:37.777984 4632 generic.go:334] "Generic (PLEG): container finished" podID="2e270cfe-55fc-4855-87ff-4313a0ad319c" containerID="5659996057e94d41164ea9a63a709e951333a865b02ec58f8d7d3cf3acc64dbc" exitCode=0 Mar 13 10:18:37 crc kubenswrapper[4632]: I0313 10:18:37.779326 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" event={"ID":"2e270cfe-55fc-4855-87ff-4313a0ad319c","Type":"ContainerDied","Data":"5659996057e94d41164ea9a63a709e951333a865b02ec58f8d7d3cf3acc64dbc"} Mar 13 10:18:37 crc kubenswrapper[4632]: I0313 10:18:37.787492 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bww79" event={"ID":"adab5e58-1b8e-4170-b244-d45be51beccb","Type":"ContainerStarted","Data":"69a89f03ba249c5eb9ce15b60d6967b375feee74919b0e43506625501d0e271b"} Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.174121 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.288120 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4x2pj\" (UniqueName: \"kubernetes.io/projected/2e270cfe-55fc-4855-87ff-4313a0ad319c-kube-api-access-4x2pj\") pod \"2e270cfe-55fc-4855-87ff-4313a0ad319c\" (UID: \"2e270cfe-55fc-4855-87ff-4313a0ad319c\") " Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.288187 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2e270cfe-55fc-4855-87ff-4313a0ad319c-bundle\") pod \"2e270cfe-55fc-4855-87ff-4313a0ad319c\" (UID: \"2e270cfe-55fc-4855-87ff-4313a0ad319c\") " Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.288224 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2e270cfe-55fc-4855-87ff-4313a0ad319c-util\") pod \"2e270cfe-55fc-4855-87ff-4313a0ad319c\" (UID: \"2e270cfe-55fc-4855-87ff-4313a0ad319c\") " Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.288690 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e270cfe-55fc-4855-87ff-4313a0ad319c-bundle" (OuterVolumeSpecName: "bundle") pod "2e270cfe-55fc-4855-87ff-4313a0ad319c" (UID: "2e270cfe-55fc-4855-87ff-4313a0ad319c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.295798 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e270cfe-55fc-4855-87ff-4313a0ad319c-kube-api-access-4x2pj" (OuterVolumeSpecName: "kube-api-access-4x2pj") pod "2e270cfe-55fc-4855-87ff-4313a0ad319c" (UID: "2e270cfe-55fc-4855-87ff-4313a0ad319c"). InnerVolumeSpecName "kube-api-access-4x2pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.300548 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e270cfe-55fc-4855-87ff-4313a0ad319c-util" (OuterVolumeSpecName: "util") pod "2e270cfe-55fc-4855-87ff-4313a0ad319c" (UID: "2e270cfe-55fc-4855-87ff-4313a0ad319c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.389600 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4x2pj\" (UniqueName: \"kubernetes.io/projected/2e270cfe-55fc-4855-87ff-4313a0ad319c-kube-api-access-4x2pj\") on node \"crc\" DevicePath \"\"" Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.389639 4632 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/2e270cfe-55fc-4855-87ff-4313a0ad319c-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.389648 4632 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/2e270cfe-55fc-4855-87ff-4313a0ad319c-util\") on node \"crc\" DevicePath \"\"" Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.819825 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" event={"ID":"2e270cfe-55fc-4855-87ff-4313a0ad319c","Type":"ContainerDied","Data":"30f776f8c54e2a7f9a372e8f33dd1f58da40deb74ed5a948205ed020d75644c7"} Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.819846 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg" Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.819885 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30f776f8c54e2a7f9a372e8f33dd1f58da40deb74ed5a948205ed020d75644c7" Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.822106 4632 generic.go:334] "Generic (PLEG): container finished" podID="adab5e58-1b8e-4170-b244-d45be51beccb" containerID="69a89f03ba249c5eb9ce15b60d6967b375feee74919b0e43506625501d0e271b" exitCode=0 Mar 13 10:18:39 crc kubenswrapper[4632]: I0313 10:18:39.822175 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bww79" event={"ID":"adab5e58-1b8e-4170-b244-d45be51beccb","Type":"ContainerDied","Data":"69a89f03ba249c5eb9ce15b60d6967b375feee74919b0e43506625501d0e271b"} Mar 13 10:18:40 crc kubenswrapper[4632]: I0313 10:18:40.466204 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:18:40 crc kubenswrapper[4632]: I0313 10:18:40.466592 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:18:40 crc kubenswrapper[4632]: I0313 10:18:40.466649 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:18:40 crc kubenswrapper[4632]: I0313 10:18:40.467263 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7fcd863f1a2b3af4768aa1d32979163bc846d3d472acea1e8c27ffcf3dfe0ffc"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:18:40 crc kubenswrapper[4632]: I0313 10:18:40.467309 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://7fcd863f1a2b3af4768aa1d32979163bc846d3d472acea1e8c27ffcf3dfe0ffc" gracePeriod=600 Mar 13 10:18:40 crc kubenswrapper[4632]: I0313 10:18:40.832564 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="7fcd863f1a2b3af4768aa1d32979163bc846d3d472acea1e8c27ffcf3dfe0ffc" exitCode=0 Mar 13 10:18:40 crc kubenswrapper[4632]: I0313 10:18:40.832640 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"7fcd863f1a2b3af4768aa1d32979163bc846d3d472acea1e8c27ffcf3dfe0ffc"} Mar 13 10:18:40 crc kubenswrapper[4632]: I0313 10:18:40.832688 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"624a339b1e1f8b218223c2e3440b7f9925bb18567bb6def4fcf3bfc022198658"} Mar 13 10:18:40 crc kubenswrapper[4632]: I0313 10:18:40.832708 4632 scope.go:117] "RemoveContainer" containerID="313e3b067f9ea051953ab56cbddeb09cc8cceb68240f33ca492d13584077681c" Mar 13 10:18:40 crc kubenswrapper[4632]: I0313 10:18:40.837522 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bww79" event={"ID":"adab5e58-1b8e-4170-b244-d45be51beccb","Type":"ContainerStarted","Data":"7470f34a42aacbea955a79fad4dd4ed78868cab12e40fc4e1449bd061f3deb93"} Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.451249 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bww79" podStartSLOduration=4.938159852 podStartE2EDuration="8.451229694s" podCreationTimestamp="2026-03-13 10:18:35 +0000 UTC" firstStartedPulling="2026-03-13 10:18:36.769069777 +0000 UTC m=+890.791599910" lastFinishedPulling="2026-03-13 10:18:40.282139619 +0000 UTC m=+894.304669752" observedRunningTime="2026-03-13 10:18:40.878307507 +0000 UTC m=+894.900837660" watchObservedRunningTime="2026-03-13 10:18:43.451229694 +0000 UTC m=+897.473759827" Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.454793 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-bzmdv"] Mar 13 10:18:43 crc kubenswrapper[4632]: E0313 10:18:43.455045 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e270cfe-55fc-4855-87ff-4313a0ad319c" containerName="pull" Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.455066 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e270cfe-55fc-4855-87ff-4313a0ad319c" containerName="pull" Mar 13 10:18:43 crc kubenswrapper[4632]: E0313 10:18:43.455092 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e270cfe-55fc-4855-87ff-4313a0ad319c" containerName="util" Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.455103 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e270cfe-55fc-4855-87ff-4313a0ad319c" containerName="util" Mar 13 10:18:43 crc kubenswrapper[4632]: E0313 10:18:43.455114 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e270cfe-55fc-4855-87ff-4313a0ad319c" containerName="extract" Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.455120 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e270cfe-55fc-4855-87ff-4313a0ad319c" containerName="extract" Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.455209 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e270cfe-55fc-4855-87ff-4313a0ad319c" containerName="extract" Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.455586 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-bzmdv" Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.457856 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-lxf9r" Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.458546 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.460118 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.471704 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-bzmdv"] Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.475103 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9jgc\" (UniqueName: \"kubernetes.io/projected/3b679db2-06cc-4796-945a-5ced45b39053-kube-api-access-t9jgc\") pod \"nmstate-operator-796d4cfff4-bzmdv\" (UID: \"3b679db2-06cc-4796-945a-5ced45b39053\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-bzmdv" Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.576400 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9jgc\" (UniqueName: \"kubernetes.io/projected/3b679db2-06cc-4796-945a-5ced45b39053-kube-api-access-t9jgc\") pod \"nmstate-operator-796d4cfff4-bzmdv\" (UID: \"3b679db2-06cc-4796-945a-5ced45b39053\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-bzmdv" Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.615806 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9jgc\" (UniqueName: \"kubernetes.io/projected/3b679db2-06cc-4796-945a-5ced45b39053-kube-api-access-t9jgc\") pod \"nmstate-operator-796d4cfff4-bzmdv\" (UID: \"3b679db2-06cc-4796-945a-5ced45b39053\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-bzmdv" Mar 13 10:18:43 crc kubenswrapper[4632]: I0313 10:18:43.778969 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-bzmdv" Mar 13 10:18:44 crc kubenswrapper[4632]: I0313 10:18:44.306473 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-bzmdv"] Mar 13 10:18:44 crc kubenswrapper[4632]: W0313 10:18:44.312601 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b679db2_06cc_4796_945a_5ced45b39053.slice/crio-6497a9252cbcc6d30b8d952e7753224e2d9e98fa6b8ef76f6a046249c1701353 WatchSource:0}: Error finding container 6497a9252cbcc6d30b8d952e7753224e2d9e98fa6b8ef76f6a046249c1701353: Status 404 returned error can't find the container with id 6497a9252cbcc6d30b8d952e7753224e2d9e98fa6b8ef76f6a046249c1701353 Mar 13 10:18:44 crc kubenswrapper[4632]: I0313 10:18:44.875806 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-bzmdv" event={"ID":"3b679db2-06cc-4796-945a-5ced45b39053","Type":"ContainerStarted","Data":"6497a9252cbcc6d30b8d952e7753224e2d9e98fa6b8ef76f6a046249c1701353"} Mar 13 10:18:45 crc kubenswrapper[4632]: I0313 10:18:45.961046 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:45 crc kubenswrapper[4632]: I0313 10:18:45.961095 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:47 crc kubenswrapper[4632]: I0313 10:18:47.015321 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bww79" podUID="adab5e58-1b8e-4170-b244-d45be51beccb" containerName="registry-server" probeResult="failure" output=< Mar 13 10:18:47 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:18:47 crc kubenswrapper[4632]: > Mar 13 10:18:47 crc kubenswrapper[4632]: I0313 10:18:47.902319 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-bzmdv" event={"ID":"3b679db2-06cc-4796-945a-5ced45b39053","Type":"ContainerStarted","Data":"d94fc64c10cc29b0260c1ee26784f074d44dac3354acfc38028b8bde57e66c66"} Mar 13 10:18:47 crc kubenswrapper[4632]: I0313 10:18:47.921503 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-bzmdv" podStartSLOduration=2.182431379 podStartE2EDuration="4.921483144s" podCreationTimestamp="2026-03-13 10:18:43 +0000 UTC" firstStartedPulling="2026-03-13 10:18:44.31501176 +0000 UTC m=+898.337541893" lastFinishedPulling="2026-03-13 10:18:47.054063525 +0000 UTC m=+901.076593658" observedRunningTime="2026-03-13 10:18:47.917685327 +0000 UTC m=+901.940215480" watchObservedRunningTime="2026-03-13 10:18:47.921483144 +0000 UTC m=+901.944013277" Mar 13 10:18:52 crc kubenswrapper[4632]: I0313 10:18:52.985835 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-lnfrw"] Mar 13 10:18:52 crc kubenswrapper[4632]: I0313 10:18:52.987507 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lnfrw" Mar 13 10:18:52 crc kubenswrapper[4632]: I0313 10:18:52.990129 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-5fccg" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.004339 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-gcngd"] Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.005239 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.013818 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.040821 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-mpfnk"] Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.041487 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.077828 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-lnfrw"] Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.107512 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-gcngd"] Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.184038 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn"] Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.185481 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.190631 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.190932 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-st6tm" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.191623 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/33445a2b-7fa8-4198-a60a-09caeb69b8ed-dbus-socket\") pod \"nmstate-handler-mpfnk\" (UID: \"33445a2b-7fa8-4198-a60a-09caeb69b8ed\") " pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.191656 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xctsv\" (UniqueName: \"kubernetes.io/projected/33445a2b-7fa8-4198-a60a-09caeb69b8ed-kube-api-access-xctsv\") pod \"nmstate-handler-mpfnk\" (UID: \"33445a2b-7fa8-4198-a60a-09caeb69b8ed\") " pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.191680 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9bf11778-d854-4c97-acd1-ed4822ee5f47-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-gcngd\" (UID: \"9bf11778-d854-4c97-acd1-ed4822ee5f47\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.191700 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdtv6\" (UniqueName: \"kubernetes.io/projected/0c63c4bc-5c1a-4af0-b255-eb418d8a02cd-kube-api-access-fdtv6\") pod \"nmstate-metrics-9b8c8685d-lnfrw\" (UID: \"0c63c4bc-5c1a-4af0-b255-eb418d8a02cd\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lnfrw" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.191718 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/33445a2b-7fa8-4198-a60a-09caeb69b8ed-nmstate-lock\") pod \"nmstate-handler-mpfnk\" (UID: \"33445a2b-7fa8-4198-a60a-09caeb69b8ed\") " pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.191741 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgc9p\" (UniqueName: \"kubernetes.io/projected/9bf11778-d854-4c97-acd1-ed4822ee5f47-kube-api-access-zgc9p\") pod \"nmstate-webhook-5f558f5558-gcngd\" (UID: \"9bf11778-d854-4c97-acd1-ed4822ee5f47\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.191770 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/33445a2b-7fa8-4198-a60a-09caeb69b8ed-ovs-socket\") pod \"nmstate-handler-mpfnk\" (UID: \"33445a2b-7fa8-4198-a60a-09caeb69b8ed\") " pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.197429 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn"] Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.199117 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.292522 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/33445a2b-7fa8-4198-a60a-09caeb69b8ed-nmstate-lock\") pod \"nmstate-handler-mpfnk\" (UID: \"33445a2b-7fa8-4198-a60a-09caeb69b8ed\") " pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.292982 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1ca5cae6-5549-492a-a257-745bb41d3574-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-kzrvn\" (UID: \"1ca5cae6-5549-492a-a257-745bb41d3574\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.292660 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/33445a2b-7fa8-4198-a60a-09caeb69b8ed-nmstate-lock\") pod \"nmstate-handler-mpfnk\" (UID: \"33445a2b-7fa8-4198-a60a-09caeb69b8ed\") " pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.293203 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgc9p\" (UniqueName: \"kubernetes.io/projected/9bf11778-d854-4c97-acd1-ed4822ee5f47-kube-api-access-zgc9p\") pod \"nmstate-webhook-5f558f5558-gcngd\" (UID: \"9bf11778-d854-4c97-acd1-ed4822ee5f47\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.293319 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/33445a2b-7fa8-4198-a60a-09caeb69b8ed-ovs-socket\") pod \"nmstate-handler-mpfnk\" (UID: \"33445a2b-7fa8-4198-a60a-09caeb69b8ed\") " pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.293395 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/33445a2b-7fa8-4198-a60a-09caeb69b8ed-ovs-socket\") pod \"nmstate-handler-mpfnk\" (UID: \"33445a2b-7fa8-4198-a60a-09caeb69b8ed\") " pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.293409 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca5cae6-5549-492a-a257-745bb41d3574-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-kzrvn\" (UID: \"1ca5cae6-5549-492a-a257-745bb41d3574\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.293713 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54nqs\" (UniqueName: \"kubernetes.io/projected/1ca5cae6-5549-492a-a257-745bb41d3574-kube-api-access-54nqs\") pod \"nmstate-console-plugin-86f58fcf4-kzrvn\" (UID: \"1ca5cae6-5549-492a-a257-745bb41d3574\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.293804 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/33445a2b-7fa8-4198-a60a-09caeb69b8ed-dbus-socket\") pod \"nmstate-handler-mpfnk\" (UID: \"33445a2b-7fa8-4198-a60a-09caeb69b8ed\") " pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.293831 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xctsv\" (UniqueName: \"kubernetes.io/projected/33445a2b-7fa8-4198-a60a-09caeb69b8ed-kube-api-access-xctsv\") pod \"nmstate-handler-mpfnk\" (UID: \"33445a2b-7fa8-4198-a60a-09caeb69b8ed\") " pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.293870 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9bf11778-d854-4c97-acd1-ed4822ee5f47-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-gcngd\" (UID: \"9bf11778-d854-4c97-acd1-ed4822ee5f47\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.293896 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdtv6\" (UniqueName: \"kubernetes.io/projected/0c63c4bc-5c1a-4af0-b255-eb418d8a02cd-kube-api-access-fdtv6\") pod \"nmstate-metrics-9b8c8685d-lnfrw\" (UID: \"0c63c4bc-5c1a-4af0-b255-eb418d8a02cd\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lnfrw" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.294709 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/33445a2b-7fa8-4198-a60a-09caeb69b8ed-dbus-socket\") pod \"nmstate-handler-mpfnk\" (UID: \"33445a2b-7fa8-4198-a60a-09caeb69b8ed\") " pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.310088 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9bf11778-d854-4c97-acd1-ed4822ee5f47-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-gcngd\" (UID: \"9bf11778-d854-4c97-acd1-ed4822ee5f47\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.324775 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgc9p\" (UniqueName: \"kubernetes.io/projected/9bf11778-d854-4c97-acd1-ed4822ee5f47-kube-api-access-zgc9p\") pod \"nmstate-webhook-5f558f5558-gcngd\" (UID: \"9bf11778-d854-4c97-acd1-ed4822ee5f47\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.328728 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xctsv\" (UniqueName: \"kubernetes.io/projected/33445a2b-7fa8-4198-a60a-09caeb69b8ed-kube-api-access-xctsv\") pod \"nmstate-handler-mpfnk\" (UID: \"33445a2b-7fa8-4198-a60a-09caeb69b8ed\") " pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.336484 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdtv6\" (UniqueName: \"kubernetes.io/projected/0c63c4bc-5c1a-4af0-b255-eb418d8a02cd-kube-api-access-fdtv6\") pod \"nmstate-metrics-9b8c8685d-lnfrw\" (UID: \"0c63c4bc-5c1a-4af0-b255-eb418d8a02cd\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lnfrw" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.354954 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.395647 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1ca5cae6-5549-492a-a257-745bb41d3574-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-kzrvn\" (UID: \"1ca5cae6-5549-492a-a257-745bb41d3574\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.395758 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca5cae6-5549-492a-a257-745bb41d3574-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-kzrvn\" (UID: \"1ca5cae6-5549-492a-a257-745bb41d3574\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.395811 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54nqs\" (UniqueName: \"kubernetes.io/projected/1ca5cae6-5549-492a-a257-745bb41d3574-kube-api-access-54nqs\") pod \"nmstate-console-plugin-86f58fcf4-kzrvn\" (UID: \"1ca5cae6-5549-492a-a257-745bb41d3574\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" Mar 13 10:18:53 crc kubenswrapper[4632]: E0313 10:18:53.396839 4632 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Mar 13 10:18:53 crc kubenswrapper[4632]: E0313 10:18:53.397004 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1ca5cae6-5549-492a-a257-745bb41d3574-plugin-serving-cert podName:1ca5cae6-5549-492a-a257-745bb41d3574 nodeName:}" failed. No retries permitted until 2026-03-13 10:18:53.896980878 +0000 UTC m=+907.919511001 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/1ca5cae6-5549-492a-a257-745bb41d3574-plugin-serving-cert") pod "nmstate-console-plugin-86f58fcf4-kzrvn" (UID: "1ca5cae6-5549-492a-a257-745bb41d3574") : secret "plugin-serving-cert" not found Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.397659 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/1ca5cae6-5549-492a-a257-745bb41d3574-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-kzrvn\" (UID: \"1ca5cae6-5549-492a-a257-745bb41d3574\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.416011 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54nqs\" (UniqueName: \"kubernetes.io/projected/1ca5cae6-5549-492a-a257-745bb41d3574-kube-api-access-54nqs\") pod \"nmstate-console-plugin-86f58fcf4-kzrvn\" (UID: \"1ca5cae6-5549-492a-a257-745bb41d3574\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.434617 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5678554f8b-n7dcv"] Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.435295 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.499652 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5678554f8b-n7dcv"] Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.598313 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xb8j\" (UniqueName: \"kubernetes.io/projected/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-kube-api-access-4xb8j\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.598382 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-oauth-serving-cert\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.598457 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-console-config\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.598500 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-service-ca\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.598602 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-console-serving-cert\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.598640 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-trusted-ca-bundle\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.598700 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-console-oauth-config\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.606149 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lnfrw" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.705005 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.705449 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-console-serving-cert\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.705510 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-trusted-ca-bundle\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.705546 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-console-oauth-config\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.705607 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xb8j\" (UniqueName: \"kubernetes.io/projected/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-kube-api-access-4xb8j\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.705630 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-oauth-serving-cert\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.705672 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-console-config\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.705700 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-service-ca\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.706707 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-oauth-serving-cert\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.706879 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-console-config\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.706961 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-service-ca\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.708142 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-trusted-ca-bundle\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.709281 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-console-serving-cert\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.710904 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-console-oauth-config\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.732864 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xb8j\" (UniqueName: \"kubernetes.io/projected/a59bb7d3-da4a-4275-9dcb-b851215a9cd0-kube-api-access-4xb8j\") pod \"console-5678554f8b-n7dcv\" (UID: \"a59bb7d3-da4a-4275-9dcb-b851215a9cd0\") " pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.757791 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.908689 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca5cae6-5549-492a-a257-745bb41d3574-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-kzrvn\" (UID: \"1ca5cae6-5549-492a-a257-745bb41d3574\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.916576 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/1ca5cae6-5549-492a-a257-745bb41d3574-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-kzrvn\" (UID: \"1ca5cae6-5549-492a-a257-745bb41d3574\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" Mar 13 10:18:53 crc kubenswrapper[4632]: I0313 10:18:53.937731 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-mpfnk" event={"ID":"33445a2b-7fa8-4198-a60a-09caeb69b8ed","Type":"ContainerStarted","Data":"dd8932567ba0ab49f5ef6914d96e172cff5fd167d4c6e19ef08608b2d79885af"} Mar 13 10:18:54 crc kubenswrapper[4632]: I0313 10:18:54.117187 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" Mar 13 10:18:54 crc kubenswrapper[4632]: I0313 10:18:54.388268 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-lnfrw"] Mar 13 10:18:54 crc kubenswrapper[4632]: W0313 10:18:54.397295 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0c63c4bc_5c1a_4af0_b255_eb418d8a02cd.slice/crio-b562142a6db92fede8f830604066a7db2dbced3cab16efb3a78d56359c96acd3 WatchSource:0}: Error finding container b562142a6db92fede8f830604066a7db2dbced3cab16efb3a78d56359c96acd3: Status 404 returned error can't find the container with id b562142a6db92fede8f830604066a7db2dbced3cab16efb3a78d56359c96acd3 Mar 13 10:18:54 crc kubenswrapper[4632]: I0313 10:18:54.407016 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-gcngd"] Mar 13 10:18:54 crc kubenswrapper[4632]: W0313 10:18:54.409372 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9bf11778_d854_4c97_acd1_ed4822ee5f47.slice/crio-475f5e9b33d3a7fbbe9894ef41760606869253d3857fd3c6b9c8facde90b207d WatchSource:0}: Error finding container 475f5e9b33d3a7fbbe9894ef41760606869253d3857fd3c6b9c8facde90b207d: Status 404 returned error can't find the container with id 475f5e9b33d3a7fbbe9894ef41760606869253d3857fd3c6b9c8facde90b207d Mar 13 10:18:54 crc kubenswrapper[4632]: I0313 10:18:54.415741 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5678554f8b-n7dcv"] Mar 13 10:18:54 crc kubenswrapper[4632]: I0313 10:18:54.570563 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn"] Mar 13 10:18:54 crc kubenswrapper[4632]: W0313 10:18:54.577873 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1ca5cae6_5549_492a_a257_745bb41d3574.slice/crio-3fc6b0a0a8d409585f69921a0edff900d5c7b4f432cc49fc9f9da240899b4999 WatchSource:0}: Error finding container 3fc6b0a0a8d409585f69921a0edff900d5c7b4f432cc49fc9f9da240899b4999: Status 404 returned error can't find the container with id 3fc6b0a0a8d409585f69921a0edff900d5c7b4f432cc49fc9f9da240899b4999 Mar 13 10:18:54 crc kubenswrapper[4632]: I0313 10:18:54.946670 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5678554f8b-n7dcv" event={"ID":"a59bb7d3-da4a-4275-9dcb-b851215a9cd0","Type":"ContainerStarted","Data":"ea0d7b906ac5950b0053bf184a8dbed0c1554fc87711206c4674d1bbc3408c3d"} Mar 13 10:18:54 crc kubenswrapper[4632]: I0313 10:18:54.947352 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5678554f8b-n7dcv" event={"ID":"a59bb7d3-da4a-4275-9dcb-b851215a9cd0","Type":"ContainerStarted","Data":"cb23bd32230b0013c61ac1a9dcc579e8f217e4053f1cf49608cf48c3bca30bb7"} Mar 13 10:18:54 crc kubenswrapper[4632]: I0313 10:18:54.951067 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" event={"ID":"1ca5cae6-5549-492a-a257-745bb41d3574","Type":"ContainerStarted","Data":"3fc6b0a0a8d409585f69921a0edff900d5c7b4f432cc49fc9f9da240899b4999"} Mar 13 10:18:54 crc kubenswrapper[4632]: I0313 10:18:54.952978 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" event={"ID":"9bf11778-d854-4c97-acd1-ed4822ee5f47","Type":"ContainerStarted","Data":"475f5e9b33d3a7fbbe9894ef41760606869253d3857fd3c6b9c8facde90b207d"} Mar 13 10:18:54 crc kubenswrapper[4632]: I0313 10:18:54.955121 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lnfrw" event={"ID":"0c63c4bc-5c1a-4af0-b255-eb418d8a02cd","Type":"ContainerStarted","Data":"b562142a6db92fede8f830604066a7db2dbced3cab16efb3a78d56359c96acd3"} Mar 13 10:18:54 crc kubenswrapper[4632]: I0313 10:18:54.970413 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5678554f8b-n7dcv" podStartSLOduration=1.9703955419999999 podStartE2EDuration="1.970395542s" podCreationTimestamp="2026-03-13 10:18:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:18:54.969461708 +0000 UTC m=+908.991991841" watchObservedRunningTime="2026-03-13 10:18:54.970395542 +0000 UTC m=+908.992925675" Mar 13 10:18:56 crc kubenswrapper[4632]: I0313 10:18:56.137573 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:56 crc kubenswrapper[4632]: I0313 10:18:56.213876 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:56 crc kubenswrapper[4632]: I0313 10:18:56.373220 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bww79"] Mar 13 10:18:56 crc kubenswrapper[4632]: I0313 10:18:56.958801 4632 scope.go:117] "RemoveContainer" containerID="3025e6a57984dbcc7f1272476cb4a6a1339dea799f52af43239e5a72f7479138" Mar 13 10:18:57 crc kubenswrapper[4632]: I0313 10:18:57.976868 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bww79" podUID="adab5e58-1b8e-4170-b244-d45be51beccb" containerName="registry-server" containerID="cri-o://7470f34a42aacbea955a79fad4dd4ed78868cab12e40fc4e1449bd061f3deb93" gracePeriod=2 Mar 13 10:18:58 crc kubenswrapper[4632]: I0313 10:18:58.986913 4632 generic.go:334] "Generic (PLEG): container finished" podID="adab5e58-1b8e-4170-b244-d45be51beccb" containerID="7470f34a42aacbea955a79fad4dd4ed78868cab12e40fc4e1449bd061f3deb93" exitCode=0 Mar 13 10:18:58 crc kubenswrapper[4632]: I0313 10:18:58.987324 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bww79" event={"ID":"adab5e58-1b8e-4170-b244-d45be51beccb","Type":"ContainerDied","Data":"7470f34a42aacbea955a79fad4dd4ed78868cab12e40fc4e1449bd061f3deb93"} Mar 13 10:18:58 crc kubenswrapper[4632]: I0313 10:18:58.991026 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" event={"ID":"9bf11778-d854-4c97-acd1-ed4822ee5f47","Type":"ContainerStarted","Data":"5716a0e9287e6d22b7af01f41bdef1fc843fe9af997de193cb88b8f7571d5088"} Mar 13 10:18:58 crc kubenswrapper[4632]: I0313 10:18:58.991774 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" Mar 13 10:18:58 crc kubenswrapper[4632]: I0313 10:18:58.994878 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lnfrw" event={"ID":"0c63c4bc-5c1a-4af0-b255-eb418d8a02cd","Type":"ContainerStarted","Data":"7c1f2f888382cc87649c5b407623d5a89f5e995564511ff3d0581b77b107eaaf"} Mar 13 10:18:58 crc kubenswrapper[4632]: I0313 10:18:58.996264 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-mpfnk" event={"ID":"33445a2b-7fa8-4198-a60a-09caeb69b8ed","Type":"ContainerStarted","Data":"01dbb2c512e7a384c4a413468acce6565de99a22432dbec235a8843f26e6f1a4"} Mar 13 10:18:58 crc kubenswrapper[4632]: I0313 10:18:58.996398 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:18:59 crc kubenswrapper[4632]: I0313 10:18:59.021306 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" podStartSLOduration=3.201810704 podStartE2EDuration="7.021287514s" podCreationTimestamp="2026-03-13 10:18:52 +0000 UTC" firstStartedPulling="2026-03-13 10:18:54.411408967 +0000 UTC m=+908.433939100" lastFinishedPulling="2026-03-13 10:18:58.230885767 +0000 UTC m=+912.253415910" observedRunningTime="2026-03-13 10:18:59.009847599 +0000 UTC m=+913.032377752" watchObservedRunningTime="2026-03-13 10:18:59.021287514 +0000 UTC m=+913.043817647" Mar 13 10:18:59 crc kubenswrapper[4632]: I0313 10:18:59.209381 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:18:59 crc kubenswrapper[4632]: I0313 10:18:59.232933 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-mpfnk" podStartSLOduration=2.3914720210000002 podStartE2EDuration="7.23291618s" podCreationTimestamp="2026-03-13 10:18:52 +0000 UTC" firstStartedPulling="2026-03-13 10:18:53.393358116 +0000 UTC m=+907.415888249" lastFinishedPulling="2026-03-13 10:18:58.234802275 +0000 UTC m=+912.257332408" observedRunningTime="2026-03-13 10:18:59.030822891 +0000 UTC m=+913.053353044" watchObservedRunningTime="2026-03-13 10:18:59.23291618 +0000 UTC m=+913.255446313" Mar 13 10:18:59 crc kubenswrapper[4632]: I0313 10:18:59.234847 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adab5e58-1b8e-4170-b244-d45be51beccb-catalog-content\") pod \"adab5e58-1b8e-4170-b244-d45be51beccb\" (UID: \"adab5e58-1b8e-4170-b244-d45be51beccb\") " Mar 13 10:18:59 crc kubenswrapper[4632]: I0313 10:18:59.234992 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqhvx\" (UniqueName: \"kubernetes.io/projected/adab5e58-1b8e-4170-b244-d45be51beccb-kube-api-access-rqhvx\") pod \"adab5e58-1b8e-4170-b244-d45be51beccb\" (UID: \"adab5e58-1b8e-4170-b244-d45be51beccb\") " Mar 13 10:18:59 crc kubenswrapper[4632]: I0313 10:18:59.235043 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adab5e58-1b8e-4170-b244-d45be51beccb-utilities\") pod \"adab5e58-1b8e-4170-b244-d45be51beccb\" (UID: \"adab5e58-1b8e-4170-b244-d45be51beccb\") " Mar 13 10:18:59 crc kubenswrapper[4632]: I0313 10:18:59.236328 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adab5e58-1b8e-4170-b244-d45be51beccb-utilities" (OuterVolumeSpecName: "utilities") pod "adab5e58-1b8e-4170-b244-d45be51beccb" (UID: "adab5e58-1b8e-4170-b244-d45be51beccb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:18:59 crc kubenswrapper[4632]: I0313 10:18:59.260116 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adab5e58-1b8e-4170-b244-d45be51beccb-kube-api-access-rqhvx" (OuterVolumeSpecName: "kube-api-access-rqhvx") pod "adab5e58-1b8e-4170-b244-d45be51beccb" (UID: "adab5e58-1b8e-4170-b244-d45be51beccb"). InnerVolumeSpecName "kube-api-access-rqhvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:18:59 crc kubenswrapper[4632]: I0313 10:18:59.336514 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqhvx\" (UniqueName: \"kubernetes.io/projected/adab5e58-1b8e-4170-b244-d45be51beccb-kube-api-access-rqhvx\") on node \"crc\" DevicePath \"\"" Mar 13 10:18:59 crc kubenswrapper[4632]: I0313 10:18:59.336571 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/adab5e58-1b8e-4170-b244-d45be51beccb-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:18:59 crc kubenswrapper[4632]: I0313 10:18:59.431816 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adab5e58-1b8e-4170-b244-d45be51beccb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "adab5e58-1b8e-4170-b244-d45be51beccb" (UID: "adab5e58-1b8e-4170-b244-d45be51beccb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:18:59 crc kubenswrapper[4632]: I0313 10:18:59.438090 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/adab5e58-1b8e-4170-b244-d45be51beccb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:19:00 crc kubenswrapper[4632]: I0313 10:19:00.010057 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" event={"ID":"1ca5cae6-5549-492a-a257-745bb41d3574","Type":"ContainerStarted","Data":"c03c482f8a1d9b16c2959e5bf7664ff287487bc5fffc3e179aa213dca54a42c4"} Mar 13 10:19:00 crc kubenswrapper[4632]: I0313 10:19:00.012315 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bww79" event={"ID":"adab5e58-1b8e-4170-b244-d45be51beccb","Type":"ContainerDied","Data":"0479633a9d80d98716b8676d2d46358f35e2abf71d04127168560bd5df09cbf0"} Mar 13 10:19:00 crc kubenswrapper[4632]: I0313 10:19:00.012378 4632 scope.go:117] "RemoveContainer" containerID="7470f34a42aacbea955a79fad4dd4ed78868cab12e40fc4e1449bd061f3deb93" Mar 13 10:19:00 crc kubenswrapper[4632]: I0313 10:19:00.012459 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bww79" Mar 13 10:19:00 crc kubenswrapper[4632]: I0313 10:19:00.034429 4632 scope.go:117] "RemoveContainer" containerID="69a89f03ba249c5eb9ce15b60d6967b375feee74919b0e43506625501d0e271b" Mar 13 10:19:00 crc kubenswrapper[4632]: I0313 10:19:00.037672 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-kzrvn" podStartSLOduration=2.293789154 podStartE2EDuration="7.037652475s" podCreationTimestamp="2026-03-13 10:18:53 +0000 UTC" firstStartedPulling="2026-03-13 10:18:54.580146401 +0000 UTC m=+908.602676534" lastFinishedPulling="2026-03-13 10:18:59.324009722 +0000 UTC m=+913.346539855" observedRunningTime="2026-03-13 10:19:00.03420482 +0000 UTC m=+914.056734973" watchObservedRunningTime="2026-03-13 10:19:00.037652475 +0000 UTC m=+914.060182628" Mar 13 10:19:00 crc kubenswrapper[4632]: I0313 10:19:00.063511 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bww79"] Mar 13 10:19:00 crc kubenswrapper[4632]: I0313 10:19:00.067236 4632 scope.go:117] "RemoveContainer" containerID="61d923bcef83757034506bbf7a8f076d3bc9f8e9e2edb45de9372bf6e235a420" Mar 13 10:19:00 crc kubenswrapper[4632]: I0313 10:19:00.070375 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bww79"] Mar 13 10:19:02 crc kubenswrapper[4632]: I0313 10:19:02.025784 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lnfrw" event={"ID":"0c63c4bc-5c1a-4af0-b255-eb418d8a02cd","Type":"ContainerStarted","Data":"4c7d634e9b1bceea1fe3fad2c30e40060141b191fcf67108f5820b1526fa18ac"} Mar 13 10:19:02 crc kubenswrapper[4632]: I0313 10:19:02.042954 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-lnfrw" podStartSLOduration=3.091006074 podStartE2EDuration="10.042923383s" podCreationTimestamp="2026-03-13 10:18:52 +0000 UTC" firstStartedPulling="2026-03-13 10:18:54.399216806 +0000 UTC m=+908.421746939" lastFinishedPulling="2026-03-13 10:19:01.351134115 +0000 UTC m=+915.373664248" observedRunningTime="2026-03-13 10:19:02.040179455 +0000 UTC m=+916.062709588" watchObservedRunningTime="2026-03-13 10:19:02.042923383 +0000 UTC m=+916.065453516" Mar 13 10:19:02 crc kubenswrapper[4632]: I0313 10:19:02.054473 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adab5e58-1b8e-4170-b244-d45be51beccb" path="/var/lib/kubelet/pods/adab5e58-1b8e-4170-b244-d45be51beccb/volumes" Mar 13 10:19:03 crc kubenswrapper[4632]: I0313 10:19:03.376172 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-mpfnk" Mar 13 10:19:03 crc kubenswrapper[4632]: I0313 10:19:03.758344 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:19:03 crc kubenswrapper[4632]: I0313 10:19:03.758404 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:19:03 crc kubenswrapper[4632]: I0313 10:19:03.763382 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:19:04 crc kubenswrapper[4632]: I0313 10:19:04.051445 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 10:19:04 crc kubenswrapper[4632]: I0313 10:19:04.111840 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-zn7mn"] Mar 13 10:19:13 crc kubenswrapper[4632]: I0313 10:19:13.711557 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" Mar 13 10:19:26 crc kubenswrapper[4632]: I0313 10:19:26.989804 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj"] Mar 13 10:19:26 crc kubenswrapper[4632]: E0313 10:19:26.990682 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adab5e58-1b8e-4170-b244-d45be51beccb" containerName="registry-server" Mar 13 10:19:26 crc kubenswrapper[4632]: I0313 10:19:26.990702 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="adab5e58-1b8e-4170-b244-d45be51beccb" containerName="registry-server" Mar 13 10:19:26 crc kubenswrapper[4632]: E0313 10:19:26.990719 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adab5e58-1b8e-4170-b244-d45be51beccb" containerName="extract-content" Mar 13 10:19:26 crc kubenswrapper[4632]: I0313 10:19:26.990727 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="adab5e58-1b8e-4170-b244-d45be51beccb" containerName="extract-content" Mar 13 10:19:26 crc kubenswrapper[4632]: E0313 10:19:26.990739 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adab5e58-1b8e-4170-b244-d45be51beccb" containerName="extract-utilities" Mar 13 10:19:26 crc kubenswrapper[4632]: I0313 10:19:26.990747 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="adab5e58-1b8e-4170-b244-d45be51beccb" containerName="extract-utilities" Mar 13 10:19:26 crc kubenswrapper[4632]: I0313 10:19:26.990867 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="adab5e58-1b8e-4170-b244-d45be51beccb" containerName="registry-server" Mar 13 10:19:26 crc kubenswrapper[4632]: I0313 10:19:26.991823 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" Mar 13 10:19:26 crc kubenswrapper[4632]: I0313 10:19:26.993654 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 10:19:27 crc kubenswrapper[4632]: I0313 10:19:27.002427 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj"] Mar 13 10:19:27 crc kubenswrapper[4632]: I0313 10:19:27.075362 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hh6ml\" (UniqueName: \"kubernetes.io/projected/8c1e4d78-3f38-48b5-b157-a1a076f31b76-kube-api-access-hh6ml\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj\" (UID: \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" Mar 13 10:19:27 crc kubenswrapper[4632]: I0313 10:19:27.075567 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c1e4d78-3f38-48b5-b157-a1a076f31b76-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj\" (UID: \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" Mar 13 10:19:27 crc kubenswrapper[4632]: I0313 10:19:27.075597 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c1e4d78-3f38-48b5-b157-a1a076f31b76-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj\" (UID: \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" Mar 13 10:19:27 crc kubenswrapper[4632]: I0313 10:19:27.176552 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c1e4d78-3f38-48b5-b157-a1a076f31b76-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj\" (UID: \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" Mar 13 10:19:27 crc kubenswrapper[4632]: I0313 10:19:27.176618 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c1e4d78-3f38-48b5-b157-a1a076f31b76-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj\" (UID: \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" Mar 13 10:19:27 crc kubenswrapper[4632]: I0313 10:19:27.176676 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hh6ml\" (UniqueName: \"kubernetes.io/projected/8c1e4d78-3f38-48b5-b157-a1a076f31b76-kube-api-access-hh6ml\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj\" (UID: \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" Mar 13 10:19:27 crc kubenswrapper[4632]: I0313 10:19:27.177173 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c1e4d78-3f38-48b5-b157-a1a076f31b76-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj\" (UID: \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" Mar 13 10:19:27 crc kubenswrapper[4632]: I0313 10:19:27.177239 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c1e4d78-3f38-48b5-b157-a1a076f31b76-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj\" (UID: \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" Mar 13 10:19:27 crc kubenswrapper[4632]: I0313 10:19:27.210408 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hh6ml\" (UniqueName: \"kubernetes.io/projected/8c1e4d78-3f38-48b5-b157-a1a076f31b76-kube-api-access-hh6ml\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj\" (UID: \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" Mar 13 10:19:27 crc kubenswrapper[4632]: I0313 10:19:27.306793 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" Mar 13 10:19:27 crc kubenswrapper[4632]: I0313 10:19:27.499314 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj"] Mar 13 10:19:28 crc kubenswrapper[4632]: I0313 10:19:28.394362 4632 generic.go:334] "Generic (PLEG): container finished" podID="8c1e4d78-3f38-48b5-b157-a1a076f31b76" containerID="dd017346c528d3551d35eba6b6c7d7562e4c30d260c2251b9178a5bce35ecced" exitCode=0 Mar 13 10:19:28 crc kubenswrapper[4632]: I0313 10:19:28.394439 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" event={"ID":"8c1e4d78-3f38-48b5-b157-a1a076f31b76","Type":"ContainerDied","Data":"dd017346c528d3551d35eba6b6c7d7562e4c30d260c2251b9178a5bce35ecced"} Mar 13 10:19:28 crc kubenswrapper[4632]: I0313 10:19:28.396305 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" event={"ID":"8c1e4d78-3f38-48b5-b157-a1a076f31b76","Type":"ContainerStarted","Data":"0e375dc0f6df9d2f12b95e2145547fe9337a959c6c2f008f89398e35282d0f19"} Mar 13 10:19:28 crc kubenswrapper[4632]: I0313 10:19:28.395922 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.182702 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-zn7mn" podUID="f5a50074-5531-442f-a0e9-0578f15634c1" containerName="console" containerID="cri-o://662793b7c27b62a99fd064350b3cd52eb21f393bbf5603bbcbf03a65855922bf" gracePeriod=15 Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.407738 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-zn7mn_f5a50074-5531-442f-a0e9-0578f15634c1/console/0.log" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.408036 4632 generic.go:334] "Generic (PLEG): container finished" podID="f5a50074-5531-442f-a0e9-0578f15634c1" containerID="662793b7c27b62a99fd064350b3cd52eb21f393bbf5603bbcbf03a65855922bf" exitCode=2 Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.408078 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zn7mn" event={"ID":"f5a50074-5531-442f-a0e9-0578f15634c1","Type":"ContainerDied","Data":"662793b7c27b62a99fd064350b3cd52eb21f393bbf5603bbcbf03a65855922bf"} Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.540958 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-zn7mn_f5a50074-5531-442f-a0e9-0578f15634c1/console/0.log" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.541078 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.613557 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f5a50074-5531-442f-a0e9-0578f15634c1-console-oauth-config\") pod \"f5a50074-5531-442f-a0e9-0578f15634c1\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.613634 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-trusted-ca-bundle\") pod \"f5a50074-5531-442f-a0e9-0578f15634c1\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.613683 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f5a50074-5531-442f-a0e9-0578f15634c1-console-serving-cert\") pod \"f5a50074-5531-442f-a0e9-0578f15634c1\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.613769 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-console-config\") pod \"f5a50074-5531-442f-a0e9-0578f15634c1\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.613797 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-service-ca\") pod \"f5a50074-5531-442f-a0e9-0578f15634c1\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.613847 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-oauth-serving-cert\") pod \"f5a50074-5531-442f-a0e9-0578f15634c1\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.613871 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gpqvj\" (UniqueName: \"kubernetes.io/projected/f5a50074-5531-442f-a0e9-0578f15634c1-kube-api-access-gpqvj\") pod \"f5a50074-5531-442f-a0e9-0578f15634c1\" (UID: \"f5a50074-5531-442f-a0e9-0578f15634c1\") " Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.615328 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-console-config" (OuterVolumeSpecName: "console-config") pod "f5a50074-5531-442f-a0e9-0578f15634c1" (UID: "f5a50074-5531-442f-a0e9-0578f15634c1"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.615370 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "f5a50074-5531-442f-a0e9-0578f15634c1" (UID: "f5a50074-5531-442f-a0e9-0578f15634c1"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.615427 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-service-ca" (OuterVolumeSpecName: "service-ca") pod "f5a50074-5531-442f-a0e9-0578f15634c1" (UID: "f5a50074-5531-442f-a0e9-0578f15634c1"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.617521 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "f5a50074-5531-442f-a0e9-0578f15634c1" (UID: "f5a50074-5531-442f-a0e9-0578f15634c1"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.626991 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5a50074-5531-442f-a0e9-0578f15634c1-kube-api-access-gpqvj" (OuterVolumeSpecName: "kube-api-access-gpqvj") pod "f5a50074-5531-442f-a0e9-0578f15634c1" (UID: "f5a50074-5531-442f-a0e9-0578f15634c1"). InnerVolumeSpecName "kube-api-access-gpqvj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.627304 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5a50074-5531-442f-a0e9-0578f15634c1-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "f5a50074-5531-442f-a0e9-0578f15634c1" (UID: "f5a50074-5531-442f-a0e9-0578f15634c1"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.651480 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5a50074-5531-442f-a0e9-0578f15634c1-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "f5a50074-5531-442f-a0e9-0578f15634c1" (UID: "f5a50074-5531-442f-a0e9-0578f15634c1"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.715634 4632 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-console-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.715668 4632 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.715704 4632 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.715717 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gpqvj\" (UniqueName: \"kubernetes.io/projected/f5a50074-5531-442f-a0e9-0578f15634c1-kube-api-access-gpqvj\") on node \"crc\" DevicePath \"\"" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.715731 4632 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f5a50074-5531-442f-a0e9-0578f15634c1-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.715744 4632 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f5a50074-5531-442f-a0e9-0578f15634c1-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:19:29 crc kubenswrapper[4632]: I0313 10:19:29.715779 4632 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f5a50074-5531-442f-a0e9-0578f15634c1-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 10:19:30 crc kubenswrapper[4632]: I0313 10:19:30.416242 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-zn7mn_f5a50074-5531-442f-a0e9-0578f15634c1/console/0.log" Mar 13 10:19:30 crc kubenswrapper[4632]: I0313 10:19:30.417160 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-zn7mn" event={"ID":"f5a50074-5531-442f-a0e9-0578f15634c1","Type":"ContainerDied","Data":"c0f56571b6b9472de716bb190b1d68fe783e6f7b131b06ae9b0c01071f1d985f"} Mar 13 10:19:30 crc kubenswrapper[4632]: I0313 10:19:30.417200 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-zn7mn" Mar 13 10:19:30 crc kubenswrapper[4632]: I0313 10:19:30.417207 4632 scope.go:117] "RemoveContainer" containerID="662793b7c27b62a99fd064350b3cd52eb21f393bbf5603bbcbf03a65855922bf" Mar 13 10:19:30 crc kubenswrapper[4632]: I0313 10:19:30.421174 4632 generic.go:334] "Generic (PLEG): container finished" podID="8c1e4d78-3f38-48b5-b157-a1a076f31b76" containerID="f6eab6c1270c708d5635fc346dbf0abff23385e6fdb0b88031343e1eeb347e7e" exitCode=0 Mar 13 10:19:30 crc kubenswrapper[4632]: I0313 10:19:30.421226 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" event={"ID":"8c1e4d78-3f38-48b5-b157-a1a076f31b76","Type":"ContainerDied","Data":"f6eab6c1270c708d5635fc346dbf0abff23385e6fdb0b88031343e1eeb347e7e"} Mar 13 10:19:30 crc kubenswrapper[4632]: I0313 10:19:30.438578 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-zn7mn"] Mar 13 10:19:30 crc kubenswrapper[4632]: I0313 10:19:30.443318 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-zn7mn"] Mar 13 10:19:31 crc kubenswrapper[4632]: I0313 10:19:31.428475 4632 generic.go:334] "Generic (PLEG): container finished" podID="8c1e4d78-3f38-48b5-b157-a1a076f31b76" containerID="35bf12e34ddb56d3a7959a6cf56252b0e8e3b3a801abde467165cc190ac8dec4" exitCode=0 Mar 13 10:19:31 crc kubenswrapper[4632]: I0313 10:19:31.428558 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" event={"ID":"8c1e4d78-3f38-48b5-b157-a1a076f31b76","Type":"ContainerDied","Data":"35bf12e34ddb56d3a7959a6cf56252b0e8e3b3a801abde467165cc190ac8dec4"} Mar 13 10:19:32 crc kubenswrapper[4632]: I0313 10:19:32.060359 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5a50074-5531-442f-a0e9-0578f15634c1" path="/var/lib/kubelet/pods/f5a50074-5531-442f-a0e9-0578f15634c1/volumes" Mar 13 10:19:32 crc kubenswrapper[4632]: I0313 10:19:32.630646 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" Mar 13 10:19:32 crc kubenswrapper[4632]: I0313 10:19:32.657894 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c1e4d78-3f38-48b5-b157-a1a076f31b76-util\") pod \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\" (UID: \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\") " Mar 13 10:19:32 crc kubenswrapper[4632]: I0313 10:19:32.658042 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hh6ml\" (UniqueName: \"kubernetes.io/projected/8c1e4d78-3f38-48b5-b157-a1a076f31b76-kube-api-access-hh6ml\") pod \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\" (UID: \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\") " Mar 13 10:19:32 crc kubenswrapper[4632]: I0313 10:19:32.658077 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c1e4d78-3f38-48b5-b157-a1a076f31b76-bundle\") pod \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\" (UID: \"8c1e4d78-3f38-48b5-b157-a1a076f31b76\") " Mar 13 10:19:32 crc kubenswrapper[4632]: I0313 10:19:32.659147 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c1e4d78-3f38-48b5-b157-a1a076f31b76-bundle" (OuterVolumeSpecName: "bundle") pod "8c1e4d78-3f38-48b5-b157-a1a076f31b76" (UID: "8c1e4d78-3f38-48b5-b157-a1a076f31b76"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:19:32 crc kubenswrapper[4632]: I0313 10:19:32.664977 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c1e4d78-3f38-48b5-b157-a1a076f31b76-kube-api-access-hh6ml" (OuterVolumeSpecName: "kube-api-access-hh6ml") pod "8c1e4d78-3f38-48b5-b157-a1a076f31b76" (UID: "8c1e4d78-3f38-48b5-b157-a1a076f31b76"). InnerVolumeSpecName "kube-api-access-hh6ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:19:32 crc kubenswrapper[4632]: I0313 10:19:32.678428 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c1e4d78-3f38-48b5-b157-a1a076f31b76-util" (OuterVolumeSpecName: "util") pod "8c1e4d78-3f38-48b5-b157-a1a076f31b76" (UID: "8c1e4d78-3f38-48b5-b157-a1a076f31b76"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:19:32 crc kubenswrapper[4632]: I0313 10:19:32.759312 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hh6ml\" (UniqueName: \"kubernetes.io/projected/8c1e4d78-3f38-48b5-b157-a1a076f31b76-kube-api-access-hh6ml\") on node \"crc\" DevicePath \"\"" Mar 13 10:19:32 crc kubenswrapper[4632]: I0313 10:19:32.759360 4632 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8c1e4d78-3f38-48b5-b157-a1a076f31b76-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:19:32 crc kubenswrapper[4632]: I0313 10:19:32.759369 4632 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8c1e4d78-3f38-48b5-b157-a1a076f31b76-util\") on node \"crc\" DevicePath \"\"" Mar 13 10:19:33 crc kubenswrapper[4632]: I0313 10:19:33.443154 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" event={"ID":"8c1e4d78-3f38-48b5-b157-a1a076f31b76","Type":"ContainerDied","Data":"0e375dc0f6df9d2f12b95e2145547fe9337a959c6c2f008f89398e35282d0f19"} Mar 13 10:19:33 crc kubenswrapper[4632]: I0313 10:19:33.443206 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e375dc0f6df9d2f12b95e2145547fe9337a959c6c2f008f89398e35282d0f19" Mar 13 10:19:33 crc kubenswrapper[4632]: I0313 10:19:33.443208 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.802435 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq"] Mar 13 10:19:42 crc kubenswrapper[4632]: E0313 10:19:42.803001 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c1e4d78-3f38-48b5-b157-a1a076f31b76" containerName="extract" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.803014 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c1e4d78-3f38-48b5-b157-a1a076f31b76" containerName="extract" Mar 13 10:19:42 crc kubenswrapper[4632]: E0313 10:19:42.803025 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a50074-5531-442f-a0e9-0578f15634c1" containerName="console" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.803031 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a50074-5531-442f-a0e9-0578f15634c1" containerName="console" Mar 13 10:19:42 crc kubenswrapper[4632]: E0313 10:19:42.803047 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c1e4d78-3f38-48b5-b157-a1a076f31b76" containerName="util" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.803053 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c1e4d78-3f38-48b5-b157-a1a076f31b76" containerName="util" Mar 13 10:19:42 crc kubenswrapper[4632]: E0313 10:19:42.803065 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c1e4d78-3f38-48b5-b157-a1a076f31b76" containerName="pull" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.803070 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c1e4d78-3f38-48b5-b157-a1a076f31b76" containerName="pull" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.803158 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a50074-5531-442f-a0e9-0578f15634c1" containerName="console" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.803168 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c1e4d78-3f38-48b5-b157-a1a076f31b76" containerName="extract" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.803525 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.806550 4632 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-8ndqr" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.807505 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.807614 4632 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.808302 4632 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.808620 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.819095 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq"] Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.881738 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e62d674f-5b2c-4788-85a3-95b51621dbef-webhook-cert\") pod \"metallb-operator-controller-manager-ffdcc767b-qxvlq\" (UID: \"e62d674f-5b2c-4788-85a3-95b51621dbef\") " pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.881810 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72kgf\" (UniqueName: \"kubernetes.io/projected/e62d674f-5b2c-4788-85a3-95b51621dbef-kube-api-access-72kgf\") pod \"metallb-operator-controller-manager-ffdcc767b-qxvlq\" (UID: \"e62d674f-5b2c-4788-85a3-95b51621dbef\") " pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.881882 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e62d674f-5b2c-4788-85a3-95b51621dbef-apiservice-cert\") pod \"metallb-operator-controller-manager-ffdcc767b-qxvlq\" (UID: \"e62d674f-5b2c-4788-85a3-95b51621dbef\") " pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.983450 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e62d674f-5b2c-4788-85a3-95b51621dbef-apiservice-cert\") pod \"metallb-operator-controller-manager-ffdcc767b-qxvlq\" (UID: \"e62d674f-5b2c-4788-85a3-95b51621dbef\") " pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.983731 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e62d674f-5b2c-4788-85a3-95b51621dbef-webhook-cert\") pod \"metallb-operator-controller-manager-ffdcc767b-qxvlq\" (UID: \"e62d674f-5b2c-4788-85a3-95b51621dbef\") " pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.983759 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72kgf\" (UniqueName: \"kubernetes.io/projected/e62d674f-5b2c-4788-85a3-95b51621dbef-kube-api-access-72kgf\") pod \"metallb-operator-controller-manager-ffdcc767b-qxvlq\" (UID: \"e62d674f-5b2c-4788-85a3-95b51621dbef\") " pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.990907 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e62d674f-5b2c-4788-85a3-95b51621dbef-apiservice-cert\") pod \"metallb-operator-controller-manager-ffdcc767b-qxvlq\" (UID: \"e62d674f-5b2c-4788-85a3-95b51621dbef\") " pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" Mar 13 10:19:42 crc kubenswrapper[4632]: I0313 10:19:42.991786 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e62d674f-5b2c-4788-85a3-95b51621dbef-webhook-cert\") pod \"metallb-operator-controller-manager-ffdcc767b-qxvlq\" (UID: \"e62d674f-5b2c-4788-85a3-95b51621dbef\") " pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.002124 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72kgf\" (UniqueName: \"kubernetes.io/projected/e62d674f-5b2c-4788-85a3-95b51621dbef-kube-api-access-72kgf\") pod \"metallb-operator-controller-manager-ffdcc767b-qxvlq\" (UID: \"e62d674f-5b2c-4788-85a3-95b51621dbef\") " pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.065297 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l"] Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.066133 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.068020 4632 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.068972 4632 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.069338 4632 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-w8ltf" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.084348 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/712b2002-4fce-4983-926a-99a4b2dc7a8c-webhook-cert\") pod \"metallb-operator-webhook-server-6c7bf5ddc5-v6t5l\" (UID: \"712b2002-4fce-4983-926a-99a4b2dc7a8c\") " pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.089231 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l"] Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.090558 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/712b2002-4fce-4983-926a-99a4b2dc7a8c-apiservice-cert\") pod \"metallb-operator-webhook-server-6c7bf5ddc5-v6t5l\" (UID: \"712b2002-4fce-4983-926a-99a4b2dc7a8c\") " pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.090770 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zcw9\" (UniqueName: \"kubernetes.io/projected/712b2002-4fce-4983-926a-99a4b2dc7a8c-kube-api-access-7zcw9\") pod \"metallb-operator-webhook-server-6c7bf5ddc5-v6t5l\" (UID: \"712b2002-4fce-4983-926a-99a4b2dc7a8c\") " pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.121227 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.191527 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/712b2002-4fce-4983-926a-99a4b2dc7a8c-apiservice-cert\") pod \"metallb-operator-webhook-server-6c7bf5ddc5-v6t5l\" (UID: \"712b2002-4fce-4983-926a-99a4b2dc7a8c\") " pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.192323 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zcw9\" (UniqueName: \"kubernetes.io/projected/712b2002-4fce-4983-926a-99a4b2dc7a8c-kube-api-access-7zcw9\") pod \"metallb-operator-webhook-server-6c7bf5ddc5-v6t5l\" (UID: \"712b2002-4fce-4983-926a-99a4b2dc7a8c\") " pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.194965 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/712b2002-4fce-4983-926a-99a4b2dc7a8c-webhook-cert\") pod \"metallb-operator-webhook-server-6c7bf5ddc5-v6t5l\" (UID: \"712b2002-4fce-4983-926a-99a4b2dc7a8c\") " pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.194830 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/712b2002-4fce-4983-926a-99a4b2dc7a8c-apiservice-cert\") pod \"metallb-operator-webhook-server-6c7bf5ddc5-v6t5l\" (UID: \"712b2002-4fce-4983-926a-99a4b2dc7a8c\") " pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.201505 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/712b2002-4fce-4983-926a-99a4b2dc7a8c-webhook-cert\") pod \"metallb-operator-webhook-server-6c7bf5ddc5-v6t5l\" (UID: \"712b2002-4fce-4983-926a-99a4b2dc7a8c\") " pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.243687 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zcw9\" (UniqueName: \"kubernetes.io/projected/712b2002-4fce-4983-926a-99a4b2dc7a8c-kube-api-access-7zcw9\") pod \"metallb-operator-webhook-server-6c7bf5ddc5-v6t5l\" (UID: \"712b2002-4fce-4983-926a-99a4b2dc7a8c\") " pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.382347 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.603721 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq"] Mar 13 10:19:43 crc kubenswrapper[4632]: W0313 10:19:43.625385 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode62d674f_5b2c_4788_85a3_95b51621dbef.slice/crio-36539b45fbaabd48077b1e3bf16bc5b42f8aeb664d9ffa7f5fd14bd9c877f6b3 WatchSource:0}: Error finding container 36539b45fbaabd48077b1e3bf16bc5b42f8aeb664d9ffa7f5fd14bd9c877f6b3: Status 404 returned error can't find the container with id 36539b45fbaabd48077b1e3bf16bc5b42f8aeb664d9ffa7f5fd14bd9c877f6b3 Mar 13 10:19:43 crc kubenswrapper[4632]: I0313 10:19:43.843798 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l"] Mar 13 10:19:43 crc kubenswrapper[4632]: W0313 10:19:43.863009 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod712b2002_4fce_4983_926a_99a4b2dc7a8c.slice/crio-a456a30659a7a3870cf65fd46e2de49b22835f180cfc7ed293faa6d6aedac332 WatchSource:0}: Error finding container a456a30659a7a3870cf65fd46e2de49b22835f180cfc7ed293faa6d6aedac332: Status 404 returned error can't find the container with id a456a30659a7a3870cf65fd46e2de49b22835f180cfc7ed293faa6d6aedac332 Mar 13 10:19:44 crc kubenswrapper[4632]: I0313 10:19:44.509773 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" event={"ID":"e62d674f-5b2c-4788-85a3-95b51621dbef","Type":"ContainerStarted","Data":"36539b45fbaabd48077b1e3bf16bc5b42f8aeb664d9ffa7f5fd14bd9c877f6b3"} Mar 13 10:19:44 crc kubenswrapper[4632]: I0313 10:19:44.511217 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" event={"ID":"712b2002-4fce-4983-926a-99a4b2dc7a8c","Type":"ContainerStarted","Data":"a456a30659a7a3870cf65fd46e2de49b22835f180cfc7ed293faa6d6aedac332"} Mar 13 10:19:52 crc kubenswrapper[4632]: I0313 10:19:52.568025 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" event={"ID":"e62d674f-5b2c-4788-85a3-95b51621dbef","Type":"ContainerStarted","Data":"d10aebbd8471a57b9c8fb772d9957bcb409f35eba90c3ff18f788ca610c86020"} Mar 13 10:19:52 crc kubenswrapper[4632]: I0313 10:19:52.568565 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" Mar 13 10:19:52 crc kubenswrapper[4632]: I0313 10:19:52.569858 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" event={"ID":"712b2002-4fce-4983-926a-99a4b2dc7a8c","Type":"ContainerStarted","Data":"4f3954f3545fe16f459ae776b9c8dd134a4be9b3933eebfce6eab0b02f3d82e6"} Mar 13 10:19:52 crc kubenswrapper[4632]: I0313 10:19:52.570033 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" Mar 13 10:19:52 crc kubenswrapper[4632]: I0313 10:19:52.592978 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" podStartSLOduration=2.822942367 podStartE2EDuration="10.592931729s" podCreationTimestamp="2026-03-13 10:19:42 +0000 UTC" firstStartedPulling="2026-03-13 10:19:43.632470075 +0000 UTC m=+957.655000198" lastFinishedPulling="2026-03-13 10:19:51.402459427 +0000 UTC m=+965.424989560" observedRunningTime="2026-03-13 10:19:52.585197835 +0000 UTC m=+966.607727988" watchObservedRunningTime="2026-03-13 10:19:52.592931729 +0000 UTC m=+966.615461902" Mar 13 10:19:52 crc kubenswrapper[4632]: I0313 10:19:52.605861 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" podStartSLOduration=2.008069632 podStartE2EDuration="9.60584267s" podCreationTimestamp="2026-03-13 10:19:43 +0000 UTC" firstStartedPulling="2026-03-13 10:19:43.866537801 +0000 UTC m=+957.889067934" lastFinishedPulling="2026-03-13 10:19:51.464310829 +0000 UTC m=+965.486840972" observedRunningTime="2026-03-13 10:19:52.603739328 +0000 UTC m=+966.626269481" watchObservedRunningTime="2026-03-13 10:19:52.60584267 +0000 UTC m=+966.628372803" Mar 13 10:20:00 crc kubenswrapper[4632]: I0313 10:20:00.140755 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556620-42vs6"] Mar 13 10:20:00 crc kubenswrapper[4632]: I0313 10:20:00.142258 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556620-42vs6" Mar 13 10:20:00 crc kubenswrapper[4632]: I0313 10:20:00.151915 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556620-42vs6"] Mar 13 10:20:00 crc kubenswrapper[4632]: I0313 10:20:00.153099 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:20:00 crc kubenswrapper[4632]: I0313 10:20:00.153524 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:20:00 crc kubenswrapper[4632]: I0313 10:20:00.154103 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:20:00 crc kubenswrapper[4632]: I0313 10:20:00.232888 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6c8k\" (UniqueName: \"kubernetes.io/projected/28dbef1d-ca7f-4387-80af-8dffbfe92895-kube-api-access-w6c8k\") pod \"auto-csr-approver-29556620-42vs6\" (UID: \"28dbef1d-ca7f-4387-80af-8dffbfe92895\") " pod="openshift-infra/auto-csr-approver-29556620-42vs6" Mar 13 10:20:00 crc kubenswrapper[4632]: I0313 10:20:00.334256 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6c8k\" (UniqueName: \"kubernetes.io/projected/28dbef1d-ca7f-4387-80af-8dffbfe92895-kube-api-access-w6c8k\") pod \"auto-csr-approver-29556620-42vs6\" (UID: \"28dbef1d-ca7f-4387-80af-8dffbfe92895\") " pod="openshift-infra/auto-csr-approver-29556620-42vs6" Mar 13 10:20:00 crc kubenswrapper[4632]: I0313 10:20:00.363195 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6c8k\" (UniqueName: \"kubernetes.io/projected/28dbef1d-ca7f-4387-80af-8dffbfe92895-kube-api-access-w6c8k\") pod \"auto-csr-approver-29556620-42vs6\" (UID: \"28dbef1d-ca7f-4387-80af-8dffbfe92895\") " pod="openshift-infra/auto-csr-approver-29556620-42vs6" Mar 13 10:20:00 crc kubenswrapper[4632]: I0313 10:20:00.464264 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556620-42vs6" Mar 13 10:20:00 crc kubenswrapper[4632]: I0313 10:20:00.696524 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556620-42vs6"] Mar 13 10:20:01 crc kubenswrapper[4632]: I0313 10:20:01.619932 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556620-42vs6" event={"ID":"28dbef1d-ca7f-4387-80af-8dffbfe92895","Type":"ContainerStarted","Data":"b8eaff64006265fcd0d329ca24eef7b4cfa19bbf794b80db42c9037947401d93"} Mar 13 10:20:02 crc kubenswrapper[4632]: I0313 10:20:02.628246 4632 generic.go:334] "Generic (PLEG): container finished" podID="28dbef1d-ca7f-4387-80af-8dffbfe92895" containerID="746bd1f1584c6b468985171d618d35f15871608c045fd5e9f4070c7ace66e505" exitCode=0 Mar 13 10:20:02 crc kubenswrapper[4632]: I0313 10:20:02.628316 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556620-42vs6" event={"ID":"28dbef1d-ca7f-4387-80af-8dffbfe92895","Type":"ContainerDied","Data":"746bd1f1584c6b468985171d618d35f15871608c045fd5e9f4070c7ace66e505"} Mar 13 10:20:03 crc kubenswrapper[4632]: I0313 10:20:03.390274 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" Mar 13 10:20:03 crc kubenswrapper[4632]: I0313 10:20:03.919605 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556620-42vs6" Mar 13 10:20:03 crc kubenswrapper[4632]: I0313 10:20:03.985844 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w6c8k\" (UniqueName: \"kubernetes.io/projected/28dbef1d-ca7f-4387-80af-8dffbfe92895-kube-api-access-w6c8k\") pod \"28dbef1d-ca7f-4387-80af-8dffbfe92895\" (UID: \"28dbef1d-ca7f-4387-80af-8dffbfe92895\") " Mar 13 10:20:03 crc kubenswrapper[4632]: I0313 10:20:03.990798 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28dbef1d-ca7f-4387-80af-8dffbfe92895-kube-api-access-w6c8k" (OuterVolumeSpecName: "kube-api-access-w6c8k") pod "28dbef1d-ca7f-4387-80af-8dffbfe92895" (UID: "28dbef1d-ca7f-4387-80af-8dffbfe92895"). InnerVolumeSpecName "kube-api-access-w6c8k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:20:04 crc kubenswrapper[4632]: I0313 10:20:04.087741 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w6c8k\" (UniqueName: \"kubernetes.io/projected/28dbef1d-ca7f-4387-80af-8dffbfe92895-kube-api-access-w6c8k\") on node \"crc\" DevicePath \"\"" Mar 13 10:20:04 crc kubenswrapper[4632]: I0313 10:20:04.638788 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556620-42vs6" event={"ID":"28dbef1d-ca7f-4387-80af-8dffbfe92895","Type":"ContainerDied","Data":"b8eaff64006265fcd0d329ca24eef7b4cfa19bbf794b80db42c9037947401d93"} Mar 13 10:20:04 crc kubenswrapper[4632]: I0313 10:20:04.639217 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b8eaff64006265fcd0d329ca24eef7b4cfa19bbf794b80db42c9037947401d93" Mar 13 10:20:04 crc kubenswrapper[4632]: I0313 10:20:04.639130 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556620-42vs6" Mar 13 10:20:04 crc kubenswrapper[4632]: I0313 10:20:04.971901 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556614-7pzwt"] Mar 13 10:20:04 crc kubenswrapper[4632]: I0313 10:20:04.975508 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556614-7pzwt"] Mar 13 10:20:06 crc kubenswrapper[4632]: I0313 10:20:06.060873 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f80bfe67-be24-45e3-9e57-b67389f8cc63" path="/var/lib/kubelet/pods/f80bfe67-be24-45e3-9e57-b67389f8cc63/volumes" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.125661 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-ffdcc767b-qxvlq" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.858664 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-lvlxj"] Mar 13 10:20:23 crc kubenswrapper[4632]: E0313 10:20:23.858925 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28dbef1d-ca7f-4387-80af-8dffbfe92895" containerName="oc" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.858978 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="28dbef1d-ca7f-4387-80af-8dffbfe92895" containerName="oc" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.859113 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="28dbef1d-ca7f-4387-80af-8dffbfe92895" containerName="oc" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.861120 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.864347 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8"] Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.865372 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.866763 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.866862 4632 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.867157 4632 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-cmb4l" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.869015 4632 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.887325 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8"] Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.953389 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-tztd9"] Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.954291 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-tztd9" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.957713 4632 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.957927 4632 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.958115 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.958266 4632 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-v97kj" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.966559 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92cl4\" (UniqueName: \"kubernetes.io/projected/b33bccd8-6f28-4ffe-9500-069a52aab5df-kube-api-access-92cl4\") pod \"frr-k8s-webhook-server-bcc4b6f68-9zbh8\" (UID: \"b33bccd8-6f28-4ffe-9500-069a52aab5df\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.966633 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/85b58bb0-63f5-4c85-8759-ce28d2c7db58-frr-conf\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.966666 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxgsd\" (UniqueName: \"kubernetes.io/projected/85b58bb0-63f5-4c85-8759-ce28d2c7db58-kube-api-access-lxgsd\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.966813 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/85b58bb0-63f5-4c85-8759-ce28d2c7db58-frr-startup\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.966892 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85b58bb0-63f5-4c85-8759-ce28d2c7db58-metrics-certs\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.966976 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b33bccd8-6f28-4ffe-9500-069a52aab5df-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-9zbh8\" (UID: \"b33bccd8-6f28-4ffe-9500-069a52aab5df\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.967063 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/85b58bb0-63f5-4c85-8759-ce28d2c7db58-reloader\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.967087 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/85b58bb0-63f5-4c85-8759-ce28d2c7db58-frr-sockets\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.967133 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/85b58bb0-63f5-4c85-8759-ce28d2c7db58-metrics\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.987834 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-62bwr"] Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.988687 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-62bwr" Mar 13 10:20:23 crc kubenswrapper[4632]: I0313 10:20:23.991380 4632 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.002834 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-62bwr"] Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.068682 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxgsd\" (UniqueName: \"kubernetes.io/projected/85b58bb0-63f5-4c85-8759-ce28d2c7db58-kube-api-access-lxgsd\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.068739 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/277ddd7f-fd9c-4b27-9563-c904f1dffd40-metrics-certs\") pod \"controller-7bb4cc7c98-62bwr\" (UID: \"277ddd7f-fd9c-4b27-9563-c904f1dffd40\") " pod="metallb-system/controller-7bb4cc7c98-62bwr" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.068766 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/85b58bb0-63f5-4c85-8759-ce28d2c7db58-frr-startup\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.068793 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-metrics-certs\") pod \"speaker-tztd9\" (UID: \"8f51973a-596d-40dc-9b5b-b2c95a60ea0c\") " pod="metallb-system/speaker-tztd9" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.068815 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85b58bb0-63f5-4c85-8759-ce28d2c7db58-metrics-certs\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.068842 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b33bccd8-6f28-4ffe-9500-069a52aab5df-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-9zbh8\" (UID: \"b33bccd8-6f28-4ffe-9500-069a52aab5df\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.068865 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-metallb-excludel2\") pod \"speaker-tztd9\" (UID: \"8f51973a-596d-40dc-9b5b-b2c95a60ea0c\") " pod="metallb-system/speaker-tztd9" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.068887 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffpbj\" (UniqueName: \"kubernetes.io/projected/277ddd7f-fd9c-4b27-9563-c904f1dffd40-kube-api-access-ffpbj\") pod \"controller-7bb4cc7c98-62bwr\" (UID: \"277ddd7f-fd9c-4b27-9563-c904f1dffd40\") " pod="metallb-system/controller-7bb4cc7c98-62bwr" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.068908 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/277ddd7f-fd9c-4b27-9563-c904f1dffd40-cert\") pod \"controller-7bb4cc7c98-62bwr\" (UID: \"277ddd7f-fd9c-4b27-9563-c904f1dffd40\") " pod="metallb-system/controller-7bb4cc7c98-62bwr" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.068923 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/85b58bb0-63f5-4c85-8759-ce28d2c7db58-reloader\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.068953 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/85b58bb0-63f5-4c85-8759-ce28d2c7db58-frr-sockets\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.068971 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-memberlist\") pod \"speaker-tztd9\" (UID: \"8f51973a-596d-40dc-9b5b-b2c95a60ea0c\") " pod="metallb-system/speaker-tztd9" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.068989 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/85b58bb0-63f5-4c85-8759-ce28d2c7db58-metrics\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.069019 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92cl4\" (UniqueName: \"kubernetes.io/projected/b33bccd8-6f28-4ffe-9500-069a52aab5df-kube-api-access-92cl4\") pod \"frr-k8s-webhook-server-bcc4b6f68-9zbh8\" (UID: \"b33bccd8-6f28-4ffe-9500-069a52aab5df\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.069039 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd9n6\" (UniqueName: \"kubernetes.io/projected/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-kube-api-access-qd9n6\") pod \"speaker-tztd9\" (UID: \"8f51973a-596d-40dc-9b5b-b2c95a60ea0c\") " pod="metallb-system/speaker-tztd9" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.069058 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/85b58bb0-63f5-4c85-8759-ce28d2c7db58-frr-conf\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.069620 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/85b58bb0-63f5-4c85-8759-ce28d2c7db58-frr-conf\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.069684 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/85b58bb0-63f5-4c85-8759-ce28d2c7db58-reloader\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.069893 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/85b58bb0-63f5-4c85-8759-ce28d2c7db58-frr-sockets\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.070188 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/85b58bb0-63f5-4c85-8759-ce28d2c7db58-metrics\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: E0313 10:20:24.070517 4632 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Mar 13 10:20:24 crc kubenswrapper[4632]: E0313 10:20:24.070611 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b33bccd8-6f28-4ffe-9500-069a52aab5df-cert podName:b33bccd8-6f28-4ffe-9500-069a52aab5df nodeName:}" failed. No retries permitted until 2026-03-13 10:20:24.570585087 +0000 UTC m=+998.593115220 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b33bccd8-6f28-4ffe-9500-069a52aab5df-cert") pod "frr-k8s-webhook-server-bcc4b6f68-9zbh8" (UID: "b33bccd8-6f28-4ffe-9500-069a52aab5df") : secret "frr-k8s-webhook-server-cert" not found Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.071203 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/85b58bb0-63f5-4c85-8759-ce28d2c7db58-frr-startup\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.079793 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/85b58bb0-63f5-4c85-8759-ce28d2c7db58-metrics-certs\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.087517 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92cl4\" (UniqueName: \"kubernetes.io/projected/b33bccd8-6f28-4ffe-9500-069a52aab5df-kube-api-access-92cl4\") pod \"frr-k8s-webhook-server-bcc4b6f68-9zbh8\" (UID: \"b33bccd8-6f28-4ffe-9500-069a52aab5df\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.088245 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxgsd\" (UniqueName: \"kubernetes.io/projected/85b58bb0-63f5-4c85-8759-ce28d2c7db58-kube-api-access-lxgsd\") pod \"frr-k8s-lvlxj\" (UID: \"85b58bb0-63f5-4c85-8759-ce28d2c7db58\") " pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.170635 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-metrics-certs\") pod \"speaker-tztd9\" (UID: \"8f51973a-596d-40dc-9b5b-b2c95a60ea0c\") " pod="metallb-system/speaker-tztd9" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.170752 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-metallb-excludel2\") pod \"speaker-tztd9\" (UID: \"8f51973a-596d-40dc-9b5b-b2c95a60ea0c\") " pod="metallb-system/speaker-tztd9" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.170786 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffpbj\" (UniqueName: \"kubernetes.io/projected/277ddd7f-fd9c-4b27-9563-c904f1dffd40-kube-api-access-ffpbj\") pod \"controller-7bb4cc7c98-62bwr\" (UID: \"277ddd7f-fd9c-4b27-9563-c904f1dffd40\") " pod="metallb-system/controller-7bb4cc7c98-62bwr" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.170813 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/277ddd7f-fd9c-4b27-9563-c904f1dffd40-cert\") pod \"controller-7bb4cc7c98-62bwr\" (UID: \"277ddd7f-fd9c-4b27-9563-c904f1dffd40\") " pod="metallb-system/controller-7bb4cc7c98-62bwr" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.170839 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-memberlist\") pod \"speaker-tztd9\" (UID: \"8f51973a-596d-40dc-9b5b-b2c95a60ea0c\") " pod="metallb-system/speaker-tztd9" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.170882 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd9n6\" (UniqueName: \"kubernetes.io/projected/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-kube-api-access-qd9n6\") pod \"speaker-tztd9\" (UID: \"8f51973a-596d-40dc-9b5b-b2c95a60ea0c\") " pod="metallb-system/speaker-tztd9" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.170919 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/277ddd7f-fd9c-4b27-9563-c904f1dffd40-metrics-certs\") pod \"controller-7bb4cc7c98-62bwr\" (UID: \"277ddd7f-fd9c-4b27-9563-c904f1dffd40\") " pod="metallb-system/controller-7bb4cc7c98-62bwr" Mar 13 10:20:24 crc kubenswrapper[4632]: E0313 10:20:24.171788 4632 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 13 10:20:24 crc kubenswrapper[4632]: E0313 10:20:24.171859 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-memberlist podName:8f51973a-596d-40dc-9b5b-b2c95a60ea0c nodeName:}" failed. No retries permitted until 2026-03-13 10:20:24.671838072 +0000 UTC m=+998.694368205 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-memberlist") pod "speaker-tztd9" (UID: "8f51973a-596d-40dc-9b5b-b2c95a60ea0c") : secret "metallb-memberlist" not found Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.172470 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-metallb-excludel2\") pod \"speaker-tztd9\" (UID: \"8f51973a-596d-40dc-9b5b-b2c95a60ea0c\") " pod="metallb-system/speaker-tztd9" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.178281 4632 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.178692 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/277ddd7f-fd9c-4b27-9563-c904f1dffd40-metrics-certs\") pod \"controller-7bb4cc7c98-62bwr\" (UID: \"277ddd7f-fd9c-4b27-9563-c904f1dffd40\") " pod="metallb-system/controller-7bb4cc7c98-62bwr" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.179376 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.186022 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/277ddd7f-fd9c-4b27-9563-c904f1dffd40-cert\") pod \"controller-7bb4cc7c98-62bwr\" (UID: \"277ddd7f-fd9c-4b27-9563-c904f1dffd40\") " pod="metallb-system/controller-7bb4cc7c98-62bwr" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.199190 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffpbj\" (UniqueName: \"kubernetes.io/projected/277ddd7f-fd9c-4b27-9563-c904f1dffd40-kube-api-access-ffpbj\") pod \"controller-7bb4cc7c98-62bwr\" (UID: \"277ddd7f-fd9c-4b27-9563-c904f1dffd40\") " pod="metallb-system/controller-7bb4cc7c98-62bwr" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.201769 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd9n6\" (UniqueName: \"kubernetes.io/projected/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-kube-api-access-qd9n6\") pod \"speaker-tztd9\" (UID: \"8f51973a-596d-40dc-9b5b-b2c95a60ea0c\") " pod="metallb-system/speaker-tztd9" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.206586 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-metrics-certs\") pod \"speaker-tztd9\" (UID: \"8f51973a-596d-40dc-9b5b-b2c95a60ea0c\") " pod="metallb-system/speaker-tztd9" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.312280 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-62bwr" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.576618 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b33bccd8-6f28-4ffe-9500-069a52aab5df-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-9zbh8\" (UID: \"b33bccd8-6f28-4ffe-9500-069a52aab5df\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.582272 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b33bccd8-6f28-4ffe-9500-069a52aab5df-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-9zbh8\" (UID: \"b33bccd8-6f28-4ffe-9500-069a52aab5df\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.677975 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-memberlist\") pod \"speaker-tztd9\" (UID: \"8f51973a-596d-40dc-9b5b-b2c95a60ea0c\") " pod="metallb-system/speaker-tztd9" Mar 13 10:20:24 crc kubenswrapper[4632]: E0313 10:20:24.678152 4632 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 13 10:20:24 crc kubenswrapper[4632]: E0313 10:20:24.678249 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-memberlist podName:8f51973a-596d-40dc-9b5b-b2c95a60ea0c nodeName:}" failed. No retries permitted until 2026-03-13 10:20:25.678228918 +0000 UTC m=+999.700759051 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-memberlist") pod "speaker-tztd9" (UID: "8f51973a-596d-40dc-9b5b-b2c95a60ea0c") : secret "metallb-memberlist" not found Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.746344 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lvlxj" event={"ID":"85b58bb0-63f5-4c85-8759-ce28d2c7db58","Type":"ContainerStarted","Data":"78c4f38e92514e8e35bc3a4d59f7da89119cb64f0e178cf6c63c5d138a8a7177"} Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.760906 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-62bwr"] Mar 13 10:20:24 crc kubenswrapper[4632]: W0313 10:20:24.766808 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod277ddd7f_fd9c_4b27_9563_c904f1dffd40.slice/crio-6f91d639a73c89c4bef35530e3553cae46d82fd04f4d5159a503dd42c5f41663 WatchSource:0}: Error finding container 6f91d639a73c89c4bef35530e3553cae46d82fd04f4d5159a503dd42c5f41663: Status 404 returned error can't find the container with id 6f91d639a73c89c4bef35530e3553cae46d82fd04f4d5159a503dd42c5f41663 Mar 13 10:20:24 crc kubenswrapper[4632]: I0313 10:20:24.789522 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 10:20:25 crc kubenswrapper[4632]: I0313 10:20:25.084051 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8"] Mar 13 10:20:25 crc kubenswrapper[4632]: I0313 10:20:25.690783 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-memberlist\") pod \"speaker-tztd9\" (UID: \"8f51973a-596d-40dc-9b5b-b2c95a60ea0c\") " pod="metallb-system/speaker-tztd9" Mar 13 10:20:25 crc kubenswrapper[4632]: I0313 10:20:25.697624 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/8f51973a-596d-40dc-9b5b-b2c95a60ea0c-memberlist\") pod \"speaker-tztd9\" (UID: \"8f51973a-596d-40dc-9b5b-b2c95a60ea0c\") " pod="metallb-system/speaker-tztd9" Mar 13 10:20:25 crc kubenswrapper[4632]: I0313 10:20:25.754805 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-62bwr" event={"ID":"277ddd7f-fd9c-4b27-9563-c904f1dffd40","Type":"ContainerStarted","Data":"6d322cabee18aef00b71e74c7d9c0afec843e21a91f4426253de3249590b9941"} Mar 13 10:20:25 crc kubenswrapper[4632]: I0313 10:20:25.755097 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-62bwr" event={"ID":"277ddd7f-fd9c-4b27-9563-c904f1dffd40","Type":"ContainerStarted","Data":"b70c01aadbba4c08d00cd500d541a93419f2965f917c721c2a529c0f228795b5"} Mar 13 10:20:25 crc kubenswrapper[4632]: I0313 10:20:25.755198 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-62bwr" event={"ID":"277ddd7f-fd9c-4b27-9563-c904f1dffd40","Type":"ContainerStarted","Data":"6f91d639a73c89c4bef35530e3553cae46d82fd04f4d5159a503dd42c5f41663"} Mar 13 10:20:25 crc kubenswrapper[4632]: I0313 10:20:25.755303 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-62bwr" Mar 13 10:20:25 crc kubenswrapper[4632]: I0313 10:20:25.755578 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" event={"ID":"b33bccd8-6f28-4ffe-9500-069a52aab5df","Type":"ContainerStarted","Data":"6ccfd3b438da53aa3a9cc24cd575fef5e9f4147fe4893d65dafa89b9bddf7863"} Mar 13 10:20:25 crc kubenswrapper[4632]: I0313 10:20:25.777081 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-tztd9" Mar 13 10:20:25 crc kubenswrapper[4632]: I0313 10:20:25.782708 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-62bwr" podStartSLOduration=2.782685855 podStartE2EDuration="2.782685855s" podCreationTimestamp="2026-03-13 10:20:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:20:25.773510307 +0000 UTC m=+999.796040450" watchObservedRunningTime="2026-03-13 10:20:25.782685855 +0000 UTC m=+999.805216008" Mar 13 10:20:26 crc kubenswrapper[4632]: I0313 10:20:26.792610 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tztd9" event={"ID":"8f51973a-596d-40dc-9b5b-b2c95a60ea0c","Type":"ContainerStarted","Data":"2ea54c19d6a4ac011be67fd99761d87872512667c4d68bb1f20f2ba64f27c6b9"} Mar 13 10:20:26 crc kubenswrapper[4632]: I0313 10:20:26.792908 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tztd9" event={"ID":"8f51973a-596d-40dc-9b5b-b2c95a60ea0c","Type":"ContainerStarted","Data":"7369028ab3380b8162926288f2a66e0780eba331066b6d04106bd606debba692"} Mar 13 10:20:26 crc kubenswrapper[4632]: I0313 10:20:26.792923 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tztd9" event={"ID":"8f51973a-596d-40dc-9b5b-b2c95a60ea0c","Type":"ContainerStarted","Data":"21d4836cf410138123a02393e40bd0af724c4c3a101092ac3a79a03eed782a1d"} Mar 13 10:20:26 crc kubenswrapper[4632]: I0313 10:20:26.793523 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-tztd9" Mar 13 10:20:26 crc kubenswrapper[4632]: I0313 10:20:26.841205 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-tztd9" podStartSLOduration=3.841183327 podStartE2EDuration="3.841183327s" podCreationTimestamp="2026-03-13 10:20:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:20:26.840297755 +0000 UTC m=+1000.862827888" watchObservedRunningTime="2026-03-13 10:20:26.841183327 +0000 UTC m=+1000.863713470" Mar 13 10:20:34 crc kubenswrapper[4632]: I0313 10:20:34.320404 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-62bwr" Mar 13 10:20:36 crc kubenswrapper[4632]: I0313 10:20:36.049394 4632 generic.go:334] "Generic (PLEG): container finished" podID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerID="5d35a7bcb38219d00926d9e32f8bdc06f3388e1497f4e4293b861382b0d06c02" exitCode=0 Mar 13 10:20:36 crc kubenswrapper[4632]: I0313 10:20:36.057247 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 10:20:36 crc kubenswrapper[4632]: I0313 10:20:36.057277 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lvlxj" event={"ID":"85b58bb0-63f5-4c85-8759-ce28d2c7db58","Type":"ContainerDied","Data":"5d35a7bcb38219d00926d9e32f8bdc06f3388e1497f4e4293b861382b0d06c02"} Mar 13 10:20:36 crc kubenswrapper[4632]: I0313 10:20:36.057297 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" event={"ID":"b33bccd8-6f28-4ffe-9500-069a52aab5df","Type":"ContainerStarted","Data":"6b277b3621566e90d2ea8a306394444270adbf026557398f5520284a63c356df"} Mar 13 10:20:36 crc kubenswrapper[4632]: I0313 10:20:36.140381 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" podStartSLOduration=2.474485731 podStartE2EDuration="13.140357057s" podCreationTimestamp="2026-03-13 10:20:23 +0000 UTC" firstStartedPulling="2026-03-13 10:20:25.090932867 +0000 UTC m=+999.113463000" lastFinishedPulling="2026-03-13 10:20:35.756804193 +0000 UTC m=+1009.779334326" observedRunningTime="2026-03-13 10:20:36.118626555 +0000 UTC m=+1010.141156698" watchObservedRunningTime="2026-03-13 10:20:36.140357057 +0000 UTC m=+1010.162887190" Mar 13 10:20:37 crc kubenswrapper[4632]: I0313 10:20:37.064178 4632 generic.go:334] "Generic (PLEG): container finished" podID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerID="7f411b298d721dfc2f203bd0d068c8eb0744299f10d8bf319ffb996f2c67fa11" exitCode=0 Mar 13 10:20:37 crc kubenswrapper[4632]: I0313 10:20:37.064242 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lvlxj" event={"ID":"85b58bb0-63f5-4c85-8759-ce28d2c7db58","Type":"ContainerDied","Data":"7f411b298d721dfc2f203bd0d068c8eb0744299f10d8bf319ffb996f2c67fa11"} Mar 13 10:20:38 crc kubenswrapper[4632]: I0313 10:20:38.071392 4632 generic.go:334] "Generic (PLEG): container finished" podID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerID="eca84017b08d4a3daf0b100d49fcc9129ce1ea0f23816d19bcc8797ce9117ebd" exitCode=0 Mar 13 10:20:38 crc kubenswrapper[4632]: I0313 10:20:38.071438 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lvlxj" event={"ID":"85b58bb0-63f5-4c85-8759-ce28d2c7db58","Type":"ContainerDied","Data":"eca84017b08d4a3daf0b100d49fcc9129ce1ea0f23816d19bcc8797ce9117ebd"} Mar 13 10:20:39 crc kubenswrapper[4632]: I0313 10:20:39.080106 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lvlxj" event={"ID":"85b58bb0-63f5-4c85-8759-ce28d2c7db58","Type":"ContainerStarted","Data":"ac86591594d77be34bb7d30cd47d6e21c76bb21f414a9eee2975a4ab00905070"} Mar 13 10:20:39 crc kubenswrapper[4632]: I0313 10:20:39.080143 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lvlxj" event={"ID":"85b58bb0-63f5-4c85-8759-ce28d2c7db58","Type":"ContainerStarted","Data":"f1bb9b44e61c382d8baef6eea87e50c583aaf39ef98e6660ba4c14110edbbd50"} Mar 13 10:20:39 crc kubenswrapper[4632]: I0313 10:20:39.080154 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lvlxj" event={"ID":"85b58bb0-63f5-4c85-8759-ce28d2c7db58","Type":"ContainerStarted","Data":"7460db0e48987b1f59a273cebeb41489b6ba6de15b26f36b4087497d7870a4f2"} Mar 13 10:20:39 crc kubenswrapper[4632]: I0313 10:20:39.080163 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lvlxj" event={"ID":"85b58bb0-63f5-4c85-8759-ce28d2c7db58","Type":"ContainerStarted","Data":"856e616e36ae078de77d4b4ce66fe7580c0b8b24fdb8aaa99573dc768bc62e8a"} Mar 13 10:20:39 crc kubenswrapper[4632]: I0313 10:20:39.080172 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lvlxj" event={"ID":"85b58bb0-63f5-4c85-8759-ce28d2c7db58","Type":"ContainerStarted","Data":"59d96a138cf7adeb4d273db270ca9998a9b75447d7d6c92e875e751afba3f9b8"} Mar 13 10:20:39 crc kubenswrapper[4632]: I0313 10:20:39.080180 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lvlxj" event={"ID":"85b58bb0-63f5-4c85-8759-ce28d2c7db58","Type":"ContainerStarted","Data":"bfa960455f207db762d901f5af9c2b35ade8cd1c5f43d1bc1d4a40a5bfd8199d"} Mar 13 10:20:39 crc kubenswrapper[4632]: I0313 10:20:39.081301 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:39 crc kubenswrapper[4632]: I0313 10:20:39.109780 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-lvlxj" podStartSLOduration=4.699453806 podStartE2EDuration="16.109758674s" podCreationTimestamp="2026-03-13 10:20:23 +0000 UTC" firstStartedPulling="2026-03-13 10:20:24.326331513 +0000 UTC m=+998.348861656" lastFinishedPulling="2026-03-13 10:20:35.736636391 +0000 UTC m=+1009.759166524" observedRunningTime="2026-03-13 10:20:39.107710682 +0000 UTC m=+1013.130240835" watchObservedRunningTime="2026-03-13 10:20:39.109758674 +0000 UTC m=+1013.132288807" Mar 13 10:20:39 crc kubenswrapper[4632]: I0313 10:20:39.180523 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:39 crc kubenswrapper[4632]: I0313 10:20:39.222777 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:40 crc kubenswrapper[4632]: I0313 10:20:40.461352 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:20:40 crc kubenswrapper[4632]: I0313 10:20:40.461654 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:20:45 crc kubenswrapper[4632]: I0313 10:20:45.783411 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-tztd9" Mar 13 10:20:48 crc kubenswrapper[4632]: I0313 10:20:48.483033 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-2jqnk"] Mar 13 10:20:48 crc kubenswrapper[4632]: I0313 10:20:48.484087 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 10:20:48 crc kubenswrapper[4632]: I0313 10:20:48.487546 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 13 10:20:48 crc kubenswrapper[4632]: I0313 10:20:48.488332 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 13 10:20:48 crc kubenswrapper[4632]: I0313 10:20:48.488608 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-zrjwg" Mar 13 10:20:48 crc kubenswrapper[4632]: I0313 10:20:48.515704 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2jqnk"] Mar 13 10:20:48 crc kubenswrapper[4632]: I0313 10:20:48.606024 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h96l4\" (UniqueName: \"kubernetes.io/projected/7de02b7f-4e1c-4ba1-9659-c864e9080092-kube-api-access-h96l4\") pod \"openstack-operator-index-2jqnk\" (UID: \"7de02b7f-4e1c-4ba1-9659-c864e9080092\") " pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 10:20:48 crc kubenswrapper[4632]: I0313 10:20:48.706813 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h96l4\" (UniqueName: \"kubernetes.io/projected/7de02b7f-4e1c-4ba1-9659-c864e9080092-kube-api-access-h96l4\") pod \"openstack-operator-index-2jqnk\" (UID: \"7de02b7f-4e1c-4ba1-9659-c864e9080092\") " pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 10:20:48 crc kubenswrapper[4632]: I0313 10:20:48.728480 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h96l4\" (UniqueName: \"kubernetes.io/projected/7de02b7f-4e1c-4ba1-9659-c864e9080092-kube-api-access-h96l4\") pod \"openstack-operator-index-2jqnk\" (UID: \"7de02b7f-4e1c-4ba1-9659-c864e9080092\") " pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 10:20:48 crc kubenswrapper[4632]: I0313 10:20:48.811877 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 10:20:49 crc kubenswrapper[4632]: I0313 10:20:49.044323 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-2jqnk"] Mar 13 10:20:49 crc kubenswrapper[4632]: I0313 10:20:49.139985 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2jqnk" event={"ID":"7de02b7f-4e1c-4ba1-9659-c864e9080092","Type":"ContainerStarted","Data":"04de8e81d8d4902a0ce7d3552a9cd405d1dd0a14387d654e4100f584583eff01"} Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.189576 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2jqnk" event={"ID":"7de02b7f-4e1c-4ba1-9659-c864e9080092","Type":"ContainerStarted","Data":"956a137089f814594898450d78be9fb64aa26d046a610f28da0c9756520f90c8"} Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.664260 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-2jqnk" podStartSLOduration=2.481847726 podStartE2EDuration="5.664235343s" podCreationTimestamp="2026-03-13 10:20:48 +0000 UTC" firstStartedPulling="2026-03-13 10:20:49.073227875 +0000 UTC m=+1023.095758028" lastFinishedPulling="2026-03-13 10:20:52.255615512 +0000 UTC m=+1026.278145645" observedRunningTime="2026-03-13 10:20:53.21167938 +0000 UTC m=+1027.234209513" watchObservedRunningTime="2026-03-13 10:20:53.664235343 +0000 UTC m=+1027.686765476" Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.665615 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j2w2m"] Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.667049 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.685542 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j2w2m"] Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.770981 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95e13797-40e9-4942-a7b5-6174fa448654-catalog-content\") pod \"certified-operators-j2w2m\" (UID: \"95e13797-40e9-4942-a7b5-6174fa448654\") " pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.771038 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95e13797-40e9-4942-a7b5-6174fa448654-utilities\") pod \"certified-operators-j2w2m\" (UID: \"95e13797-40e9-4942-a7b5-6174fa448654\") " pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.771217 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-48kk8\" (UniqueName: \"kubernetes.io/projected/95e13797-40e9-4942-a7b5-6174fa448654-kube-api-access-48kk8\") pod \"certified-operators-j2w2m\" (UID: \"95e13797-40e9-4942-a7b5-6174fa448654\") " pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.873778 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-48kk8\" (UniqueName: \"kubernetes.io/projected/95e13797-40e9-4942-a7b5-6174fa448654-kube-api-access-48kk8\") pod \"certified-operators-j2w2m\" (UID: \"95e13797-40e9-4942-a7b5-6174fa448654\") " pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.873860 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95e13797-40e9-4942-a7b5-6174fa448654-catalog-content\") pod \"certified-operators-j2w2m\" (UID: \"95e13797-40e9-4942-a7b5-6174fa448654\") " pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.873883 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95e13797-40e9-4942-a7b5-6174fa448654-utilities\") pod \"certified-operators-j2w2m\" (UID: \"95e13797-40e9-4942-a7b5-6174fa448654\") " pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.874465 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95e13797-40e9-4942-a7b5-6174fa448654-utilities\") pod \"certified-operators-j2w2m\" (UID: \"95e13797-40e9-4942-a7b5-6174fa448654\") " pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.874527 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95e13797-40e9-4942-a7b5-6174fa448654-catalog-content\") pod \"certified-operators-j2w2m\" (UID: \"95e13797-40e9-4942-a7b5-6174fa448654\") " pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.911524 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-48kk8\" (UniqueName: \"kubernetes.io/projected/95e13797-40e9-4942-a7b5-6174fa448654-kube-api-access-48kk8\") pod \"certified-operators-j2w2m\" (UID: \"95e13797-40e9-4942-a7b5-6174fa448654\") " pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:20:53 crc kubenswrapper[4632]: I0313 10:20:53.984457 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:20:54 crc kubenswrapper[4632]: I0313 10:20:54.185047 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-lvlxj" Mar 13 10:20:54 crc kubenswrapper[4632]: I0313 10:20:54.565449 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j2w2m"] Mar 13 10:20:54 crc kubenswrapper[4632]: I0313 10:20:54.810570 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 10:20:55 crc kubenswrapper[4632]: I0313 10:20:55.203417 4632 generic.go:334] "Generic (PLEG): container finished" podID="95e13797-40e9-4942-a7b5-6174fa448654" containerID="6e108feffd870d15032d03dad27338b300eaf2cbdd55be7000c4f4a141b8f3ad" exitCode=0 Mar 13 10:20:55 crc kubenswrapper[4632]: I0313 10:20:55.203452 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2w2m" event={"ID":"95e13797-40e9-4942-a7b5-6174fa448654","Type":"ContainerDied","Data":"6e108feffd870d15032d03dad27338b300eaf2cbdd55be7000c4f4a141b8f3ad"} Mar 13 10:20:55 crc kubenswrapper[4632]: I0313 10:20:55.203476 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2w2m" event={"ID":"95e13797-40e9-4942-a7b5-6174fa448654","Type":"ContainerStarted","Data":"a5d6997e64c1bba55b267a1de22e67ed4e90ac2795eea19781b2283baf2f9305"} Mar 13 10:20:56 crc kubenswrapper[4632]: I0313 10:20:56.211438 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2w2m" event={"ID":"95e13797-40e9-4942-a7b5-6174fa448654","Type":"ContainerStarted","Data":"eff3d4047aa48452ec24bfeecbd51ff4236d75ad1f2c6051aceaf13600d9b721"} Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.218536 4632 generic.go:334] "Generic (PLEG): container finished" podID="95e13797-40e9-4942-a7b5-6174fa448654" containerID="eff3d4047aa48452ec24bfeecbd51ff4236d75ad1f2c6051aceaf13600d9b721" exitCode=0 Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.218585 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2w2m" event={"ID":"95e13797-40e9-4942-a7b5-6174fa448654","Type":"ContainerDied","Data":"eff3d4047aa48452ec24bfeecbd51ff4236d75ad1f2c6051aceaf13600d9b721"} Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.667901 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fdtl7"] Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.669887 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.679822 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fdtl7"] Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.729231 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-catalog-content\") pod \"community-operators-fdtl7\" (UID: \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\") " pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.729327 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-utilities\") pod \"community-operators-fdtl7\" (UID: \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\") " pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.729408 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgb8q\" (UniqueName: \"kubernetes.io/projected/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-kube-api-access-sgb8q\") pod \"community-operators-fdtl7\" (UID: \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\") " pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.830177 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-catalog-content\") pod \"community-operators-fdtl7\" (UID: \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\") " pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.832107 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-utilities\") pod \"community-operators-fdtl7\" (UID: \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\") " pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.832157 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgb8q\" (UniqueName: \"kubernetes.io/projected/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-kube-api-access-sgb8q\") pod \"community-operators-fdtl7\" (UID: \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\") " pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.831146 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-catalog-content\") pod \"community-operators-fdtl7\" (UID: \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\") " pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.832784 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-utilities\") pod \"community-operators-fdtl7\" (UID: \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\") " pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.860740 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgb8q\" (UniqueName: \"kubernetes.io/projected/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-kube-api-access-sgb8q\") pod \"community-operators-fdtl7\" (UID: \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\") " pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.966686 4632 scope.go:117] "RemoveContainer" containerID="fcf5d9f69f7435b287086bfcb908c42e9330ebc2ef407226d11b60f145efd8de" Mar 13 10:20:57 crc kubenswrapper[4632]: I0313 10:20:57.987716 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:20:58 crc kubenswrapper[4632]: I0313 10:20:58.254281 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2w2m" event={"ID":"95e13797-40e9-4942-a7b5-6174fa448654","Type":"ContainerStarted","Data":"f3b5913884cb0cd302a4b84c7cdf0903689fa6580713d76d4c3ad0f4d72eb034"} Mar 13 10:20:58 crc kubenswrapper[4632]: I0313 10:20:58.293068 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j2w2m" podStartSLOduration=2.758044938 podStartE2EDuration="5.293049504s" podCreationTimestamp="2026-03-13 10:20:53 +0000 UTC" firstStartedPulling="2026-03-13 10:20:55.205126562 +0000 UTC m=+1029.227656705" lastFinishedPulling="2026-03-13 10:20:57.740131138 +0000 UTC m=+1031.762661271" observedRunningTime="2026-03-13 10:20:58.292338696 +0000 UTC m=+1032.314868829" watchObservedRunningTime="2026-03-13 10:20:58.293049504 +0000 UTC m=+1032.315579637" Mar 13 10:20:58 crc kubenswrapper[4632]: I0313 10:20:58.721788 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fdtl7"] Mar 13 10:20:58 crc kubenswrapper[4632]: I0313 10:20:58.812921 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 10:20:58 crc kubenswrapper[4632]: I0313 10:20:58.812983 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 10:20:58 crc kubenswrapper[4632]: I0313 10:20:58.850853 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 10:20:59 crc kubenswrapper[4632]: I0313 10:20:59.261499 4632 generic.go:334] "Generic (PLEG): container finished" podID="a01bcaf0-e2c1-495b-bc6d-a57978c7817b" containerID="e0b81f7099d2b6e974290907876a0f00faa065b0c428d84017f9ee0db229bc73" exitCode=0 Mar 13 10:20:59 crc kubenswrapper[4632]: I0313 10:20:59.261564 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdtl7" event={"ID":"a01bcaf0-e2c1-495b-bc6d-a57978c7817b","Type":"ContainerDied","Data":"e0b81f7099d2b6e974290907876a0f00faa065b0c428d84017f9ee0db229bc73"} Mar 13 10:20:59 crc kubenswrapper[4632]: I0313 10:20:59.261605 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdtl7" event={"ID":"a01bcaf0-e2c1-495b-bc6d-a57978c7817b","Type":"ContainerStarted","Data":"b009a6765de2ecd804f8d033b53e49abb32cd27ba05ed7eacca8b430a75a2575"} Mar 13 10:20:59 crc kubenswrapper[4632]: I0313 10:20:59.295881 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 10:21:01 crc kubenswrapper[4632]: I0313 10:21:01.463434 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x5lbm"] Mar 13 10:21:01 crc kubenswrapper[4632]: I0313 10:21:01.465251 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:01 crc kubenswrapper[4632]: I0313 10:21:01.477691 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5lbm"] Mar 13 10:21:01 crc kubenswrapper[4632]: I0313 10:21:01.587739 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d96ff75-88fd-4637-9199-806314e5276d-utilities\") pod \"redhat-marketplace-x5lbm\" (UID: \"2d96ff75-88fd-4637-9199-806314e5276d\") " pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:01 crc kubenswrapper[4632]: I0313 10:21:01.587823 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7vjg\" (UniqueName: \"kubernetes.io/projected/2d96ff75-88fd-4637-9199-806314e5276d-kube-api-access-c7vjg\") pod \"redhat-marketplace-x5lbm\" (UID: \"2d96ff75-88fd-4637-9199-806314e5276d\") " pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:01 crc kubenswrapper[4632]: I0313 10:21:01.587872 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d96ff75-88fd-4637-9199-806314e5276d-catalog-content\") pod \"redhat-marketplace-x5lbm\" (UID: \"2d96ff75-88fd-4637-9199-806314e5276d\") " pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:01 crc kubenswrapper[4632]: I0313 10:21:01.689369 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7vjg\" (UniqueName: \"kubernetes.io/projected/2d96ff75-88fd-4637-9199-806314e5276d-kube-api-access-c7vjg\") pod \"redhat-marketplace-x5lbm\" (UID: \"2d96ff75-88fd-4637-9199-806314e5276d\") " pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:01 crc kubenswrapper[4632]: I0313 10:21:01.689466 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d96ff75-88fd-4637-9199-806314e5276d-catalog-content\") pod \"redhat-marketplace-x5lbm\" (UID: \"2d96ff75-88fd-4637-9199-806314e5276d\") " pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:01 crc kubenswrapper[4632]: I0313 10:21:01.689519 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d96ff75-88fd-4637-9199-806314e5276d-utilities\") pod \"redhat-marketplace-x5lbm\" (UID: \"2d96ff75-88fd-4637-9199-806314e5276d\") " pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:01 crc kubenswrapper[4632]: I0313 10:21:01.690226 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d96ff75-88fd-4637-9199-806314e5276d-catalog-content\") pod \"redhat-marketplace-x5lbm\" (UID: \"2d96ff75-88fd-4637-9199-806314e5276d\") " pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:01 crc kubenswrapper[4632]: I0313 10:21:01.690249 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d96ff75-88fd-4637-9199-806314e5276d-utilities\") pod \"redhat-marketplace-x5lbm\" (UID: \"2d96ff75-88fd-4637-9199-806314e5276d\") " pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:01 crc kubenswrapper[4632]: I0313 10:21:01.711654 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7vjg\" (UniqueName: \"kubernetes.io/projected/2d96ff75-88fd-4637-9199-806314e5276d-kube-api-access-c7vjg\") pod \"redhat-marketplace-x5lbm\" (UID: \"2d96ff75-88fd-4637-9199-806314e5276d\") " pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:01 crc kubenswrapper[4632]: I0313 10:21:01.791213 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:03 crc kubenswrapper[4632]: I0313 10:21:03.985641 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:21:03 crc kubenswrapper[4632]: I0313 10:21:03.986021 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:21:05 crc kubenswrapper[4632]: I0313 10:21:05.032930 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-j2w2m" podUID="95e13797-40e9-4942-a7b5-6174fa448654" containerName="registry-server" probeResult="failure" output=< Mar 13 10:21:05 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:21:05 crc kubenswrapper[4632]: > Mar 13 10:21:05 crc kubenswrapper[4632]: I0313 10:21:05.290816 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5lbm"] Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.308353 4632 generic.go:334] "Generic (PLEG): container finished" podID="2d96ff75-88fd-4637-9199-806314e5276d" containerID="9d09e614abc972c9bc41b522cbf719729037565923e2784a67b05584e2614a0c" exitCode=0 Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.308451 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5lbm" event={"ID":"2d96ff75-88fd-4637-9199-806314e5276d","Type":"ContainerDied","Data":"9d09e614abc972c9bc41b522cbf719729037565923e2784a67b05584e2614a0c"} Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.308750 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5lbm" event={"ID":"2d96ff75-88fd-4637-9199-806314e5276d","Type":"ContainerStarted","Data":"21e3226d1d96fc021bc557ff5418b6dec4ea17fe511f69b4fe8609410d428008"} Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.311925 4632 generic.go:334] "Generic (PLEG): container finished" podID="a01bcaf0-e2c1-495b-bc6d-a57978c7817b" containerID="90d2b05d55e78b9f9829c2ee4bf7bbc01510b17dfbff9b47dda76cd10b610010" exitCode=0 Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.311978 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdtl7" event={"ID":"a01bcaf0-e2c1-495b-bc6d-a57978c7817b","Type":"ContainerDied","Data":"90d2b05d55e78b9f9829c2ee4bf7bbc01510b17dfbff9b47dda76cd10b610010"} Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.385865 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m"] Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.387018 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.394862 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-wbvdr" Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.410984 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m"] Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.476125 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrkd5\" (UniqueName: \"kubernetes.io/projected/13abf84a-b499-4439-ab4e-1c34bcf07308-kube-api-access-wrkd5\") pod \"cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m\" (UID: \"13abf84a-b499-4439-ab4e-1c34bcf07308\") " pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.476232 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/13abf84a-b499-4439-ab4e-1c34bcf07308-util\") pod \"cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m\" (UID: \"13abf84a-b499-4439-ab4e-1c34bcf07308\") " pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.476293 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/13abf84a-b499-4439-ab4e-1c34bcf07308-bundle\") pod \"cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m\" (UID: \"13abf84a-b499-4439-ab4e-1c34bcf07308\") " pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.577132 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrkd5\" (UniqueName: \"kubernetes.io/projected/13abf84a-b499-4439-ab4e-1c34bcf07308-kube-api-access-wrkd5\") pod \"cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m\" (UID: \"13abf84a-b499-4439-ab4e-1c34bcf07308\") " pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.577238 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/13abf84a-b499-4439-ab4e-1c34bcf07308-util\") pod \"cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m\" (UID: \"13abf84a-b499-4439-ab4e-1c34bcf07308\") " pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.577295 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/13abf84a-b499-4439-ab4e-1c34bcf07308-bundle\") pod \"cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m\" (UID: \"13abf84a-b499-4439-ab4e-1c34bcf07308\") " pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.577994 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/13abf84a-b499-4439-ab4e-1c34bcf07308-bundle\") pod \"cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m\" (UID: \"13abf84a-b499-4439-ab4e-1c34bcf07308\") " pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.578591 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/13abf84a-b499-4439-ab4e-1c34bcf07308-util\") pod \"cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m\" (UID: \"13abf84a-b499-4439-ab4e-1c34bcf07308\") " pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.604197 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrkd5\" (UniqueName: \"kubernetes.io/projected/13abf84a-b499-4439-ab4e-1c34bcf07308-kube-api-access-wrkd5\") pod \"cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m\" (UID: \"13abf84a-b499-4439-ab4e-1c34bcf07308\") " pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.701304 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" Mar 13 10:21:06 crc kubenswrapper[4632]: I0313 10:21:06.997758 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m"] Mar 13 10:21:07 crc kubenswrapper[4632]: W0313 10:21:07.026954 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod13abf84a_b499_4439_ab4e_1c34bcf07308.slice/crio-c2b7069b8c86dedbcbc82d2f9138beb8a86a6fd02f2f8774925b78243b5b613e WatchSource:0}: Error finding container c2b7069b8c86dedbcbc82d2f9138beb8a86a6fd02f2f8774925b78243b5b613e: Status 404 returned error can't find the container with id c2b7069b8c86dedbcbc82d2f9138beb8a86a6fd02f2f8774925b78243b5b613e Mar 13 10:21:07 crc kubenswrapper[4632]: I0313 10:21:07.321604 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5lbm" event={"ID":"2d96ff75-88fd-4637-9199-806314e5276d","Type":"ContainerStarted","Data":"559890583b5bdac795810d916d2ba129170ca28c94f2aae83a30fbc62b754214"} Mar 13 10:21:07 crc kubenswrapper[4632]: I0313 10:21:07.324977 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdtl7" event={"ID":"a01bcaf0-e2c1-495b-bc6d-a57978c7817b","Type":"ContainerStarted","Data":"e2fda5f5d13b5663527978c3fcfd3cccd016f3f222280313a4a0c5c88b5212d7"} Mar 13 10:21:07 crc kubenswrapper[4632]: I0313 10:21:07.326779 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" event={"ID":"13abf84a-b499-4439-ab4e-1c34bcf07308","Type":"ContainerStarted","Data":"7318b74c846ee470b6f66c429df22fea7568a67857fa4aca7afcbc6ac2e7ef17"} Mar 13 10:21:07 crc kubenswrapper[4632]: I0313 10:21:07.326901 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" event={"ID":"13abf84a-b499-4439-ab4e-1c34bcf07308","Type":"ContainerStarted","Data":"c2b7069b8c86dedbcbc82d2f9138beb8a86a6fd02f2f8774925b78243b5b613e"} Mar 13 10:21:07 crc kubenswrapper[4632]: I0313 10:21:07.388426 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fdtl7" podStartSLOduration=2.7759761210000002 podStartE2EDuration="10.388405803s" podCreationTimestamp="2026-03-13 10:20:57 +0000 UTC" firstStartedPulling="2026-03-13 10:20:59.263106421 +0000 UTC m=+1033.285636554" lastFinishedPulling="2026-03-13 10:21:06.875536103 +0000 UTC m=+1040.898066236" observedRunningTime="2026-03-13 10:21:07.383778523 +0000 UTC m=+1041.406308666" watchObservedRunningTime="2026-03-13 10:21:07.388405803 +0000 UTC m=+1041.410935936" Mar 13 10:21:07 crc kubenswrapper[4632]: I0313 10:21:07.988421 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:21:07 crc kubenswrapper[4632]: I0313 10:21:07.988589 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:21:08 crc kubenswrapper[4632]: I0313 10:21:08.335126 4632 generic.go:334] "Generic (PLEG): container finished" podID="2d96ff75-88fd-4637-9199-806314e5276d" containerID="559890583b5bdac795810d916d2ba129170ca28c94f2aae83a30fbc62b754214" exitCode=0 Mar 13 10:21:08 crc kubenswrapper[4632]: I0313 10:21:08.335203 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5lbm" event={"ID":"2d96ff75-88fd-4637-9199-806314e5276d","Type":"ContainerDied","Data":"559890583b5bdac795810d916d2ba129170ca28c94f2aae83a30fbc62b754214"} Mar 13 10:21:09 crc kubenswrapper[4632]: I0313 10:21:09.039818 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-fdtl7" podUID="a01bcaf0-e2c1-495b-bc6d-a57978c7817b" containerName="registry-server" probeResult="failure" output=< Mar 13 10:21:09 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:21:09 crc kubenswrapper[4632]: > Mar 13 10:21:09 crc kubenswrapper[4632]: I0313 10:21:09.345208 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5lbm" event={"ID":"2d96ff75-88fd-4637-9199-806314e5276d","Type":"ContainerStarted","Data":"db96b68191a6b558bd0777d22d1fa122614ae61dec4e51a4b871ee53f8b057b3"} Mar 13 10:21:09 crc kubenswrapper[4632]: I0313 10:21:09.347108 4632 generic.go:334] "Generic (PLEG): container finished" podID="13abf84a-b499-4439-ab4e-1c34bcf07308" containerID="7318b74c846ee470b6f66c429df22fea7568a67857fa4aca7afcbc6ac2e7ef17" exitCode=0 Mar 13 10:21:09 crc kubenswrapper[4632]: I0313 10:21:09.347601 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" event={"ID":"13abf84a-b499-4439-ab4e-1c34bcf07308","Type":"ContainerDied","Data":"7318b74c846ee470b6f66c429df22fea7568a67857fa4aca7afcbc6ac2e7ef17"} Mar 13 10:21:09 crc kubenswrapper[4632]: I0313 10:21:09.370714 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x5lbm" podStartSLOduration=5.626338865 podStartE2EDuration="8.370698194s" podCreationTimestamp="2026-03-13 10:21:01 +0000 UTC" firstStartedPulling="2026-03-13 10:21:06.311073387 +0000 UTC m=+1040.333603530" lastFinishedPulling="2026-03-13 10:21:09.055432726 +0000 UTC m=+1043.077962859" observedRunningTime="2026-03-13 10:21:09.367309422 +0000 UTC m=+1043.389839555" watchObservedRunningTime="2026-03-13 10:21:09.370698194 +0000 UTC m=+1043.393228327" Mar 13 10:21:10 crc kubenswrapper[4632]: I0313 10:21:10.356063 4632 generic.go:334] "Generic (PLEG): container finished" podID="13abf84a-b499-4439-ab4e-1c34bcf07308" containerID="45880f5326c1f4d4af7e9ddd4f6f22fd9085395311e7514541dd5ff4a8ffb09d" exitCode=0 Mar 13 10:21:10 crc kubenswrapper[4632]: I0313 10:21:10.356134 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" event={"ID":"13abf84a-b499-4439-ab4e-1c34bcf07308","Type":"ContainerDied","Data":"45880f5326c1f4d4af7e9ddd4f6f22fd9085395311e7514541dd5ff4a8ffb09d"} Mar 13 10:21:10 crc kubenswrapper[4632]: I0313 10:21:10.460754 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:21:10 crc kubenswrapper[4632]: I0313 10:21:10.460853 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:21:11 crc kubenswrapper[4632]: I0313 10:21:11.364654 4632 generic.go:334] "Generic (PLEG): container finished" podID="13abf84a-b499-4439-ab4e-1c34bcf07308" containerID="6116cafc17e787a40f5b468c06cdc737d5e7d0fe51a208dd653b4773ccf5ac26" exitCode=0 Mar 13 10:21:11 crc kubenswrapper[4632]: I0313 10:21:11.364743 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" event={"ID":"13abf84a-b499-4439-ab4e-1c34bcf07308","Type":"ContainerDied","Data":"6116cafc17e787a40f5b468c06cdc737d5e7d0fe51a208dd653b4773ccf5ac26"} Mar 13 10:21:11 crc kubenswrapper[4632]: I0313 10:21:11.792185 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:11 crc kubenswrapper[4632]: I0313 10:21:11.794006 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:11 crc kubenswrapper[4632]: I0313 10:21:11.834257 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:12 crc kubenswrapper[4632]: I0313 10:21:12.639723 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" Mar 13 10:21:12 crc kubenswrapper[4632]: I0313 10:21:12.782102 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/13abf84a-b499-4439-ab4e-1c34bcf07308-bundle\") pod \"13abf84a-b499-4439-ab4e-1c34bcf07308\" (UID: \"13abf84a-b499-4439-ab4e-1c34bcf07308\") " Mar 13 10:21:12 crc kubenswrapper[4632]: I0313 10:21:12.782255 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrkd5\" (UniqueName: \"kubernetes.io/projected/13abf84a-b499-4439-ab4e-1c34bcf07308-kube-api-access-wrkd5\") pod \"13abf84a-b499-4439-ab4e-1c34bcf07308\" (UID: \"13abf84a-b499-4439-ab4e-1c34bcf07308\") " Mar 13 10:21:12 crc kubenswrapper[4632]: I0313 10:21:12.782288 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/13abf84a-b499-4439-ab4e-1c34bcf07308-util\") pod \"13abf84a-b499-4439-ab4e-1c34bcf07308\" (UID: \"13abf84a-b499-4439-ab4e-1c34bcf07308\") " Mar 13 10:21:12 crc kubenswrapper[4632]: I0313 10:21:12.782923 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13abf84a-b499-4439-ab4e-1c34bcf07308-bundle" (OuterVolumeSpecName: "bundle") pod "13abf84a-b499-4439-ab4e-1c34bcf07308" (UID: "13abf84a-b499-4439-ab4e-1c34bcf07308"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:21:12 crc kubenswrapper[4632]: I0313 10:21:12.794084 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/13abf84a-b499-4439-ab4e-1c34bcf07308-util" (OuterVolumeSpecName: "util") pod "13abf84a-b499-4439-ab4e-1c34bcf07308" (UID: "13abf84a-b499-4439-ab4e-1c34bcf07308"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:21:12 crc kubenswrapper[4632]: I0313 10:21:12.795279 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/13abf84a-b499-4439-ab4e-1c34bcf07308-kube-api-access-wrkd5" (OuterVolumeSpecName: "kube-api-access-wrkd5") pod "13abf84a-b499-4439-ab4e-1c34bcf07308" (UID: "13abf84a-b499-4439-ab4e-1c34bcf07308"). InnerVolumeSpecName "kube-api-access-wrkd5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:21:12 crc kubenswrapper[4632]: I0313 10:21:12.884736 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrkd5\" (UniqueName: \"kubernetes.io/projected/13abf84a-b499-4439-ab4e-1c34bcf07308-kube-api-access-wrkd5\") on node \"crc\" DevicePath \"\"" Mar 13 10:21:12 crc kubenswrapper[4632]: I0313 10:21:12.884776 4632 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/13abf84a-b499-4439-ab4e-1c34bcf07308-util\") on node \"crc\" DevicePath \"\"" Mar 13 10:21:12 crc kubenswrapper[4632]: I0313 10:21:12.884787 4632 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/13abf84a-b499-4439-ab4e-1c34bcf07308-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:21:13 crc kubenswrapper[4632]: I0313 10:21:13.382170 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" event={"ID":"13abf84a-b499-4439-ab4e-1c34bcf07308","Type":"ContainerDied","Data":"c2b7069b8c86dedbcbc82d2f9138beb8a86a6fd02f2f8774925b78243b5b613e"} Mar 13 10:21:13 crc kubenswrapper[4632]: I0313 10:21:13.382239 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c2b7069b8c86dedbcbc82d2f9138beb8a86a6fd02f2f8774925b78243b5b613e" Mar 13 10:21:13 crc kubenswrapper[4632]: I0313 10:21:13.382203 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m" Mar 13 10:21:14 crc kubenswrapper[4632]: I0313 10:21:14.023772 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:21:14 crc kubenswrapper[4632]: I0313 10:21:14.064640 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:21:17 crc kubenswrapper[4632]: I0313 10:21:17.502777 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq"] Mar 13 10:21:17 crc kubenswrapper[4632]: E0313 10:21:17.503120 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13abf84a-b499-4439-ab4e-1c34bcf07308" containerName="pull" Mar 13 10:21:17 crc kubenswrapper[4632]: I0313 10:21:17.503149 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="13abf84a-b499-4439-ab4e-1c34bcf07308" containerName="pull" Mar 13 10:21:17 crc kubenswrapper[4632]: E0313 10:21:17.503172 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13abf84a-b499-4439-ab4e-1c34bcf07308" containerName="extract" Mar 13 10:21:17 crc kubenswrapper[4632]: I0313 10:21:17.503182 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="13abf84a-b499-4439-ab4e-1c34bcf07308" containerName="extract" Mar 13 10:21:17 crc kubenswrapper[4632]: E0313 10:21:17.503194 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="13abf84a-b499-4439-ab4e-1c34bcf07308" containerName="util" Mar 13 10:21:17 crc kubenswrapper[4632]: I0313 10:21:17.503203 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="13abf84a-b499-4439-ab4e-1c34bcf07308" containerName="util" Mar 13 10:21:17 crc kubenswrapper[4632]: I0313 10:21:17.503334 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="13abf84a-b499-4439-ab4e-1c34bcf07308" containerName="extract" Mar 13 10:21:17 crc kubenswrapper[4632]: I0313 10:21:17.503827 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq" Mar 13 10:21:17 crc kubenswrapper[4632]: I0313 10:21:17.508499 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-7gsqz" Mar 13 10:21:17 crc kubenswrapper[4632]: I0313 10:21:17.547096 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2npzm\" (UniqueName: \"kubernetes.io/projected/82fe7ef6-50a5-41d4-9419-787812e16bd6-kube-api-access-2npzm\") pod \"openstack-operator-controller-init-865685cd99-ls9jq\" (UID: \"82fe7ef6-50a5-41d4-9419-787812e16bd6\") " pod="openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq" Mar 13 10:21:17 crc kubenswrapper[4632]: I0313 10:21:17.567191 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq"] Mar 13 10:21:17 crc kubenswrapper[4632]: I0313 10:21:17.648225 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2npzm\" (UniqueName: \"kubernetes.io/projected/82fe7ef6-50a5-41d4-9419-787812e16bd6-kube-api-access-2npzm\") pod \"openstack-operator-controller-init-865685cd99-ls9jq\" (UID: \"82fe7ef6-50a5-41d4-9419-787812e16bd6\") " pod="openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq" Mar 13 10:21:17 crc kubenswrapper[4632]: I0313 10:21:17.671201 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2npzm\" (UniqueName: \"kubernetes.io/projected/82fe7ef6-50a5-41d4-9419-787812e16bd6-kube-api-access-2npzm\") pod \"openstack-operator-controller-init-865685cd99-ls9jq\" (UID: \"82fe7ef6-50a5-41d4-9419-787812e16bd6\") " pod="openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq" Mar 13 10:21:17 crc kubenswrapper[4632]: I0313 10:21:17.857314 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq" Mar 13 10:21:17 crc kubenswrapper[4632]: I0313 10:21:17.857894 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j2w2m"] Mar 13 10:21:17 crc kubenswrapper[4632]: I0313 10:21:17.858203 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j2w2m" podUID="95e13797-40e9-4942-a7b5-6174fa448654" containerName="registry-server" containerID="cri-o://f3b5913884cb0cd302a4b84c7cdf0903689fa6580713d76d4c3ad0f4d72eb034" gracePeriod=2 Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.073714 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.161802 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.246823 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq"] Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.365448 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.419665 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq" event={"ID":"82fe7ef6-50a5-41d4-9419-787812e16bd6","Type":"ContainerStarted","Data":"84d12dff21de88c28f1e03859a2f3517c9de7ca42f5ece138cc13f2a26293e57"} Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.424853 4632 generic.go:334] "Generic (PLEG): container finished" podID="95e13797-40e9-4942-a7b5-6174fa448654" containerID="f3b5913884cb0cd302a4b84c7cdf0903689fa6580713d76d4c3ad0f4d72eb034" exitCode=0 Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.425680 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j2w2m" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.425972 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2w2m" event={"ID":"95e13797-40e9-4942-a7b5-6174fa448654","Type":"ContainerDied","Data":"f3b5913884cb0cd302a4b84c7cdf0903689fa6580713d76d4c3ad0f4d72eb034"} Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.426001 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j2w2m" event={"ID":"95e13797-40e9-4942-a7b5-6174fa448654","Type":"ContainerDied","Data":"a5d6997e64c1bba55b267a1de22e67ed4e90ac2795eea19781b2283baf2f9305"} Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.426017 4632 scope.go:117] "RemoveContainer" containerID="f3b5913884cb0cd302a4b84c7cdf0903689fa6580713d76d4c3ad0f4d72eb034" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.453971 4632 scope.go:117] "RemoveContainer" containerID="eff3d4047aa48452ec24bfeecbd51ff4236d75ad1f2c6051aceaf13600d9b721" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.461438 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-48kk8\" (UniqueName: \"kubernetes.io/projected/95e13797-40e9-4942-a7b5-6174fa448654-kube-api-access-48kk8\") pod \"95e13797-40e9-4942-a7b5-6174fa448654\" (UID: \"95e13797-40e9-4942-a7b5-6174fa448654\") " Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.467059 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95e13797-40e9-4942-a7b5-6174fa448654-kube-api-access-48kk8" (OuterVolumeSpecName: "kube-api-access-48kk8") pod "95e13797-40e9-4942-a7b5-6174fa448654" (UID: "95e13797-40e9-4942-a7b5-6174fa448654"). InnerVolumeSpecName "kube-api-access-48kk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.478109 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95e13797-40e9-4942-a7b5-6174fa448654-catalog-content\") pod \"95e13797-40e9-4942-a7b5-6174fa448654\" (UID: \"95e13797-40e9-4942-a7b5-6174fa448654\") " Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.478222 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95e13797-40e9-4942-a7b5-6174fa448654-utilities\") pod \"95e13797-40e9-4942-a7b5-6174fa448654\" (UID: \"95e13797-40e9-4942-a7b5-6174fa448654\") " Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.478826 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-48kk8\" (UniqueName: \"kubernetes.io/projected/95e13797-40e9-4942-a7b5-6174fa448654-kube-api-access-48kk8\") on node \"crc\" DevicePath \"\"" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.479242 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95e13797-40e9-4942-a7b5-6174fa448654-utilities" (OuterVolumeSpecName: "utilities") pod "95e13797-40e9-4942-a7b5-6174fa448654" (UID: "95e13797-40e9-4942-a7b5-6174fa448654"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.492987 4632 scope.go:117] "RemoveContainer" containerID="6e108feffd870d15032d03dad27338b300eaf2cbdd55be7000c4f4a141b8f3ad" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.515082 4632 scope.go:117] "RemoveContainer" containerID="f3b5913884cb0cd302a4b84c7cdf0903689fa6580713d76d4c3ad0f4d72eb034" Mar 13 10:21:18 crc kubenswrapper[4632]: E0313 10:21:18.515611 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3b5913884cb0cd302a4b84c7cdf0903689fa6580713d76d4c3ad0f4d72eb034\": container with ID starting with f3b5913884cb0cd302a4b84c7cdf0903689fa6580713d76d4c3ad0f4d72eb034 not found: ID does not exist" containerID="f3b5913884cb0cd302a4b84c7cdf0903689fa6580713d76d4c3ad0f4d72eb034" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.515658 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3b5913884cb0cd302a4b84c7cdf0903689fa6580713d76d4c3ad0f4d72eb034"} err="failed to get container status \"f3b5913884cb0cd302a4b84c7cdf0903689fa6580713d76d4c3ad0f4d72eb034\": rpc error: code = NotFound desc = could not find container \"f3b5913884cb0cd302a4b84c7cdf0903689fa6580713d76d4c3ad0f4d72eb034\": container with ID starting with f3b5913884cb0cd302a4b84c7cdf0903689fa6580713d76d4c3ad0f4d72eb034 not found: ID does not exist" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.515690 4632 scope.go:117] "RemoveContainer" containerID="eff3d4047aa48452ec24bfeecbd51ff4236d75ad1f2c6051aceaf13600d9b721" Mar 13 10:21:18 crc kubenswrapper[4632]: E0313 10:21:18.516166 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eff3d4047aa48452ec24bfeecbd51ff4236d75ad1f2c6051aceaf13600d9b721\": container with ID starting with eff3d4047aa48452ec24bfeecbd51ff4236d75ad1f2c6051aceaf13600d9b721 not found: ID does not exist" containerID="eff3d4047aa48452ec24bfeecbd51ff4236d75ad1f2c6051aceaf13600d9b721" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.516189 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eff3d4047aa48452ec24bfeecbd51ff4236d75ad1f2c6051aceaf13600d9b721"} err="failed to get container status \"eff3d4047aa48452ec24bfeecbd51ff4236d75ad1f2c6051aceaf13600d9b721\": rpc error: code = NotFound desc = could not find container \"eff3d4047aa48452ec24bfeecbd51ff4236d75ad1f2c6051aceaf13600d9b721\": container with ID starting with eff3d4047aa48452ec24bfeecbd51ff4236d75ad1f2c6051aceaf13600d9b721 not found: ID does not exist" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.516255 4632 scope.go:117] "RemoveContainer" containerID="6e108feffd870d15032d03dad27338b300eaf2cbdd55be7000c4f4a141b8f3ad" Mar 13 10:21:18 crc kubenswrapper[4632]: E0313 10:21:18.516466 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e108feffd870d15032d03dad27338b300eaf2cbdd55be7000c4f4a141b8f3ad\": container with ID starting with 6e108feffd870d15032d03dad27338b300eaf2cbdd55be7000c4f4a141b8f3ad not found: ID does not exist" containerID="6e108feffd870d15032d03dad27338b300eaf2cbdd55be7000c4f4a141b8f3ad" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.516488 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e108feffd870d15032d03dad27338b300eaf2cbdd55be7000c4f4a141b8f3ad"} err="failed to get container status \"6e108feffd870d15032d03dad27338b300eaf2cbdd55be7000c4f4a141b8f3ad\": rpc error: code = NotFound desc = could not find container \"6e108feffd870d15032d03dad27338b300eaf2cbdd55be7000c4f4a141b8f3ad\": container with ID starting with 6e108feffd870d15032d03dad27338b300eaf2cbdd55be7000c4f4a141b8f3ad not found: ID does not exist" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.538974 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95e13797-40e9-4942-a7b5-6174fa448654-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "95e13797-40e9-4942-a7b5-6174fa448654" (UID: "95e13797-40e9-4942-a7b5-6174fa448654"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.583655 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/95e13797-40e9-4942-a7b5-6174fa448654-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.583694 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/95e13797-40e9-4942-a7b5-6174fa448654-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.774284 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j2w2m"] Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.785354 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j2w2m"] Mar 13 10:21:18 crc kubenswrapper[4632]: I0313 10:21:18.902764 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fdtl7"] Mar 13 10:21:19 crc kubenswrapper[4632]: I0313 10:21:19.257411 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lgwff"] Mar 13 10:21:19 crc kubenswrapper[4632]: I0313 10:21:19.257690 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lgwff" podUID="0d0fc567-0682-4bbc-981b-b4d1df62aa4e" containerName="registry-server" containerID="cri-o://6fa56a0ef2065ba4287ddc46b227aad0c8d55e685aeeb7889682c05acb775492" gracePeriod=2 Mar 13 10:21:19 crc kubenswrapper[4632]: I0313 10:21:19.450033 4632 generic.go:334] "Generic (PLEG): container finished" podID="0d0fc567-0682-4bbc-981b-b4d1df62aa4e" containerID="6fa56a0ef2065ba4287ddc46b227aad0c8d55e685aeeb7889682c05acb775492" exitCode=0 Mar 13 10:21:19 crc kubenswrapper[4632]: I0313 10:21:19.450093 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lgwff" event={"ID":"0d0fc567-0682-4bbc-981b-b4d1df62aa4e","Type":"ContainerDied","Data":"6fa56a0ef2065ba4287ddc46b227aad0c8d55e685aeeb7889682c05acb775492"} Mar 13 10:21:19 crc kubenswrapper[4632]: I0313 10:21:19.942892 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.002792 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sw5gz\" (UniqueName: \"kubernetes.io/projected/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-kube-api-access-sw5gz\") pod \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\" (UID: \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\") " Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.002921 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-catalog-content\") pod \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\" (UID: \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\") " Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.003023 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-utilities\") pod \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\" (UID: \"0d0fc567-0682-4bbc-981b-b4d1df62aa4e\") " Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.003464 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-utilities" (OuterVolumeSpecName: "utilities") pod "0d0fc567-0682-4bbc-981b-b4d1df62aa4e" (UID: "0d0fc567-0682-4bbc-981b-b4d1df62aa4e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.006320 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-kube-api-access-sw5gz" (OuterVolumeSpecName: "kube-api-access-sw5gz") pod "0d0fc567-0682-4bbc-981b-b4d1df62aa4e" (UID: "0d0fc567-0682-4bbc-981b-b4d1df62aa4e"). InnerVolumeSpecName "kube-api-access-sw5gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.069049 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95e13797-40e9-4942-a7b5-6174fa448654" path="/var/lib/kubelet/pods/95e13797-40e9-4942-a7b5-6174fa448654/volumes" Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.108564 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sw5gz\" (UniqueName: \"kubernetes.io/projected/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-kube-api-access-sw5gz\") on node \"crc\" DevicePath \"\"" Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.108600 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.109392 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0d0fc567-0682-4bbc-981b-b4d1df62aa4e" (UID: "0d0fc567-0682-4bbc-981b-b4d1df62aa4e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.209714 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0d0fc567-0682-4bbc-981b-b4d1df62aa4e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.472257 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lgwff" event={"ID":"0d0fc567-0682-4bbc-981b-b4d1df62aa4e","Type":"ContainerDied","Data":"4266d075e4d04ea92ccfdc02ec4b3551e54779fe4f2f2c386ac2a209fda18404"} Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.472319 4632 scope.go:117] "RemoveContainer" containerID="6fa56a0ef2065ba4287ddc46b227aad0c8d55e685aeeb7889682c05acb775492" Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.472448 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lgwff" Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.587100 4632 scope.go:117] "RemoveContainer" containerID="6ed6c1b1b2793ab4b788dc1723932bf9c4121a7bf0945a697809d4c945eec749" Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.592400 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lgwff"] Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.623667 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lgwff"] Mar 13 10:21:20 crc kubenswrapper[4632]: I0313 10:21:20.638086 4632 scope.go:117] "RemoveContainer" containerID="647f33468b0e454866917d9beec3f31ad6bc8dca469daccdfcf7e8df5de24312" Mar 13 10:21:21 crc kubenswrapper[4632]: I0313 10:21:21.865720 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:22 crc kubenswrapper[4632]: I0313 10:21:22.055146 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d0fc567-0682-4bbc-981b-b4d1df62aa4e" path="/var/lib/kubelet/pods/0d0fc567-0682-4bbc-981b-b4d1df62aa4e/volumes" Mar 13 10:21:25 crc kubenswrapper[4632]: I0313 10:21:25.509363 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq" event={"ID":"82fe7ef6-50a5-41d4-9419-787812e16bd6","Type":"ContainerStarted","Data":"5b061cfc2623671cbdb73e62714370004bd271740be45f371ce11c87da85cf57"} Mar 13 10:21:25 crc kubenswrapper[4632]: I0313 10:21:25.509673 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq" Mar 13 10:21:25 crc kubenswrapper[4632]: I0313 10:21:25.539300 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq" podStartSLOduration=1.716355558 podStartE2EDuration="8.539280562s" podCreationTimestamp="2026-03-13 10:21:17 +0000 UTC" firstStartedPulling="2026-03-13 10:21:18.262296863 +0000 UTC m=+1052.284826996" lastFinishedPulling="2026-03-13 10:21:25.085221867 +0000 UTC m=+1059.107752000" observedRunningTime="2026-03-13 10:21:25.534837206 +0000 UTC m=+1059.557367349" watchObservedRunningTime="2026-03-13 10:21:25.539280562 +0000 UTC m=+1059.561810695" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.065370 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5lbm"] Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.065704 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x5lbm" podUID="2d96ff75-88fd-4637-9199-806314e5276d" containerName="registry-server" containerID="cri-o://db96b68191a6b558bd0777d22d1fa122614ae61dec4e51a4b871ee53f8b057b3" gracePeriod=2 Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.462868 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.516258 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d96ff75-88fd-4637-9199-806314e5276d-catalog-content\") pod \"2d96ff75-88fd-4637-9199-806314e5276d\" (UID: \"2d96ff75-88fd-4637-9199-806314e5276d\") " Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.516312 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d96ff75-88fd-4637-9199-806314e5276d-utilities\") pod \"2d96ff75-88fd-4637-9199-806314e5276d\" (UID: \"2d96ff75-88fd-4637-9199-806314e5276d\") " Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.516340 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7vjg\" (UniqueName: \"kubernetes.io/projected/2d96ff75-88fd-4637-9199-806314e5276d-kube-api-access-c7vjg\") pod \"2d96ff75-88fd-4637-9199-806314e5276d\" (UID: \"2d96ff75-88fd-4637-9199-806314e5276d\") " Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.517699 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d96ff75-88fd-4637-9199-806314e5276d-utilities" (OuterVolumeSpecName: "utilities") pod "2d96ff75-88fd-4637-9199-806314e5276d" (UID: "2d96ff75-88fd-4637-9199-806314e5276d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.519635 4632 generic.go:334] "Generic (PLEG): container finished" podID="2d96ff75-88fd-4637-9199-806314e5276d" containerID="db96b68191a6b558bd0777d22d1fa122614ae61dec4e51a4b871ee53f8b057b3" exitCode=0 Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.520422 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x5lbm" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.520904 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5lbm" event={"ID":"2d96ff75-88fd-4637-9199-806314e5276d","Type":"ContainerDied","Data":"db96b68191a6b558bd0777d22d1fa122614ae61dec4e51a4b871ee53f8b057b3"} Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.520950 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x5lbm" event={"ID":"2d96ff75-88fd-4637-9199-806314e5276d","Type":"ContainerDied","Data":"21e3226d1d96fc021bc557ff5418b6dec4ea17fe511f69b4fe8609410d428008"} Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.520970 4632 scope.go:117] "RemoveContainer" containerID="db96b68191a6b558bd0777d22d1fa122614ae61dec4e51a4b871ee53f8b057b3" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.544571 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d96ff75-88fd-4637-9199-806314e5276d-kube-api-access-c7vjg" (OuterVolumeSpecName: "kube-api-access-c7vjg") pod "2d96ff75-88fd-4637-9199-806314e5276d" (UID: "2d96ff75-88fd-4637-9199-806314e5276d"). InnerVolumeSpecName "kube-api-access-c7vjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.560664 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d96ff75-88fd-4637-9199-806314e5276d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2d96ff75-88fd-4637-9199-806314e5276d" (UID: "2d96ff75-88fd-4637-9199-806314e5276d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.561383 4632 scope.go:117] "RemoveContainer" containerID="559890583b5bdac795810d916d2ba129170ca28c94f2aae83a30fbc62b754214" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.586986 4632 scope.go:117] "RemoveContainer" containerID="9d09e614abc972c9bc41b522cbf719729037565923e2784a67b05584e2614a0c" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.605144 4632 scope.go:117] "RemoveContainer" containerID="db96b68191a6b558bd0777d22d1fa122614ae61dec4e51a4b871ee53f8b057b3" Mar 13 10:21:26 crc kubenswrapper[4632]: E0313 10:21:26.605706 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db96b68191a6b558bd0777d22d1fa122614ae61dec4e51a4b871ee53f8b057b3\": container with ID starting with db96b68191a6b558bd0777d22d1fa122614ae61dec4e51a4b871ee53f8b057b3 not found: ID does not exist" containerID="db96b68191a6b558bd0777d22d1fa122614ae61dec4e51a4b871ee53f8b057b3" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.605807 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db96b68191a6b558bd0777d22d1fa122614ae61dec4e51a4b871ee53f8b057b3"} err="failed to get container status \"db96b68191a6b558bd0777d22d1fa122614ae61dec4e51a4b871ee53f8b057b3\": rpc error: code = NotFound desc = could not find container \"db96b68191a6b558bd0777d22d1fa122614ae61dec4e51a4b871ee53f8b057b3\": container with ID starting with db96b68191a6b558bd0777d22d1fa122614ae61dec4e51a4b871ee53f8b057b3 not found: ID does not exist" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.605896 4632 scope.go:117] "RemoveContainer" containerID="559890583b5bdac795810d916d2ba129170ca28c94f2aae83a30fbc62b754214" Mar 13 10:21:26 crc kubenswrapper[4632]: E0313 10:21:26.606687 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"559890583b5bdac795810d916d2ba129170ca28c94f2aae83a30fbc62b754214\": container with ID starting with 559890583b5bdac795810d916d2ba129170ca28c94f2aae83a30fbc62b754214 not found: ID does not exist" containerID="559890583b5bdac795810d916d2ba129170ca28c94f2aae83a30fbc62b754214" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.606737 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"559890583b5bdac795810d916d2ba129170ca28c94f2aae83a30fbc62b754214"} err="failed to get container status \"559890583b5bdac795810d916d2ba129170ca28c94f2aae83a30fbc62b754214\": rpc error: code = NotFound desc = could not find container \"559890583b5bdac795810d916d2ba129170ca28c94f2aae83a30fbc62b754214\": container with ID starting with 559890583b5bdac795810d916d2ba129170ca28c94f2aae83a30fbc62b754214 not found: ID does not exist" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.606776 4632 scope.go:117] "RemoveContainer" containerID="9d09e614abc972c9bc41b522cbf719729037565923e2784a67b05584e2614a0c" Mar 13 10:21:26 crc kubenswrapper[4632]: E0313 10:21:26.607229 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d09e614abc972c9bc41b522cbf719729037565923e2784a67b05584e2614a0c\": container with ID starting with 9d09e614abc972c9bc41b522cbf719729037565923e2784a67b05584e2614a0c not found: ID does not exist" containerID="9d09e614abc972c9bc41b522cbf719729037565923e2784a67b05584e2614a0c" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.607308 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d09e614abc972c9bc41b522cbf719729037565923e2784a67b05584e2614a0c"} err="failed to get container status \"9d09e614abc972c9bc41b522cbf719729037565923e2784a67b05584e2614a0c\": rpc error: code = NotFound desc = could not find container \"9d09e614abc972c9bc41b522cbf719729037565923e2784a67b05584e2614a0c\": container with ID starting with 9d09e614abc972c9bc41b522cbf719729037565923e2784a67b05584e2614a0c not found: ID does not exist" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.618128 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2d96ff75-88fd-4637-9199-806314e5276d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.618176 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2d96ff75-88fd-4637-9199-806314e5276d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.618186 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7vjg\" (UniqueName: \"kubernetes.io/projected/2d96ff75-88fd-4637-9199-806314e5276d-kube-api-access-c7vjg\") on node \"crc\" DevicePath \"\"" Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.849218 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5lbm"] Mar 13 10:21:26 crc kubenswrapper[4632]: I0313 10:21:26.860493 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x5lbm"] Mar 13 10:21:28 crc kubenswrapper[4632]: I0313 10:21:28.052512 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d96ff75-88fd-4637-9199-806314e5276d" path="/var/lib/kubelet/pods/2d96ff75-88fd-4637-9199-806314e5276d/volumes" Mar 13 10:21:37 crc kubenswrapper[4632]: I0313 10:21:37.860985 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq" Mar 13 10:21:40 crc kubenswrapper[4632]: I0313 10:21:40.460448 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:21:40 crc kubenswrapper[4632]: I0313 10:21:40.460517 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:21:40 crc kubenswrapper[4632]: I0313 10:21:40.460565 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:21:40 crc kubenswrapper[4632]: I0313 10:21:40.461218 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"624a339b1e1f8b218223c2e3440b7f9925bb18567bb6def4fcf3bfc022198658"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:21:40 crc kubenswrapper[4632]: I0313 10:21:40.461273 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://624a339b1e1f8b218223c2e3440b7f9925bb18567bb6def4fcf3bfc022198658" gracePeriod=600 Mar 13 10:21:40 crc kubenswrapper[4632]: I0313 10:21:40.646715 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="624a339b1e1f8b218223c2e3440b7f9925bb18567bb6def4fcf3bfc022198658" exitCode=0 Mar 13 10:21:40 crc kubenswrapper[4632]: I0313 10:21:40.646784 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"624a339b1e1f8b218223c2e3440b7f9925bb18567bb6def4fcf3bfc022198658"} Mar 13 10:21:40 crc kubenswrapper[4632]: I0313 10:21:40.647048 4632 scope.go:117] "RemoveContainer" containerID="7fcd863f1a2b3af4768aa1d32979163bc846d3d472acea1e8c27ffcf3dfe0ffc" Mar 13 10:21:41 crc kubenswrapper[4632]: I0313 10:21:41.656188 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"e9a22f93dffae95945f5e47a3d15b0ebe11dc6b72712dcbe34fa0191ff687b27"} Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.278572 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs"] Mar 13 10:21:57 crc kubenswrapper[4632]: E0313 10:21:57.279265 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d0fc567-0682-4bbc-981b-b4d1df62aa4e" containerName="extract-content" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.279279 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d0fc567-0682-4bbc-981b-b4d1df62aa4e" containerName="extract-content" Mar 13 10:21:57 crc kubenswrapper[4632]: E0313 10:21:57.279294 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d0fc567-0682-4bbc-981b-b4d1df62aa4e" containerName="extract-utilities" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.279300 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d0fc567-0682-4bbc-981b-b4d1df62aa4e" containerName="extract-utilities" Mar 13 10:21:57 crc kubenswrapper[4632]: E0313 10:21:57.279309 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d96ff75-88fd-4637-9199-806314e5276d" containerName="extract-utilities" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.279317 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d96ff75-88fd-4637-9199-806314e5276d" containerName="extract-utilities" Mar 13 10:21:57 crc kubenswrapper[4632]: E0313 10:21:57.279328 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e13797-40e9-4942-a7b5-6174fa448654" containerName="extract-content" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.279334 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e13797-40e9-4942-a7b5-6174fa448654" containerName="extract-content" Mar 13 10:21:57 crc kubenswrapper[4632]: E0313 10:21:57.279344 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e13797-40e9-4942-a7b5-6174fa448654" containerName="registry-server" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.279349 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e13797-40e9-4942-a7b5-6174fa448654" containerName="registry-server" Mar 13 10:21:57 crc kubenswrapper[4632]: E0313 10:21:57.279361 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e13797-40e9-4942-a7b5-6174fa448654" containerName="extract-utilities" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.279366 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e13797-40e9-4942-a7b5-6174fa448654" containerName="extract-utilities" Mar 13 10:21:57 crc kubenswrapper[4632]: E0313 10:21:57.279373 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d0fc567-0682-4bbc-981b-b4d1df62aa4e" containerName="registry-server" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.279379 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d0fc567-0682-4bbc-981b-b4d1df62aa4e" containerName="registry-server" Mar 13 10:21:57 crc kubenswrapper[4632]: E0313 10:21:57.279397 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d96ff75-88fd-4637-9199-806314e5276d" containerName="extract-content" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.279407 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d96ff75-88fd-4637-9199-806314e5276d" containerName="extract-content" Mar 13 10:21:57 crc kubenswrapper[4632]: E0313 10:21:57.279417 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d96ff75-88fd-4637-9199-806314e5276d" containerName="registry-server" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.279423 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d96ff75-88fd-4637-9199-806314e5276d" containerName="registry-server" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.279528 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d96ff75-88fd-4637-9199-806314e5276d" containerName="registry-server" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.279542 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="95e13797-40e9-4942-a7b5-6174fa448654" containerName="registry-server" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.279552 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d0fc567-0682-4bbc-981b-b4d1df62aa4e" containerName="registry-server" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.279913 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.283563 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-8pjg4" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.291308 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.295078 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.295761 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.297580 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-5z6b8" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.320587 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.321572 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.323503 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-xb4n2" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.360173 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.361494 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.367439 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-d8hjt" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.377411 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.378198 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.383824 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-7sbvl" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.391471 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mp8hr\" (UniqueName: \"kubernetes.io/projected/20f92131-aca4-41ea-9144-a23bd9216f49-kube-api-access-mp8hr\") pod \"glance-operator-controller-manager-5964f64c48-qg79l\" (UID: \"20f92131-aca4-41ea-9144-a23bd9216f49\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.391584 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wpcd\" (UniqueName: \"kubernetes.io/projected/3f3a462e-4d89-45b3-8611-181aca5f8558-kube-api-access-7wpcd\") pod \"cinder-operator-controller-manager-984cd4dcf-f6c87\" (UID: \"3f3a462e-4d89-45b3-8611-181aca5f8558\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.391653 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96kl7\" (UniqueName: \"kubernetes.io/projected/68c5eb80-4214-42c5-a08d-de6012969621-kube-api-access-96kl7\") pod \"barbican-operator-controller-manager-677bd678f7-wj9qs\" (UID: \"68c5eb80-4214-42c5-a08d-de6012969621\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.391731 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-99gkt\" (UniqueName: \"kubernetes.io/projected/75d652c7-8521-4039-913a-fa625f89b094-kube-api-access-99gkt\") pod \"designate-operator-controller-manager-66d56f6ff4-cfcgn\" (UID: \"75d652c7-8521-4039-913a-fa625f89b094\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.391799 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9mvv\" (UniqueName: \"kubernetes.io/projected/ff6d4dcb-9eb8-44fc-951e-f2aecd77a639-kube-api-access-q9mvv\") pod \"heat-operator-controller-manager-77b6666d85-cgh6c\" (UID: \"ff6d4dcb-9eb8-44fc-951e-f2aecd77a639\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.396219 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.404230 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.415218 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.426162 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.431595 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-fpwd2" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.439254 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.484172 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.493208 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzvbs\" (UniqueName: \"kubernetes.io/projected/9a963f9c-ac58-4e21-abfa-fca1279a192d-kube-api-access-pzvbs\") pod \"horizon-operator-controller-manager-6d9d6b584d-2rv7s\" (UID: \"9a963f9c-ac58-4e21-abfa-fca1279a192d\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.493528 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mp8hr\" (UniqueName: \"kubernetes.io/projected/20f92131-aca4-41ea-9144-a23bd9216f49-kube-api-access-mp8hr\") pod \"glance-operator-controller-manager-5964f64c48-qg79l\" (UID: \"20f92131-aca4-41ea-9144-a23bd9216f49\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.493674 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wpcd\" (UniqueName: \"kubernetes.io/projected/3f3a462e-4d89-45b3-8611-181aca5f8558-kube-api-access-7wpcd\") pod \"cinder-operator-controller-manager-984cd4dcf-f6c87\" (UID: \"3f3a462e-4d89-45b3-8611-181aca5f8558\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.493807 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96kl7\" (UniqueName: \"kubernetes.io/projected/68c5eb80-4214-42c5-a08d-de6012969621-kube-api-access-96kl7\") pod \"barbican-operator-controller-manager-677bd678f7-wj9qs\" (UID: \"68c5eb80-4214-42c5-a08d-de6012969621\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.493930 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-99gkt\" (UniqueName: \"kubernetes.io/projected/75d652c7-8521-4039-913a-fa625f89b094-kube-api-access-99gkt\") pod \"designate-operator-controller-manager-66d56f6ff4-cfcgn\" (UID: \"75d652c7-8521-4039-913a-fa625f89b094\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.494144 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9mvv\" (UniqueName: \"kubernetes.io/projected/ff6d4dcb-9eb8-44fc-951e-f2aecd77a639-kube-api-access-q9mvv\") pod \"heat-operator-controller-manager-77b6666d85-cgh6c\" (UID: \"ff6d4dcb-9eb8-44fc-951e-f2aecd77a639\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.507063 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.526201 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.527034 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.528505 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-99gkt\" (UniqueName: \"kubernetes.io/projected/75d652c7-8521-4039-913a-fa625f89b094-kube-api-access-99gkt\") pod \"designate-operator-controller-manager-66d56f6ff4-cfcgn\" (UID: \"75d652c7-8521-4039-913a-fa625f89b094\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.528517 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9mvv\" (UniqueName: \"kubernetes.io/projected/ff6d4dcb-9eb8-44fc-951e-f2aecd77a639-kube-api-access-q9mvv\") pod \"heat-operator-controller-manager-77b6666d85-cgh6c\" (UID: \"ff6d4dcb-9eb8-44fc-951e-f2aecd77a639\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.538209 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mp8hr\" (UniqueName: \"kubernetes.io/projected/20f92131-aca4-41ea-9144-a23bd9216f49-kube-api-access-mp8hr\") pod \"glance-operator-controller-manager-5964f64c48-qg79l\" (UID: \"20f92131-aca4-41ea-9144-a23bd9216f49\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.538583 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-qf888" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.542077 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wpcd\" (UniqueName: \"kubernetes.io/projected/3f3a462e-4d89-45b3-8611-181aca5f8558-kube-api-access-7wpcd\") pod \"cinder-operator-controller-manager-984cd4dcf-f6c87\" (UID: \"3f3a462e-4d89-45b3-8611-181aca5f8558\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.549307 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.582059 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96kl7\" (UniqueName: \"kubernetes.io/projected/68c5eb80-4214-42c5-a08d-de6012969621-kube-api-access-96kl7\") pod \"barbican-operator-controller-manager-677bd678f7-wj9qs\" (UID: \"68c5eb80-4214-42c5-a08d-de6012969621\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.586452 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.587464 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.590855 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-cxrjj" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.595419 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert\") pod \"infra-operator-controller-manager-5995f4446f-flfxh\" (UID: \"1542a9c8-92f6-4bc9-8231-829f649b0b8f\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.595504 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnh7r\" (UniqueName: \"kubernetes.io/projected/1542a9c8-92f6-4bc9-8231-829f649b0b8f-kube-api-access-pnh7r\") pod \"infra-operator-controller-manager-5995f4446f-flfxh\" (UID: \"1542a9c8-92f6-4bc9-8231-829f649b0b8f\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.595533 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psrhj\" (UniqueName: \"kubernetes.io/projected/c8fc6f03-c43b-4ade-92a8-acc5537a4eeb-kube-api-access-psrhj\") pod \"ironic-operator-controller-manager-6bbb499bbc-wtzrw\" (UID: \"c8fc6f03-c43b-4ade-92a8-acc5537a4eeb\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.595571 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzvbs\" (UniqueName: \"kubernetes.io/projected/9a963f9c-ac58-4e21-abfa-fca1279a192d-kube-api-access-pzvbs\") pod \"horizon-operator-controller-manager-6d9d6b584d-2rv7s\" (UID: \"9a963f9c-ac58-4e21-abfa-fca1279a192d\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.601531 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.620029 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.620473 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.634809 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.635600 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.646523 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.650001 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-wdj9j" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.665008 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.670250 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.675727 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzvbs\" (UniqueName: \"kubernetes.io/projected/9a963f9c-ac58-4e21-abfa-fca1279a192d-kube-api-access-pzvbs\") pod \"horizon-operator-controller-manager-6d9d6b584d-2rv7s\" (UID: \"9a963f9c-ac58-4e21-abfa-fca1279a192d\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.676047 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.700445 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert\") pod \"infra-operator-controller-manager-5995f4446f-flfxh\" (UID: \"1542a9c8-92f6-4bc9-8231-829f649b0b8f\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.700526 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnh7r\" (UniqueName: \"kubernetes.io/projected/1542a9c8-92f6-4bc9-8231-829f649b0b8f-kube-api-access-pnh7r\") pod \"infra-operator-controller-manager-5995f4446f-flfxh\" (UID: \"1542a9c8-92f6-4bc9-8231-829f649b0b8f\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.700548 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psrhj\" (UniqueName: \"kubernetes.io/projected/c8fc6f03-c43b-4ade-92a8-acc5537a4eeb-kube-api-access-psrhj\") pod \"ironic-operator-controller-manager-6bbb499bbc-wtzrw\" (UID: \"c8fc6f03-c43b-4ade-92a8-acc5537a4eeb\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" Mar 13 10:21:57 crc kubenswrapper[4632]: E0313 10:21:57.701317 4632 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 10:21:57 crc kubenswrapper[4632]: E0313 10:21:57.701367 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert podName:1542a9c8-92f6-4bc9-8231-829f649b0b8f nodeName:}" failed. No retries permitted until 2026-03-13 10:21:58.201348944 +0000 UTC m=+1092.223879077 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert") pod "infra-operator-controller-manager-5995f4446f-flfxh" (UID: "1542a9c8-92f6-4bc9-8231-829f649b0b8f") : secret "infra-operator-webhook-server-cert" not found Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.701775 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.713608 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.714423 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.719430 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-qnd2l" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.733814 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnh7r\" (UniqueName: \"kubernetes.io/projected/1542a9c8-92f6-4bc9-8231-829f649b0b8f-kube-api-access-pnh7r\") pod \"infra-operator-controller-manager-5995f4446f-flfxh\" (UID: \"1542a9c8-92f6-4bc9-8231-829f649b0b8f\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.743761 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.744618 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.749920 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-77vkb" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.751073 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psrhj\" (UniqueName: \"kubernetes.io/projected/c8fc6f03-c43b-4ade-92a8-acc5537a4eeb-kube-api-access-psrhj\") pod \"ironic-operator-controller-manager-6bbb499bbc-wtzrw\" (UID: \"c8fc6f03-c43b-4ade-92a8-acc5537a4eeb\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.767239 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.767348 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.784470 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.802142 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.803116 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.803418 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzwsh\" (UniqueName: \"kubernetes.io/projected/9040a0e0-2a56-4331-ba50-b19ff05ef0c0-kube-api-access-bzwsh\") pod \"mariadb-operator-controller-manager-658d4cdd5-szd7c\" (UID: \"9040a0e0-2a56-4331-ba50-b19ff05ef0c0\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.803507 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xct9m\" (UniqueName: \"kubernetes.io/projected/7b491335-6a73-46de-8098-f27ff4c6f795-kube-api-access-xct9m\") pod \"manila-operator-controller-manager-68f45f9d9f-sxw8d\" (UID: \"7b491335-6a73-46de-8098-f27ff4c6f795\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.803573 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pd2q\" (UniqueName: \"kubernetes.io/projected/f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda-kube-api-access-8pd2q\") pod \"keystone-operator-controller-manager-684f77d66d-6nb82\" (UID: \"f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.807001 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-bnhzq" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.817709 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.827867 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-569cc54c5-628ss"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.828713 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.840633 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-sn5sp" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.863010 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.863821 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.875109 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-q9f6p" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.882729 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.913023 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pd2q\" (UniqueName: \"kubernetes.io/projected/f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda-kube-api-access-8pd2q\") pod \"keystone-operator-controller-manager-684f77d66d-6nb82\" (UID: \"f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.913089 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzwsh\" (UniqueName: \"kubernetes.io/projected/9040a0e0-2a56-4331-ba50-b19ff05ef0c0-kube-api-access-bzwsh\") pod \"mariadb-operator-controller-manager-658d4cdd5-szd7c\" (UID: \"9040a0e0-2a56-4331-ba50-b19ff05ef0c0\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.913147 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xct9m\" (UniqueName: \"kubernetes.io/projected/7b491335-6a73-46de-8098-f27ff4c6f795-kube-api-access-xct9m\") pod \"manila-operator-controller-manager-68f45f9d9f-sxw8d\" (UID: \"7b491335-6a73-46de-8098-f27ff4c6f795\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.932777 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.941645 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xct9m\" (UniqueName: \"kubernetes.io/projected/7b491335-6a73-46de-8098-f27ff4c6f795-kube-api-access-xct9m\") pod \"manila-operator-controller-manager-68f45f9d9f-sxw8d\" (UID: \"7b491335-6a73-46de-8098-f27ff4c6f795\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.946720 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzwsh\" (UniqueName: \"kubernetes.io/projected/9040a0e0-2a56-4331-ba50-b19ff05ef0c0-kube-api-access-bzwsh\") pod \"mariadb-operator-controller-manager-658d4cdd5-szd7c\" (UID: \"9040a0e0-2a56-4331-ba50-b19ff05ef0c0\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.954543 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.955681 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.965345 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.957590 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pd2q\" (UniqueName: \"kubernetes.io/projected/f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda-kube-api-access-8pd2q\") pod \"keystone-operator-controller-manager-684f77d66d-6nb82\" (UID: \"f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.969912 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-rcfbr" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.970420 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.979331 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.980108 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.983869 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-x42p7" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.989870 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n"] Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.992015 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" Mar 13 10:21:57 crc kubenswrapper[4632]: I0313 10:21:57.994839 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-jg474" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.001103 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.016092 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.016795 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl4tq\" (UniqueName: \"kubernetes.io/projected/c33d0da9-5a04-42d6-80d3-2f558b4a90b0-kube-api-access-dl4tq\") pod \"neutron-operator-controller-manager-776c5696bf-bkmbn\" (UID: \"c33d0da9-5a04-42d6-80d3-2f558b4a90b0\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.016841 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsp6d\" (UniqueName: \"kubernetes.io/projected/9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5-kube-api-access-zsp6d\") pod \"octavia-operator-controller-manager-5f4f55cb5c-62gpm\" (UID: \"9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.016899 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qp7zl\" (UniqueName: \"kubernetes.io/projected/d04e9aa6-f234-4ffa-81e2-1a2407addb77-kube-api-access-qp7zl\") pod \"nova-operator-controller-manager-569cc54c5-628ss\" (UID: \"d04e9aa6-f234-4ffa-81e2-1a2407addb77\") " pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.034036 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-569cc54c5-628ss"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.169522 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt4gr\" (UniqueName: \"kubernetes.io/projected/e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5-kube-api-access-nt4gr\") pod \"placement-operator-controller-manager-574d45c66c-qkr9n\" (UID: \"e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.177160 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsqv6\" (UniqueName: \"kubernetes.io/projected/0a9d48f4-d68b-4ef9-826e-ed619c761405-kube-api-access-zsqv6\") pod \"ovn-operator-controller-manager-bbc5b68f9-4m8kf\" (UID: \"0a9d48f4-d68b-4ef9-826e-ed619c761405\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.177456 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl4tq\" (UniqueName: \"kubernetes.io/projected/c33d0da9-5a04-42d6-80d3-2f558b4a90b0-kube-api-access-dl4tq\") pod \"neutron-operator-controller-manager-776c5696bf-bkmbn\" (UID: \"c33d0da9-5a04-42d6-80d3-2f558b4a90b0\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.177641 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2q22k\" (UniqueName: \"kubernetes.io/projected/2d221857-ee77-4165-a351-ecd5fc424970-kube-api-access-2q22k\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7v927j\" (UID: \"2d221857-ee77-4165-a351-ecd5fc424970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.177816 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsp6d\" (UniqueName: \"kubernetes.io/projected/9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5-kube-api-access-zsp6d\") pod \"octavia-operator-controller-manager-5f4f55cb5c-62gpm\" (UID: \"9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.181306 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7v927j\" (UID: \"2d221857-ee77-4165-a351-ecd5fc424970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.181423 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qp7zl\" (UniqueName: \"kubernetes.io/projected/d04e9aa6-f234-4ffa-81e2-1a2407addb77-kube-api-access-qp7zl\") pod \"nova-operator-controller-manager-569cc54c5-628ss\" (UID: \"d04e9aa6-f234-4ffa-81e2-1a2407addb77\") " pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.179800 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.223804 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qp7zl\" (UniqueName: \"kubernetes.io/projected/d04e9aa6-f234-4ffa-81e2-1a2407addb77-kube-api-access-qp7zl\") pod \"nova-operator-controller-manager-569cc54c5-628ss\" (UID: \"d04e9aa6-f234-4ffa-81e2-1a2407addb77\") " pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.231265 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl4tq\" (UniqueName: \"kubernetes.io/projected/c33d0da9-5a04-42d6-80d3-2f558b4a90b0-kube-api-access-dl4tq\") pod \"neutron-operator-controller-manager-776c5696bf-bkmbn\" (UID: \"c33d0da9-5a04-42d6-80d3-2f558b4a90b0\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.231831 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.263721 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.283596 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.288335 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert\") pod \"infra-operator-controller-manager-5995f4446f-flfxh\" (UID: \"1542a9c8-92f6-4bc9-8231-829f649b0b8f\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.288384 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt4gr\" (UniqueName: \"kubernetes.io/projected/e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5-kube-api-access-nt4gr\") pod \"placement-operator-controller-manager-574d45c66c-qkr9n\" (UID: \"e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.288409 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsqv6\" (UniqueName: \"kubernetes.io/projected/0a9d48f4-d68b-4ef9-826e-ed619c761405-kube-api-access-zsqv6\") pod \"ovn-operator-controller-manager-bbc5b68f9-4m8kf\" (UID: \"0a9d48f4-d68b-4ef9-826e-ed619c761405\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.288449 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2q22k\" (UniqueName: \"kubernetes.io/projected/2d221857-ee77-4165-a351-ecd5fc424970-kube-api-access-2q22k\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7v927j\" (UID: \"2d221857-ee77-4165-a351-ecd5fc424970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.288501 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7v927j\" (UID: \"2d221857-ee77-4165-a351-ecd5fc424970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.288657 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsp6d\" (UniqueName: \"kubernetes.io/projected/9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5-kube-api-access-zsp6d\") pod \"octavia-operator-controller-manager-5f4f55cb5c-62gpm\" (UID: \"9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" Mar 13 10:21:58 crc kubenswrapper[4632]: E0313 10:21:58.289097 4632 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 10:21:58 crc kubenswrapper[4632]: E0313 10:21:58.289144 4632 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 10:21:58 crc kubenswrapper[4632]: E0313 10:21:58.289147 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert podName:2d221857-ee77-4165-a351-ecd5fc424970 nodeName:}" failed. No retries permitted until 2026-03-13 10:21:58.789133206 +0000 UTC m=+1092.811663339 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7v927j" (UID: "2d221857-ee77-4165-a351-ecd5fc424970") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 10:21:58 crc kubenswrapper[4632]: E0313 10:21:58.289377 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert podName:1542a9c8-92f6-4bc9-8231-829f649b0b8f nodeName:}" failed. No retries permitted until 2026-03-13 10:21:59.289348271 +0000 UTC m=+1093.311878404 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert") pod "infra-operator-controller-manager-5995f4446f-flfxh" (UID: "1542a9c8-92f6-4bc9-8231-829f649b0b8f") : secret "infra-operator-webhook-server-cert" not found Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.316083 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-qbfg2"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.318846 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-677c674df7-qbfg2" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.326113 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-wr8sw" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.327121 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.383600 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt4gr\" (UniqueName: \"kubernetes.io/projected/e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5-kube-api-access-nt4gr\") pod \"placement-operator-controller-manager-574d45c66c-qkr9n\" (UID: \"e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.385331 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.390892 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8p7rc\" (UniqueName: \"kubernetes.io/projected/2d8a9f3a-6631-4c1e-8381-3bc313837ca0-kube-api-access-8p7rc\") pod \"swift-operator-controller-manager-677c674df7-qbfg2\" (UID: \"2d8a9f3a-6631-4c1e-8381-3bc313837ca0\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-qbfg2" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.399761 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2q22k\" (UniqueName: \"kubernetes.io/projected/2d221857-ee77-4165-a351-ecd5fc424970-kube-api-access-2q22k\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7v927j\" (UID: \"2d221857-ee77-4165-a351-ecd5fc424970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.401090 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsqv6\" (UniqueName: \"kubernetes.io/projected/0a9d48f4-d68b-4ef9-826e-ed619c761405-kube-api-access-zsqv6\") pod \"ovn-operator-controller-manager-bbc5b68f9-4m8kf\" (UID: \"0a9d48f4-d68b-4ef9-826e-ed619c761405\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.425738 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-qbfg2"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.449836 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.450835 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.471114 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-ds7sc" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.494698 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8p7rc\" (UniqueName: \"kubernetes.io/projected/2d8a9f3a-6631-4c1e-8381-3bc313837ca0-kube-api-access-8p7rc\") pod \"swift-operator-controller-manager-677c674df7-qbfg2\" (UID: \"2d8a9f3a-6631-4c1e-8381-3bc313837ca0\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-qbfg2" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.495032 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmz6h\" (UniqueName: \"kubernetes.io/projected/ee081327-4c3f-4c0a-9085-71085c6487b5-kube-api-access-fmz6h\") pod \"telemetry-operator-controller-manager-6cd66dbd4b-nt7np\" (UID: \"ee081327-4c3f-4c0a-9085-71085c6487b5\") " pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.495303 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.541999 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.565068 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.565964 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.579478 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-8vng2" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.580063 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.609625 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmz6h\" (UniqueName: \"kubernetes.io/projected/ee081327-4c3f-4c0a-9085-71085c6487b5-kube-api-access-fmz6h\") pod \"telemetry-operator-controller-manager-6cd66dbd4b-nt7np\" (UID: \"ee081327-4c3f-4c0a-9085-71085c6487b5\") " pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.609772 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4vf2\" (UniqueName: \"kubernetes.io/projected/7bab78c8-7dac-48dc-a426-ccd4ae00a428-kube-api-access-j4vf2\") pod \"test-operator-controller-manager-5c5cb9c4d7-jwrgq\" (UID: \"7bab78c8-7dac-48dc-a426-ccd4ae00a428\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.620696 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.621519 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.654141 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-z8lfp" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.669995 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.695479 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8p7rc\" (UniqueName: \"kubernetes.io/projected/2d8a9f3a-6631-4c1e-8381-3bc313837ca0-kube-api-access-8p7rc\") pod \"swift-operator-controller-manager-677c674df7-qbfg2\" (UID: \"2d8a9f3a-6631-4c1e-8381-3bc313837ca0\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-qbfg2" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.722527 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmz6h\" (UniqueName: \"kubernetes.io/projected/ee081327-4c3f-4c0a-9085-71085c6487b5-kube-api-access-fmz6h\") pod \"telemetry-operator-controller-manager-6cd66dbd4b-nt7np\" (UID: \"ee081327-4c3f-4c0a-9085-71085c6487b5\") " pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.754398 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-677c674df7-qbfg2" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.789090 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.822965 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7v927j\" (UID: \"2d221857-ee77-4165-a351-ecd5fc424970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.823146 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4vf2\" (UniqueName: \"kubernetes.io/projected/7bab78c8-7dac-48dc-a426-ccd4ae00a428-kube-api-access-j4vf2\") pod \"test-operator-controller-manager-5c5cb9c4d7-jwrgq\" (UID: \"7bab78c8-7dac-48dc-a426-ccd4ae00a428\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.823292 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rk5b8\" (UniqueName: \"kubernetes.io/projected/e0d1d349-d63d-498b-ae15-3121f9ae73f8-kube-api-access-rk5b8\") pod \"watcher-operator-controller-manager-6dd88c6f67-kv8b2\" (UID: \"e0d1d349-d63d-498b-ae15-3121f9ae73f8\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" Mar 13 10:21:58 crc kubenswrapper[4632]: E0313 10:21:58.823491 4632 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 10:21:58 crc kubenswrapper[4632]: E0313 10:21:58.823785 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert podName:2d221857-ee77-4165-a351-ecd5fc424970 nodeName:}" failed. No retries permitted until 2026-03-13 10:21:59.823768638 +0000 UTC m=+1093.846298771 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7v927j" (UID: "2d221857-ee77-4165-a351-ecd5fc424970") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.855273 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.870741 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4vf2\" (UniqueName: \"kubernetes.io/projected/7bab78c8-7dac-48dc-a426-ccd4ae00a428-kube-api-access-j4vf2\") pod \"test-operator-controller-manager-5c5cb9c4d7-jwrgq\" (UID: \"7bab78c8-7dac-48dc-a426-ccd4ae00a428\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.918061 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.918994 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.923674 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.923799 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.923909 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-thpwj" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.926639 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltgld\" (UniqueName: \"kubernetes.io/projected/3fdb377f-5a78-4687-82e1-50718514290d-kube-api-access-ltgld\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.926792 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.926819 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.926904 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rk5b8\" (UniqueName: \"kubernetes.io/projected/e0d1d349-d63d-498b-ae15-3121f9ae73f8-kube-api-access-rk5b8\") pod \"watcher-operator-controller-manager-6dd88c6f67-kv8b2\" (UID: \"e0d1d349-d63d-498b-ae15-3121f9ae73f8\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.936016 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.945741 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2lzt8"] Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.946816 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2lzt8" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.951053 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-78jxg" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.956702 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rk5b8\" (UniqueName: \"kubernetes.io/projected/e0d1d349-d63d-498b-ae15-3121f9ae73f8-kube-api-access-rk5b8\") pod \"watcher-operator-controller-manager-6dd88c6f67-kv8b2\" (UID: \"e0d1d349-d63d-498b-ae15-3121f9ae73f8\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" Mar 13 10:21:58 crc kubenswrapper[4632]: I0313 10:21:58.971734 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2lzt8"] Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.031644 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltgld\" (UniqueName: \"kubernetes.io/projected/3fdb377f-5a78-4687-82e1-50718514290d-kube-api-access-ltgld\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.031985 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp7js\" (UniqueName: \"kubernetes.io/projected/daba1153-3b28-4234-8dd0-ec20160abbfe-kube-api-access-dp7js\") pod \"rabbitmq-cluster-operator-manager-668c99d594-2lzt8\" (UID: \"daba1153-3b28-4234-8dd0-ec20160abbfe\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2lzt8" Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.032009 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.032026 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:21:59 crc kubenswrapper[4632]: E0313 10:21:59.032515 4632 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 10:21:59 crc kubenswrapper[4632]: E0313 10:21:59.032574 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs podName:3fdb377f-5a78-4687-82e1-50718514290d nodeName:}" failed. No retries permitted until 2026-03-13 10:21:59.532558299 +0000 UTC m=+1093.555088432 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs") pod "openstack-operator-controller-manager-85c677895b-thbc4" (UID: "3fdb377f-5a78-4687-82e1-50718514290d") : secret "webhook-server-cert" not found Mar 13 10:21:59 crc kubenswrapper[4632]: E0313 10:21:59.033727 4632 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 10:21:59 crc kubenswrapper[4632]: E0313 10:21:59.033756 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs podName:3fdb377f-5a78-4687-82e1-50718514290d nodeName:}" failed. No retries permitted until 2026-03-13 10:21:59.533747888 +0000 UTC m=+1093.556278011 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs") pod "openstack-operator-controller-manager-85c677895b-thbc4" (UID: "3fdb377f-5a78-4687-82e1-50718514290d") : secret "metrics-server-cert" not found Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.060791 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.092756 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltgld\" (UniqueName: \"kubernetes.io/projected/3fdb377f-5a78-4687-82e1-50718514290d-kube-api-access-ltgld\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.106257 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87"] Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.107741 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.132908 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp7js\" (UniqueName: \"kubernetes.io/projected/daba1153-3b28-4234-8dd0-ec20160abbfe-kube-api-access-dp7js\") pod \"rabbitmq-cluster-operator-manager-668c99d594-2lzt8\" (UID: \"daba1153-3b28-4234-8dd0-ec20160abbfe\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2lzt8" Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.197932 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs"] Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.204453 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp7js\" (UniqueName: \"kubernetes.io/projected/daba1153-3b28-4234-8dd0-ec20160abbfe-kube-api-access-dp7js\") pod \"rabbitmq-cluster-operator-manager-668c99d594-2lzt8\" (UID: \"daba1153-3b28-4234-8dd0-ec20160abbfe\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2lzt8" Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.212879 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s"] Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.340231 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert\") pod \"infra-operator-controller-manager-5995f4446f-flfxh\" (UID: \"1542a9c8-92f6-4bc9-8231-829f649b0b8f\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:21:59 crc kubenswrapper[4632]: E0313 10:21:59.341095 4632 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 10:21:59 crc kubenswrapper[4632]: E0313 10:21:59.341656 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert podName:1542a9c8-92f6-4bc9-8231-829f649b0b8f nodeName:}" failed. No retries permitted until 2026-03-13 10:22:01.341633679 +0000 UTC m=+1095.364163822 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert") pod "infra-operator-controller-manager-5995f4446f-flfxh" (UID: "1542a9c8-92f6-4bc9-8231-829f649b0b8f") : secret "infra-operator-webhook-server-cert" not found Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.354115 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2lzt8" Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.553128 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.553169 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:21:59 crc kubenswrapper[4632]: E0313 10:21:59.555330 4632 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 10:21:59 crc kubenswrapper[4632]: E0313 10:21:59.555408 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs podName:3fdb377f-5a78-4687-82e1-50718514290d nodeName:}" failed. No retries permitted until 2026-03-13 10:22:00.555387908 +0000 UTC m=+1094.577918041 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs") pod "openstack-operator-controller-manager-85c677895b-thbc4" (UID: "3fdb377f-5a78-4687-82e1-50718514290d") : secret "metrics-server-cert" not found Mar 13 10:21:59 crc kubenswrapper[4632]: E0313 10:21:59.555922 4632 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 10:21:59 crc kubenswrapper[4632]: E0313 10:21:59.556014 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs podName:3fdb377f-5a78-4687-82e1-50718514290d nodeName:}" failed. No retries permitted until 2026-03-13 10:22:00.556001523 +0000 UTC m=+1094.578531656 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs") pod "openstack-operator-controller-manager-85c677895b-thbc4" (UID: "3fdb377f-5a78-4687-82e1-50718514290d") : secret "webhook-server-cert" not found Mar 13 10:21:59 crc kubenswrapper[4632]: W0313 10:21:59.741161 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc8fc6f03_c43b_4ade_92a8_acc5537a4eeb.slice/crio-8b14b6b02d647f7be0233d6ae3c76aafb06cdac55c64087f8d1a18d6d170164d WatchSource:0}: Error finding container 8b14b6b02d647f7be0233d6ae3c76aafb06cdac55c64087f8d1a18d6d170164d: Status 404 returned error can't find the container with id 8b14b6b02d647f7be0233d6ae3c76aafb06cdac55c64087f8d1a18d6d170164d Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.801158 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw"] Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.822416 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c"] Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.832601 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l"] Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.844071 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn"] Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.876231 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l" event={"ID":"20f92131-aca4-41ea-9144-a23bd9216f49","Type":"ContainerStarted","Data":"1b12cd6affbbba27a91255f8a7d15ae6836df88d5d0c65e64972b933b667f736"} Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.879788 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm"] Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.886552 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7v927j\" (UID: \"2d221857-ee77-4165-a351-ecd5fc424970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:21:59 crc kubenswrapper[4632]: E0313 10:21:59.886799 4632 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 10:21:59 crc kubenswrapper[4632]: E0313 10:21:59.888278 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert podName:2d221857-ee77-4165-a351-ecd5fc424970 nodeName:}" failed. No retries permitted until 2026-03-13 10:22:01.888255306 +0000 UTC m=+1095.910785439 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7v927j" (UID: "2d221857-ee77-4165-a351-ecd5fc424970") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.890347 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" event={"ID":"9a963f9c-ac58-4e21-abfa-fca1279a192d","Type":"ContainerStarted","Data":"b828a39b1e00247b03288b6f0c7d7291f0cebe51d87ba54cc05d5c9092c5939a"} Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.903428 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c" event={"ID":"9040a0e0-2a56-4331-ba50-b19ff05ef0c0","Type":"ContainerStarted","Data":"ecdb7d21968cab437082542a10cb34e09bc2b54ea5965a8ce16bde9dfc8089f3"} Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.911789 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82"] Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.916427 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn" event={"ID":"75d652c7-8521-4039-913a-fa625f89b094","Type":"ContainerStarted","Data":"c2dfe6e0745050c82493b5f6751e24433eca74094dba7adfe32b436c8ed15f81"} Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.922667 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c"] Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.933214 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn"] Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.933515 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" event={"ID":"68c5eb80-4214-42c5-a08d-de6012969621","Type":"ContainerStarted","Data":"afddeda00568af10f04e19906cded4b4cc285f511804b2870e8288b8e948d902"} Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.940478 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" event={"ID":"c8fc6f03-c43b-4ade-92a8-acc5537a4eeb","Type":"ContainerStarted","Data":"8b14b6b02d647f7be0233d6ae3c76aafb06cdac55c64087f8d1a18d6d170164d"} Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.944633 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87" event={"ID":"3f3a462e-4d89-45b3-8611-181aca5f8558","Type":"ContainerStarted","Data":"068122cb5d809545fff29ef97ce88c07259a8c6b589d361afe704f82996ee9df"} Mar 13 10:21:59 crc kubenswrapper[4632]: I0313 10:21:59.975375 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d"] Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.134535 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556622-7428t"] Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.135770 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556622-7428t" Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.139471 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.139807 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.146235 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556622-7428t"] Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.148707 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.173151 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2lzt8"] Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.196714 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb5hx\" (UniqueName: \"kubernetes.io/projected/bedc1d17-f5c4-4a62-ab0c-f20a002e859b-kube-api-access-mb5hx\") pod \"auto-csr-approver-29556622-7428t\" (UID: \"bedc1d17-f5c4-4a62-ab0c-f20a002e859b\") " pod="openshift-infra/auto-csr-approver-29556622-7428t" Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.216992 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf"] Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.251000 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2"] Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.255039 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n"] Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.266910 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-qbfg2"] Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.279608 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-569cc54c5-628ss"] Mar 13 10:22:00 crc kubenswrapper[4632]: W0313 10:22:00.279738 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d8a9f3a_6631_4c1e_8381_3bc313837ca0.slice/crio-318f9130b56f9833106599e46f4be878dc926c055c7a426d7e877917e09016bd WatchSource:0}: Error finding container 318f9130b56f9833106599e46f4be878dc926c055c7a426d7e877917e09016bd: Status 404 returned error can't find the container with id 318f9130b56f9833106599e46f4be878dc926c055c7a426d7e877917e09016bd Mar 13 10:22:00 crc kubenswrapper[4632]: W0313 10:22:00.284294 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd04e9aa6_f234_4ffa_81e2_1a2407addb77.slice/crio-ec0c150d251d0c1342ab20ef9e2f5047cf5ccd875f9bcb8b322ae58b99a25992 WatchSource:0}: Error finding container ec0c150d251d0c1342ab20ef9e2f5047cf5ccd875f9bcb8b322ae58b99a25992: Status 404 returned error can't find the container with id ec0c150d251d0c1342ab20ef9e2f5047cf5ccd875f9bcb8b322ae58b99a25992 Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.297383 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:2bd37bdd917e3abe72613a734ce5021330242ec8cae9b8da76c57a0765152922,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-qp7zl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-569cc54c5-628ss_openstack-operators(d04e9aa6-f234-4ffa-81e2-1a2407addb77): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.298581 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" podUID="d04e9aa6-f234-4ffa-81e2-1a2407addb77" Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.300621 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb5hx\" (UniqueName: \"kubernetes.io/projected/bedc1d17-f5c4-4a62-ab0c-f20a002e859b-kube-api-access-mb5hx\") pod \"auto-csr-approver-29556622-7428t\" (UID: \"bedc1d17-f5c4-4a62-ab0c-f20a002e859b\") " pod="openshift-infra/auto-csr-approver-29556622-7428t" Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.310187 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np"] Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.323307 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:4af709a2a6a1a1abb9659dbdd6fb3818122bdec7e66009fcced0bf0949f91554,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-rk5b8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6dd88c6f67-kv8b2_openstack-operators(e0d1d349-d63d-498b-ae15-3121f9ae73f8): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.323336 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nt4gr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-574d45c66c-qkr9n_openstack-operators(e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.327091 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" podUID="e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5" Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.327174 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" podUID="e0d1d349-d63d-498b-ae15-3121f9ae73f8" Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.331137 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:27c84b712abc2df6108e22636075eec25fea0229800f38594a492fd41b02c49d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fmz6h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-6cd66dbd4b-nt7np_openstack-operators(ee081327-4c3f-4c0a-9085-71085c6487b5): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.333479 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" podUID="ee081327-4c3f-4c0a-9085-71085c6487b5" Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.335172 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq"] Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.347299 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb5hx\" (UniqueName: \"kubernetes.io/projected/bedc1d17-f5c4-4a62-ab0c-f20a002e859b-kube-api-access-mb5hx\") pod \"auto-csr-approver-29556622-7428t\" (UID: \"bedc1d17-f5c4-4a62-ab0c-f20a002e859b\") " pod="openshift-infra/auto-csr-approver-29556622-7428t" Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.373335 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-j4vf2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5c5cb9c4d7-jwrgq_openstack-operators(7bab78c8-7dac-48dc-a426-ccd4ae00a428): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.374473 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" podUID="7bab78c8-7dac-48dc-a426-ccd4ae00a428" Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.460802 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556622-7428t" Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.610840 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.611225 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.611437 4632 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.611493 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs podName:3fdb377f-5a78-4687-82e1-50718514290d nodeName:}" failed. No retries permitted until 2026-03-13 10:22:02.611475477 +0000 UTC m=+1096.634005610 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs") pod "openstack-operator-controller-manager-85c677895b-thbc4" (UID: "3fdb377f-5a78-4687-82e1-50718514290d") : secret "webhook-server-cert" not found Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.611858 4632 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.611899 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs podName:3fdb377f-5a78-4687-82e1-50718514290d nodeName:}" failed. No retries permitted until 2026-03-13 10:22:02.611889296 +0000 UTC m=+1096.634419429 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs") pod "openstack-operator-controller-manager-85c677895b-thbc4" (UID: "3fdb377f-5a78-4687-82e1-50718514290d") : secret "metrics-server-cert" not found Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.927158 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556622-7428t"] Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.970680 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" event={"ID":"7bab78c8-7dac-48dc-a426-ccd4ae00a428","Type":"ContainerStarted","Data":"0cc9f256adf45e067289c3c09cf5a3b887718e0ed4aebb00f8bdea99acfc8399"} Mar 13 10:22:00 crc kubenswrapper[4632]: E0313 10:22:00.972052 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" podUID="7bab78c8-7dac-48dc-a426-ccd4ae00a428" Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.974111 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" event={"ID":"f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda","Type":"ContainerStarted","Data":"238d41f4e6a11f56c98923b450487bbb345510f86b4677377ab990be7a3b5c6a"} Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.983070 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" event={"ID":"ff6d4dcb-9eb8-44fc-951e-f2aecd77a639","Type":"ContainerStarted","Data":"86b1d28feab0bd93d02ef0640d701cb5efd494a7316382881d4f7b85938d06cb"} Mar 13 10:22:00 crc kubenswrapper[4632]: I0313 10:22:00.986477 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2lzt8" event={"ID":"daba1153-3b28-4234-8dd0-ec20160abbfe","Type":"ContainerStarted","Data":"48107bf864a23956d2ba690d946ba69cb98e569466221aab236b2174d3456fce"} Mar 13 10:22:01 crc kubenswrapper[4632]: I0313 10:22:01.002836 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" event={"ID":"ee081327-4c3f-4c0a-9085-71085c6487b5","Type":"ContainerStarted","Data":"5252ec9dd71bcfd5c302107c0607594a44f45cfd9ef669b450dbea52c1fe870c"} Mar 13 10:22:01 crc kubenswrapper[4632]: I0313 10:22:01.005056 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" event={"ID":"e0d1d349-d63d-498b-ae15-3121f9ae73f8","Type":"ContainerStarted","Data":"8cab672d4a6a2d22b35fd58f618ec6c32836e4e3a60fbd42506cd93f4230fcbb"} Mar 13 10:22:01 crc kubenswrapper[4632]: E0313 10:22:01.010499 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:27c84b712abc2df6108e22636075eec25fea0229800f38594a492fd41b02c49d\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" podUID="ee081327-4c3f-4c0a-9085-71085c6487b5" Mar 13 10:22:01 crc kubenswrapper[4632]: I0313 10:22:01.011487 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" event={"ID":"9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5","Type":"ContainerStarted","Data":"523cef2b6f8900fac046ee029faf34f4d6f4fadca57bf80c65f3eab1f3b235fd"} Mar 13 10:22:01 crc kubenswrapper[4632]: E0313 10:22:01.011538 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4af709a2a6a1a1abb9659dbdd6fb3818122bdec7e66009fcced0bf0949f91554\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" podUID="e0d1d349-d63d-498b-ae15-3121f9ae73f8" Mar 13 10:22:01 crc kubenswrapper[4632]: I0313 10:22:01.017955 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" event={"ID":"0a9d48f4-d68b-4ef9-826e-ed619c761405","Type":"ContainerStarted","Data":"fad94d640a989a658da91c9c469a80976afb4335402e17bf8610cd318e33e922"} Mar 13 10:22:01 crc kubenswrapper[4632]: I0313 10:22:01.029343 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" event={"ID":"d04e9aa6-f234-4ffa-81e2-1a2407addb77","Type":"ContainerStarted","Data":"ec0c150d251d0c1342ab20ef9e2f5047cf5ccd875f9bcb8b322ae58b99a25992"} Mar 13 10:22:01 crc kubenswrapper[4632]: E0313 10:22:01.035363 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:2bd37bdd917e3abe72613a734ce5021330242ec8cae9b8da76c57a0765152922\\\"\"" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" podUID="d04e9aa6-f234-4ffa-81e2-1a2407addb77" Mar 13 10:22:01 crc kubenswrapper[4632]: I0313 10:22:01.040017 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn" event={"ID":"c33d0da9-5a04-42d6-80d3-2f558b4a90b0","Type":"ContainerStarted","Data":"cd5d441f5b93e630c3961ba5f4b1df277192e39becf2051dc9847a995c777521"} Mar 13 10:22:01 crc kubenswrapper[4632]: I0313 10:22:01.055233 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d" event={"ID":"7b491335-6a73-46de-8098-f27ff4c6f795","Type":"ContainerStarted","Data":"a1959e840f4a0b201ad8fbaf6f64fc480857a670b1237564c4fb9ad0250ca5ea"} Mar 13 10:22:01 crc kubenswrapper[4632]: I0313 10:22:01.067529 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" event={"ID":"e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5","Type":"ContainerStarted","Data":"8960cdc79b8729e94514f9d4f70168401027047a85c476a873705f12f51c1873"} Mar 13 10:22:01 crc kubenswrapper[4632]: E0313 10:22:01.070136 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978\\\"\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" podUID="e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5" Mar 13 10:22:01 crc kubenswrapper[4632]: I0313 10:22:01.074811 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-677c674df7-qbfg2" event={"ID":"2d8a9f3a-6631-4c1e-8381-3bc313837ca0","Type":"ContainerStarted","Data":"318f9130b56f9833106599e46f4be878dc926c055c7a426d7e877917e09016bd"} Mar 13 10:22:01 crc kubenswrapper[4632]: I0313 10:22:01.429634 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert\") pod \"infra-operator-controller-manager-5995f4446f-flfxh\" (UID: \"1542a9c8-92f6-4bc9-8231-829f649b0b8f\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:22:01 crc kubenswrapper[4632]: E0313 10:22:01.429817 4632 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 10:22:01 crc kubenswrapper[4632]: E0313 10:22:01.429865 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert podName:1542a9c8-92f6-4bc9-8231-829f649b0b8f nodeName:}" failed. No retries permitted until 2026-03-13 10:22:05.429849961 +0000 UTC m=+1099.452380094 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert") pod "infra-operator-controller-manager-5995f4446f-flfxh" (UID: "1542a9c8-92f6-4bc9-8231-829f649b0b8f") : secret "infra-operator-webhook-server-cert" not found Mar 13 10:22:01 crc kubenswrapper[4632]: I0313 10:22:01.957452 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7v927j\" (UID: \"2d221857-ee77-4165-a351-ecd5fc424970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:22:01 crc kubenswrapper[4632]: E0313 10:22:01.959754 4632 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 10:22:01 crc kubenswrapper[4632]: E0313 10:22:01.959829 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert podName:2d221857-ee77-4165-a351-ecd5fc424970 nodeName:}" failed. No retries permitted until 2026-03-13 10:22:05.959809491 +0000 UTC m=+1099.982339624 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7v927j" (UID: "2d221857-ee77-4165-a351-ecd5fc424970") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 10:22:02 crc kubenswrapper[4632]: I0313 10:22:02.125656 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556622-7428t" event={"ID":"bedc1d17-f5c4-4a62-ab0c-f20a002e859b","Type":"ContainerStarted","Data":"235384aa6e3de087f78098e5ca2543f682f052a84f878bf4d06b8cb3d539f270"} Mar 13 10:22:02 crc kubenswrapper[4632]: E0313 10:22:02.138691 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978\\\"\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" podUID="e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5" Mar 13 10:22:02 crc kubenswrapper[4632]: E0313 10:22:02.139490 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" podUID="7bab78c8-7dac-48dc-a426-ccd4ae00a428" Mar 13 10:22:02 crc kubenswrapper[4632]: E0313 10:22:02.139574 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:2bd37bdd917e3abe72613a734ce5021330242ec8cae9b8da76c57a0765152922\\\"\"" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" podUID="d04e9aa6-f234-4ffa-81e2-1a2407addb77" Mar 13 10:22:02 crc kubenswrapper[4632]: E0313 10:22:02.139715 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:4af709a2a6a1a1abb9659dbdd6fb3818122bdec7e66009fcced0bf0949f91554\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" podUID="e0d1d349-d63d-498b-ae15-3121f9ae73f8" Mar 13 10:22:02 crc kubenswrapper[4632]: E0313 10:22:02.141980 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:27c84b712abc2df6108e22636075eec25fea0229800f38594a492fd41b02c49d\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" podUID="ee081327-4c3f-4c0a-9085-71085c6487b5" Mar 13 10:22:02 crc kubenswrapper[4632]: I0313 10:22:02.691628 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:22:02 crc kubenswrapper[4632]: I0313 10:22:02.691688 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:22:02 crc kubenswrapper[4632]: E0313 10:22:02.691778 4632 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 10:22:02 crc kubenswrapper[4632]: E0313 10:22:02.691837 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs podName:3fdb377f-5a78-4687-82e1-50718514290d nodeName:}" failed. No retries permitted until 2026-03-13 10:22:06.69181931 +0000 UTC m=+1100.714349443 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs") pod "openstack-operator-controller-manager-85c677895b-thbc4" (UID: "3fdb377f-5a78-4687-82e1-50718514290d") : secret "metrics-server-cert" not found Mar 13 10:22:02 crc kubenswrapper[4632]: E0313 10:22:02.691870 4632 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 10:22:02 crc kubenswrapper[4632]: E0313 10:22:02.691907 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs podName:3fdb377f-5a78-4687-82e1-50718514290d nodeName:}" failed. No retries permitted until 2026-03-13 10:22:06.691896342 +0000 UTC m=+1100.714426475 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs") pod "openstack-operator-controller-manager-85c677895b-thbc4" (UID: "3fdb377f-5a78-4687-82e1-50718514290d") : secret "webhook-server-cert" not found Mar 13 10:22:04 crc kubenswrapper[4632]: I0313 10:22:04.154933 4632 generic.go:334] "Generic (PLEG): container finished" podID="bedc1d17-f5c4-4a62-ab0c-f20a002e859b" containerID="0e23e3344de45eadba8d2e2f7dead6b7591126ab6ec56a759524e9fc0c54694e" exitCode=0 Mar 13 10:22:04 crc kubenswrapper[4632]: I0313 10:22:04.155003 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556622-7428t" event={"ID":"bedc1d17-f5c4-4a62-ab0c-f20a002e859b","Type":"ContainerDied","Data":"0e23e3344de45eadba8d2e2f7dead6b7591126ab6ec56a759524e9fc0c54694e"} Mar 13 10:22:05 crc kubenswrapper[4632]: I0313 10:22:05.460285 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert\") pod \"infra-operator-controller-manager-5995f4446f-flfxh\" (UID: \"1542a9c8-92f6-4bc9-8231-829f649b0b8f\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:22:05 crc kubenswrapper[4632]: E0313 10:22:05.460497 4632 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 10:22:05 crc kubenswrapper[4632]: E0313 10:22:05.460763 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert podName:1542a9c8-92f6-4bc9-8231-829f649b0b8f nodeName:}" failed. No retries permitted until 2026-03-13 10:22:13.460738026 +0000 UTC m=+1107.483268249 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert") pod "infra-operator-controller-manager-5995f4446f-flfxh" (UID: "1542a9c8-92f6-4bc9-8231-829f649b0b8f") : secret "infra-operator-webhook-server-cert" not found Mar 13 10:22:05 crc kubenswrapper[4632]: I0313 10:22:05.966928 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7v927j\" (UID: \"2d221857-ee77-4165-a351-ecd5fc424970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:22:05 crc kubenswrapper[4632]: E0313 10:22:05.967133 4632 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 10:22:05 crc kubenswrapper[4632]: E0313 10:22:05.967236 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert podName:2d221857-ee77-4165-a351-ecd5fc424970 nodeName:}" failed. No retries permitted until 2026-03-13 10:22:13.967218485 +0000 UTC m=+1107.989748618 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7v927j" (UID: "2d221857-ee77-4165-a351-ecd5fc424970") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 10:22:06 crc kubenswrapper[4632]: I0313 10:22:06.674856 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556622-7428t" Mar 13 10:22:06 crc kubenswrapper[4632]: I0313 10:22:06.783414 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb5hx\" (UniqueName: \"kubernetes.io/projected/bedc1d17-f5c4-4a62-ab0c-f20a002e859b-kube-api-access-mb5hx\") pod \"bedc1d17-f5c4-4a62-ab0c-f20a002e859b\" (UID: \"bedc1d17-f5c4-4a62-ab0c-f20a002e859b\") " Mar 13 10:22:06 crc kubenswrapper[4632]: I0313 10:22:06.783849 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:22:06 crc kubenswrapper[4632]: I0313 10:22:06.783886 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:22:06 crc kubenswrapper[4632]: E0313 10:22:06.784019 4632 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 10:22:06 crc kubenswrapper[4632]: E0313 10:22:06.784076 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs podName:3fdb377f-5a78-4687-82e1-50718514290d nodeName:}" failed. No retries permitted until 2026-03-13 10:22:14.784059232 +0000 UTC m=+1108.806589365 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs") pod "openstack-operator-controller-manager-85c677895b-thbc4" (UID: "3fdb377f-5a78-4687-82e1-50718514290d") : secret "webhook-server-cert" not found Mar 13 10:22:06 crc kubenswrapper[4632]: E0313 10:22:06.784286 4632 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 10:22:06 crc kubenswrapper[4632]: E0313 10:22:06.784332 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs podName:3fdb377f-5a78-4687-82e1-50718514290d nodeName:}" failed. No retries permitted until 2026-03-13 10:22:14.784319668 +0000 UTC m=+1108.806849801 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs") pod "openstack-operator-controller-manager-85c677895b-thbc4" (UID: "3fdb377f-5a78-4687-82e1-50718514290d") : secret "metrics-server-cert" not found Mar 13 10:22:06 crc kubenswrapper[4632]: I0313 10:22:06.808298 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bedc1d17-f5c4-4a62-ab0c-f20a002e859b-kube-api-access-mb5hx" (OuterVolumeSpecName: "kube-api-access-mb5hx") pod "bedc1d17-f5c4-4a62-ab0c-f20a002e859b" (UID: "bedc1d17-f5c4-4a62-ab0c-f20a002e859b"). InnerVolumeSpecName "kube-api-access-mb5hx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:22:06 crc kubenswrapper[4632]: I0313 10:22:06.885522 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb5hx\" (UniqueName: \"kubernetes.io/projected/bedc1d17-f5c4-4a62-ab0c-f20a002e859b-kube-api-access-mb5hx\") on node \"crc\" DevicePath \"\"" Mar 13 10:22:07 crc kubenswrapper[4632]: I0313 10:22:07.175927 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556622-7428t" event={"ID":"bedc1d17-f5c4-4a62-ab0c-f20a002e859b","Type":"ContainerDied","Data":"235384aa6e3de087f78098e5ca2543f682f052a84f878bf4d06b8cb3d539f270"} Mar 13 10:22:07 crc kubenswrapper[4632]: I0313 10:22:07.175984 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="235384aa6e3de087f78098e5ca2543f682f052a84f878bf4d06b8cb3d539f270" Mar 13 10:22:07 crc kubenswrapper[4632]: I0313 10:22:07.176056 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556622-7428t" Mar 13 10:22:07 crc kubenswrapper[4632]: I0313 10:22:07.734315 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556616-8xbbs"] Mar 13 10:22:07 crc kubenswrapper[4632]: I0313 10:22:07.738901 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556616-8xbbs"] Mar 13 10:22:08 crc kubenswrapper[4632]: I0313 10:22:08.081667 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c21d462b-89d1-4844-9bfc-3f0cdf7727e9" path="/var/lib/kubelet/pods/c21d462b-89d1-4844-9bfc-3f0cdf7727e9/volumes" Mar 13 10:22:13 crc kubenswrapper[4632]: I0313 10:22:13.476445 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert\") pod \"infra-operator-controller-manager-5995f4446f-flfxh\" (UID: \"1542a9c8-92f6-4bc9-8231-829f649b0b8f\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:22:13 crc kubenswrapper[4632]: I0313 10:22:13.481166 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1542a9c8-92f6-4bc9-8231-829f649b0b8f-cert\") pod \"infra-operator-controller-manager-5995f4446f-flfxh\" (UID: \"1542a9c8-92f6-4bc9-8231-829f649b0b8f\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:22:13 crc kubenswrapper[4632]: I0313 10:22:13.669839 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:22:13 crc kubenswrapper[4632]: I0313 10:22:13.982432 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7v927j\" (UID: \"2d221857-ee77-4165-a351-ecd5fc424970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:22:13 crc kubenswrapper[4632]: I0313 10:22:13.985962 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2d221857-ee77-4165-a351-ecd5fc424970-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7v927j\" (UID: \"2d221857-ee77-4165-a351-ecd5fc424970\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:22:14 crc kubenswrapper[4632]: I0313 10:22:14.277240 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:22:14 crc kubenswrapper[4632]: E0313 10:22:14.795413 4632 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 10:22:14 crc kubenswrapper[4632]: I0313 10:22:14.796076 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:22:14 crc kubenswrapper[4632]: E0313 10:22:14.796126 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs podName:3fdb377f-5a78-4687-82e1-50718514290d nodeName:}" failed. No retries permitted until 2026-03-13 10:22:30.796106325 +0000 UTC m=+1124.818636458 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs") pod "openstack-operator-controller-manager-85c677895b-thbc4" (UID: "3fdb377f-5a78-4687-82e1-50718514290d") : secret "metrics-server-cert" not found Mar 13 10:22:14 crc kubenswrapper[4632]: I0313 10:22:14.796179 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:22:14 crc kubenswrapper[4632]: I0313 10:22:14.814675 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-webhook-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:22:15 crc kubenswrapper[4632]: E0313 10:22:15.304270 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:d9bffb59bb7f9f0a6cb103c3986fd2c1bdb13ce6349c39427a690858cbd754d6" Mar 13 10:22:15 crc kubenswrapper[4632]: E0313 10:22:15.304443 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:d9bffb59bb7f9f0a6cb103c3986fd2c1bdb13ce6349c39427a690858cbd754d6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pzvbs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-6d9d6b584d-2rv7s_openstack-operators(9a963f9c-ac58-4e21-abfa-fca1279a192d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:22:15 crc kubenswrapper[4632]: E0313 10:22:15.307593 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" podUID="9a963f9c-ac58-4e21-abfa-fca1279a192d" Mar 13 10:22:16 crc kubenswrapper[4632]: E0313 10:22:16.268745 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:d9bffb59bb7f9f0a6cb103c3986fd2c1bdb13ce6349c39427a690858cbd754d6\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" podUID="9a963f9c-ac58-4e21-abfa-fca1279a192d" Mar 13 10:22:16 crc kubenswrapper[4632]: E0313 10:22:16.880048 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721" Mar 13 10:22:16 crc kubenswrapper[4632]: E0313 10:22:16.880261 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dl4tq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-776c5696bf-bkmbn_openstack-operators(c33d0da9-5a04-42d6-80d3-2f558b4a90b0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:22:16 crc kubenswrapper[4632]: E0313 10:22:16.881499 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn" podUID="c33d0da9-5a04-42d6-80d3-2f558b4a90b0" Mar 13 10:22:17 crc kubenswrapper[4632]: E0313 10:22:17.270893 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn" podUID="c33d0da9-5a04-42d6-80d3-2f558b4a90b0" Mar 13 10:22:18 crc kubenswrapper[4632]: E0313 10:22:18.697438 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571" Mar 13 10:22:18 crc kubenswrapper[4632]: E0313 10:22:18.697604 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zsp6d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-5f4f55cb5c-62gpm_openstack-operators(9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:22:18 crc kubenswrapper[4632]: E0313 10:22:18.700511 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" podUID="9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5" Mar 13 10:22:19 crc kubenswrapper[4632]: E0313 10:22:19.282320 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" podUID="9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5" Mar 13 10:22:19 crc kubenswrapper[4632]: E0313 10:22:19.435807 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ironic-operator@sha256:9182d1816c6fdb093d6328f1b0bf39296b9eccfa495f35e2198ec4764fa6288f" Mar 13 10:22:19 crc kubenswrapper[4632]: E0313 10:22:19.435997 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ironic-operator@sha256:9182d1816c6fdb093d6328f1b0bf39296b9eccfa495f35e2198ec4764fa6288f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-psrhj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ironic-operator-controller-manager-6bbb499bbc-wtzrw_openstack-operators(c8fc6f03-c43b-4ade-92a8-acc5537a4eeb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:22:19 crc kubenswrapper[4632]: E0313 10:22:19.437162 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" podUID="c8fc6f03-c43b-4ade-92a8-acc5537a4eeb" Mar 13 10:22:20 crc kubenswrapper[4632]: E0313 10:22:20.290078 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ironic-operator@sha256:9182d1816c6fdb093d6328f1b0bf39296b9eccfa495f35e2198ec4764fa6288f\\\"\"" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" podUID="c8fc6f03-c43b-4ade-92a8-acc5537a4eeb" Mar 13 10:22:21 crc kubenswrapper[4632]: E0313 10:22:21.313570 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f" Mar 13 10:22:21 crc kubenswrapper[4632]: E0313 10:22:21.314587 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zsqv6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-bbc5b68f9-4m8kf_openstack-operators(0a9d48f4-d68b-4ef9-826e-ed619c761405): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:22:21 crc kubenswrapper[4632]: E0313 10:22:21.315832 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" podUID="0a9d48f4-d68b-4ef9-826e-ed619c761405" Mar 13 10:22:21 crc kubenswrapper[4632]: E0313 10:22:21.841703 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6" Mar 13 10:22:21 crc kubenswrapper[4632]: E0313 10:22:21.841888 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-q9mvv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-77b6666d85-cgh6c_openstack-operators(ff6d4dcb-9eb8-44fc-951e-f2aecd77a639): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:22:21 crc kubenswrapper[4632]: E0313 10:22:21.843889 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" podUID="ff6d4dcb-9eb8-44fc-951e-f2aecd77a639" Mar 13 10:22:22 crc kubenswrapper[4632]: E0313 10:22:22.305383 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:6c9aef12f50be0b974f5e35b0d69303e7f7b95e6db5d41bcdb2d9d1100e921a6\\\"\"" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" podUID="ff6d4dcb-9eb8-44fc-951e-f2aecd77a639" Mar 13 10:22:22 crc kubenswrapper[4632]: E0313 10:22:22.305777 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" podUID="0a9d48f4-d68b-4ef9-826e-ed619c761405" Mar 13 10:22:22 crc kubenswrapper[4632]: E0313 10:22:22.547855 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:571f369855b0891a2b14e54a4c1c5ae2fbbd5de4c8fddd48e81033aad4b26423" Mar 13 10:22:22 crc kubenswrapper[4632]: E0313 10:22:22.548048 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:571f369855b0891a2b14e54a4c1c5ae2fbbd5de4c8fddd48e81033aad4b26423,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-96kl7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-677bd678f7-wj9qs_openstack-operators(68c5eb80-4214-42c5-a08d-de6012969621): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:22:22 crc kubenswrapper[4632]: E0313 10:22:22.549263 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" podUID="68c5eb80-4214-42c5-a08d-de6012969621" Mar 13 10:22:23 crc kubenswrapper[4632]: E0313 10:22:23.309645 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:571f369855b0891a2b14e54a4c1c5ae2fbbd5de4c8fddd48e81033aad4b26423\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" podUID="68c5eb80-4214-42c5-a08d-de6012969621" Mar 13 10:22:24 crc kubenswrapper[4632]: E0313 10:22:24.663372 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Mar 13 10:22:24 crc kubenswrapper[4632]: E0313 10:22:24.663538 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dp7js,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-2lzt8_openstack-operators(daba1153-3b28-4234-8dd0-ec20160abbfe): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:22:24 crc kubenswrapper[4632]: E0313 10:22:24.665549 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2lzt8" podUID="daba1153-3b28-4234-8dd0-ec20160abbfe" Mar 13 10:22:25 crc kubenswrapper[4632]: E0313 10:22:25.320369 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2lzt8" podUID="daba1153-3b28-4234-8dd0-ec20160abbfe" Mar 13 10:22:25 crc kubenswrapper[4632]: I0313 10:22:25.917565 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh"] Mar 13 10:22:25 crc kubenswrapper[4632]: W0313 10:22:25.963358 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1542a9c8_92f6_4bc9_8231_829f649b0b8f.slice/crio-b11fbfeed73c462f309b92701ff644296a5764bdfe9b7decd3c8c75dab49c807 WatchSource:0}: Error finding container b11fbfeed73c462f309b92701ff644296a5764bdfe9b7decd3c8c75dab49c807: Status 404 returned error can't find the container with id b11fbfeed73c462f309b92701ff644296a5764bdfe9b7decd3c8c75dab49c807 Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.029288 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j"] Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.326787 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d" event={"ID":"7b491335-6a73-46de-8098-f27ff4c6f795","Type":"ContainerStarted","Data":"bee39db93b042daffd1b30df6ba6bf123a3260d5b83250e53a8fd157f3c524e7"} Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.327270 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.328214 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" event={"ID":"f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda","Type":"ContainerStarted","Data":"b79dde4b0109a751bfba6b9882a550b5aaf0de838fae99b2eeecdc581770755b"} Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.328963 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.329582 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" event={"ID":"ee081327-4c3f-4c0a-9085-71085c6487b5","Type":"ContainerStarted","Data":"67deb5497e0835791a752b16658e0c0dc0da7d1e14342650d8635b85be9aef64"} Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.329923 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.331310 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c" event={"ID":"9040a0e0-2a56-4331-ba50-b19ff05ef0c0","Type":"ContainerStarted","Data":"6c302ccc4aed3f78a784b9cb4fb4353ddd24cf9520fc3acd91084005d77482dd"} Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.331705 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.333780 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" event={"ID":"d04e9aa6-f234-4ffa-81e2-1a2407addb77","Type":"ContainerStarted","Data":"c892cdb883fb6ba8b138b2e24e260c237c4f7bb1b18836e7c594d3bacb5fb705"} Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.334000 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.335084 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" event={"ID":"1542a9c8-92f6-4bc9-8231-829f649b0b8f","Type":"ContainerStarted","Data":"b11fbfeed73c462f309b92701ff644296a5764bdfe9b7decd3c8c75dab49c807"} Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.337226 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" event={"ID":"e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5","Type":"ContainerStarted","Data":"6f0852ca13597fb9fc03773f58fd247d53043d5908672daa18e3613508f67bbf"} Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.337454 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.338588 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn" event={"ID":"75d652c7-8521-4039-913a-fa625f89b094","Type":"ContainerStarted","Data":"83cb70749e54e22ecd698ccb4729792c40e2011f55f7d20f7370432c8a980067"} Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.338735 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.339781 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" event={"ID":"7bab78c8-7dac-48dc-a426-ccd4ae00a428","Type":"ContainerStarted","Data":"b3fcfa0533b27651d8f7065fb7e5b6efa7d64fd553b96209205f07ed6ffdad34"} Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.340361 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.341928 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87" event={"ID":"3f3a462e-4d89-45b3-8611-181aca5f8558","Type":"ContainerStarted","Data":"9afae6bee180d74fb018ae3b7e7ec98295c2bc4a1335f0f9701be32d94102d7d"} Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.342344 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.343679 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" event={"ID":"e0d1d349-d63d-498b-ae15-3121f9ae73f8","Type":"ContainerStarted","Data":"0bec7dcb8f757d70baa9601d9b792a7b74e7b77c5edaedad986391dec0f864b4"} Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.344033 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.345433 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-677c674df7-qbfg2" event={"ID":"2d8a9f3a-6631-4c1e-8381-3bc313837ca0","Type":"ContainerStarted","Data":"ca77cf3221a73d98b6d316f0f4b658716c276f2431b18768ebe579b4b936ce38"} Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.345763 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-677c674df7-qbfg2" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.348069 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l" event={"ID":"20f92131-aca4-41ea-9144-a23bd9216f49","Type":"ContainerStarted","Data":"27e7fda6eecb21101ee5ce1ea41810a19ec88a60ab7e4cd48cde9037921e12a2"} Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.348210 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.349430 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" event={"ID":"2d221857-ee77-4165-a351-ecd5fc424970","Type":"ContainerStarted","Data":"f8df7e9acd4d145a93e181d8d0cc2a801323ffcbd2475c34f324413423693526"} Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.378284 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d" podStartSLOduration=5.138984984 podStartE2EDuration="29.378263588s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:21:59.99174984 +0000 UTC m=+1094.014279973" lastFinishedPulling="2026-03-13 10:22:24.231028444 +0000 UTC m=+1118.253558577" observedRunningTime="2026-03-13 10:22:26.37796464 +0000 UTC m=+1120.400494773" watchObservedRunningTime="2026-03-13 10:22:26.378263588 +0000 UTC m=+1120.400793721" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.411636 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" podStartSLOduration=4.089249426 podStartE2EDuration="29.411618624s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:22:00.330913518 +0000 UTC m=+1094.353443651" lastFinishedPulling="2026-03-13 10:22:25.653282716 +0000 UTC m=+1119.675812849" observedRunningTime="2026-03-13 10:22:26.405964049 +0000 UTC m=+1120.428494202" watchObservedRunningTime="2026-03-13 10:22:26.411618624 +0000 UTC m=+1120.434148757" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.485430 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn" podStartSLOduration=5.116452755 podStartE2EDuration="29.485408349s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:21:59.811537652 +0000 UTC m=+1093.834067785" lastFinishedPulling="2026-03-13 10:22:24.180493246 +0000 UTC m=+1118.203023379" observedRunningTime="2026-03-13 10:22:26.482877928 +0000 UTC m=+1120.505408081" watchObservedRunningTime="2026-03-13 10:22:26.485408349 +0000 UTC m=+1120.507938482" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.487508 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l" podStartSLOduration=5.111995689 podStartE2EDuration="29.487488099s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:21:59.804998906 +0000 UTC m=+1093.827529049" lastFinishedPulling="2026-03-13 10:22:24.180491336 +0000 UTC m=+1118.203021459" observedRunningTime="2026-03-13 10:22:26.442513023 +0000 UTC m=+1120.465043166" watchObservedRunningTime="2026-03-13 10:22:26.487488099 +0000 UTC m=+1120.510018232" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.510680 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" podStartSLOduration=4.140230447 podStartE2EDuration="29.510656762s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:22:00.323160284 +0000 UTC m=+1094.345690417" lastFinishedPulling="2026-03-13 10:22:25.693586599 +0000 UTC m=+1119.716116732" observedRunningTime="2026-03-13 10:22:26.498501231 +0000 UTC m=+1120.521031374" watchObservedRunningTime="2026-03-13 10:22:26.510656762 +0000 UTC m=+1120.533186895" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.536674 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" podStartSLOduration=4.209780851 podStartE2EDuration="29.536649384s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:22:00.323127443 +0000 UTC m=+1094.345657576" lastFinishedPulling="2026-03-13 10:22:25.649995986 +0000 UTC m=+1119.672526109" observedRunningTime="2026-03-13 10:22:26.532022183 +0000 UTC m=+1120.554552326" watchObservedRunningTime="2026-03-13 10:22:26.536649384 +0000 UTC m=+1120.559179517" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.632752 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-677c674df7-qbfg2" podStartSLOduration=5.73655299 podStartE2EDuration="29.632726141s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:22:00.284458628 +0000 UTC m=+1094.306988761" lastFinishedPulling="2026-03-13 10:22:24.180631779 +0000 UTC m=+1118.203161912" observedRunningTime="2026-03-13 10:22:26.631113562 +0000 UTC m=+1120.653643695" watchObservedRunningTime="2026-03-13 10:22:26.632726141 +0000 UTC m=+1120.655256274" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.632984 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" podStartSLOduration=5.348263786 podStartE2EDuration="29.632976136s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:21:59.945397982 +0000 UTC m=+1093.967928115" lastFinishedPulling="2026-03-13 10:22:24.230110332 +0000 UTC m=+1118.252640465" observedRunningTime="2026-03-13 10:22:26.573739551 +0000 UTC m=+1120.596269684" watchObservedRunningTime="2026-03-13 10:22:26.632976136 +0000 UTC m=+1120.655506279" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.728819 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" podStartSLOduration=4.42018119 podStartE2EDuration="29.728801688s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:22:00.373173529 +0000 UTC m=+1094.395703662" lastFinishedPulling="2026-03-13 10:22:25.681794027 +0000 UTC m=+1119.704324160" observedRunningTime="2026-03-13 10:22:26.685173455 +0000 UTC m=+1120.707703608" watchObservedRunningTime="2026-03-13 10:22:26.728801688 +0000 UTC m=+1120.751331821" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.729634 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87" podStartSLOduration=4.455868523 podStartE2EDuration="29.729626907s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:21:58.957204968 +0000 UTC m=+1092.979735101" lastFinishedPulling="2026-03-13 10:22:24.230963352 +0000 UTC m=+1118.253493485" observedRunningTime="2026-03-13 10:22:26.725890988 +0000 UTC m=+1120.748421131" watchObservedRunningTime="2026-03-13 10:22:26.729626907 +0000 UTC m=+1120.752157050" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.775521 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c" podStartSLOduration=5.402674949 podStartE2EDuration="29.775498444s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:21:59.807902536 +0000 UTC m=+1093.830432669" lastFinishedPulling="2026-03-13 10:22:24.180726031 +0000 UTC m=+1118.203256164" observedRunningTime="2026-03-13 10:22:26.770139855 +0000 UTC m=+1120.792669988" watchObservedRunningTime="2026-03-13 10:22:26.775498444 +0000 UTC m=+1120.798028587" Mar 13 10:22:26 crc kubenswrapper[4632]: I0313 10:22:26.801007 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" podStartSLOduration=4.448190179 podStartE2EDuration="29.800983803s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:22:00.297223063 +0000 UTC m=+1094.319753186" lastFinishedPulling="2026-03-13 10:22:25.650016677 +0000 UTC m=+1119.672546810" observedRunningTime="2026-03-13 10:22:26.79708495 +0000 UTC m=+1120.819615083" watchObservedRunningTime="2026-03-13 10:22:26.800983803 +0000 UTC m=+1120.823513956" Mar 13 10:22:30 crc kubenswrapper[4632]: I0313 10:22:30.873889 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:22:30 crc kubenswrapper[4632]: I0313 10:22:30.881774 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/3fdb377f-5a78-4687-82e1-50718514290d-metrics-certs\") pod \"openstack-operator-controller-manager-85c677895b-thbc4\" (UID: \"3fdb377f-5a78-4687-82e1-50718514290d\") " pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:22:31 crc kubenswrapper[4632]: I0313 10:22:31.091890 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:22:31 crc kubenswrapper[4632]: I0313 10:22:31.559379 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4"] Mar 13 10:22:32 crc kubenswrapper[4632]: I0313 10:22:32.407421 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" event={"ID":"3fdb377f-5a78-4687-82e1-50718514290d","Type":"ContainerStarted","Data":"8a9acb03a35bd7ffb9f9f29ea9128fffa2172945336e93c38861910e213e1a08"} Mar 13 10:22:32 crc kubenswrapper[4632]: I0313 10:22:32.407748 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" event={"ID":"3fdb377f-5a78-4687-82e1-50718514290d","Type":"ContainerStarted","Data":"f8dc92a47be109899fa8b210d0bf31f4e2fade68a0378a4148b8c4c29ad21bf0"} Mar 13 10:22:33 crc kubenswrapper[4632]: I0313 10:22:33.414028 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:22:35 crc kubenswrapper[4632]: I0313 10:22:35.062124 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" podStartSLOduration=37.062106769 podStartE2EDuration="37.062106769s" podCreationTimestamp="2026-03-13 10:21:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:22:33.452702793 +0000 UTC m=+1127.475232946" watchObservedRunningTime="2026-03-13 10:22:35.062106769 +0000 UTC m=+1129.084636902" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.444996 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" event={"ID":"ff6d4dcb-9eb8-44fc-951e-f2aecd77a639","Type":"ContainerStarted","Data":"45408b4b3380104d170f7dbf124450f3cb3dfa6814358710ecd0eeefe0ba8dba"} Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.446179 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.447689 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" event={"ID":"9a963f9c-ac58-4e21-abfa-fca1279a192d","Type":"ContainerStarted","Data":"5da171cf0f040b8c46b8f4cecc4cb7071dfd001352c9956b9ace45b281cc0e4d"} Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.448300 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.450553 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" event={"ID":"2d221857-ee77-4165-a351-ecd5fc424970","Type":"ContainerStarted","Data":"4bfe620f039be61e9b61330d9268e6476def1f3cae727bb56906984d0ef2ee19"} Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.450768 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.452675 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" event={"ID":"1542a9c8-92f6-4bc9-8231-829f649b0b8f","Type":"ContainerStarted","Data":"9eb54935376ed0fc9a365f9c82f2f0039889e05f8d23c20aab04b22f4c557c96"} Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.453470 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.458040 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn" event={"ID":"c33d0da9-5a04-42d6-80d3-2f558b4a90b0","Type":"ContainerStarted","Data":"45b2faabcb131c3fc71cefac2882855c661528a769ac150dbda2d0db01b89c2f"} Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.458775 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.459909 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" event={"ID":"c8fc6f03-c43b-4ade-92a8-acc5537a4eeb","Type":"ContainerStarted","Data":"ff2ebee46d2497391bf84217958c1331ef2b33747c210e706193cc32df47b16b"} Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.460396 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.461502 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" event={"ID":"9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5","Type":"ContainerStarted","Data":"97199dbb17e5b26ac45577f72ea8130835efdb3aa71d1d178692d4af3ee8d824"} Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.461996 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.463151 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" event={"ID":"0a9d48f4-d68b-4ef9-826e-ed619c761405","Type":"ContainerStarted","Data":"cf1a08befb3cdc6b962932151cafb742b4042bea18d08fd1d4b1d27bc01a5ba1"} Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.463617 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.483791 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" podStartSLOduration=4.271047755 podStartE2EDuration="40.483772803s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:21:59.945375542 +0000 UTC m=+1093.967905675" lastFinishedPulling="2026-03-13 10:22:36.15810059 +0000 UTC m=+1130.180630723" observedRunningTime="2026-03-13 10:22:37.474960833 +0000 UTC m=+1131.497490966" watchObservedRunningTime="2026-03-13 10:22:37.483772803 +0000 UTC m=+1131.506302956" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.571178 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" podStartSLOduration=4.122712669 podStartE2EDuration="40.571154133s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:22:00.271255833 +0000 UTC m=+1094.293785966" lastFinishedPulling="2026-03-13 10:22:36.719697297 +0000 UTC m=+1130.742227430" observedRunningTime="2026-03-13 10:22:37.545326275 +0000 UTC m=+1131.567856408" watchObservedRunningTime="2026-03-13 10:22:37.571154133 +0000 UTC m=+1131.593684276" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.625548 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.640971 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" podStartSLOduration=4.412925336 podStartE2EDuration="40.64093057s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:21:59.928227802 +0000 UTC m=+1093.950757935" lastFinishedPulling="2026-03-13 10:22:36.156233026 +0000 UTC m=+1130.178763169" observedRunningTime="2026-03-13 10:22:37.612131552 +0000 UTC m=+1131.634661685" watchObservedRunningTime="2026-03-13 10:22:37.64093057 +0000 UTC m=+1131.663460703" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.641794 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" podStartSLOduration=30.454470456 podStartE2EDuration="40.641786291s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:22:25.96885446 +0000 UTC m=+1119.991384593" lastFinishedPulling="2026-03-13 10:22:36.156170295 +0000 UTC m=+1130.178700428" observedRunningTime="2026-03-13 10:22:37.6346361 +0000 UTC m=+1131.657166233" watchObservedRunningTime="2026-03-13 10:22:37.641786291 +0000 UTC m=+1131.664316424" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.653698 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.673061 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" podStartSLOduration=4.320086827 podStartE2EDuration="40.673040408s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:21:59.804932244 +0000 UTC m=+1093.827462377" lastFinishedPulling="2026-03-13 10:22:36.157885825 +0000 UTC m=+1130.180415958" observedRunningTime="2026-03-13 10:22:37.664434362 +0000 UTC m=+1131.686964515" watchObservedRunningTime="2026-03-13 10:22:37.673040408 +0000 UTC m=+1131.695570551" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.682573 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.703810 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn" podStartSLOduration=4.485414439 podStartE2EDuration="40.703791463s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:21:59.945462594 +0000 UTC m=+1093.967992727" lastFinishedPulling="2026-03-13 10:22:36.163839608 +0000 UTC m=+1130.186369751" observedRunningTime="2026-03-13 10:22:37.698092647 +0000 UTC m=+1131.720622780" watchObservedRunningTime="2026-03-13 10:22:37.703791463 +0000 UTC m=+1131.726321596" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.797632 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" podStartSLOduration=3.871204305 podStartE2EDuration="40.797615526s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:21:59.229813355 +0000 UTC m=+1093.252343488" lastFinishedPulling="2026-03-13 10:22:36.156224576 +0000 UTC m=+1130.178754709" observedRunningTime="2026-03-13 10:22:37.793047797 +0000 UTC m=+1131.815577930" watchObservedRunningTime="2026-03-13 10:22:37.797615526 +0000 UTC m=+1131.820145659" Mar 13 10:22:37 crc kubenswrapper[4632]: I0313 10:22:37.797809 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" podStartSLOduration=30.747011571 podStartE2EDuration="40.797805601s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:22:26.105445996 +0000 UTC m=+1120.127976129" lastFinishedPulling="2026-03-13 10:22:36.156240026 +0000 UTC m=+1130.178770159" observedRunningTime="2026-03-13 10:22:37.765662233 +0000 UTC m=+1131.788192366" watchObservedRunningTime="2026-03-13 10:22:37.797805601 +0000 UTC m=+1131.820335734" Mar 13 10:22:38 crc kubenswrapper[4632]: I0313 10:22:38.001146 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c" Mar 13 10:22:38 crc kubenswrapper[4632]: I0313 10:22:38.189889 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" Mar 13 10:22:38 crc kubenswrapper[4632]: I0313 10:22:38.238532 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d" Mar 13 10:22:38 crc kubenswrapper[4632]: I0313 10:22:38.332035 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" Mar 13 10:22:38 crc kubenswrapper[4632]: I0313 10:22:38.470612 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" event={"ID":"68c5eb80-4214-42c5-a08d-de6012969621","Type":"ContainerStarted","Data":"579d286b9eb7e56fb8f1cb6d18127cc0ece5c920fbbbc7e2c67943e4800bb183"} Mar 13 10:22:38 crc kubenswrapper[4632]: I0313 10:22:38.504894 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" podStartSLOduration=3.244169395 podStartE2EDuration="41.504872785s" podCreationTimestamp="2026-03-13 10:21:57 +0000 UTC" firstStartedPulling="2026-03-13 10:21:59.218160496 +0000 UTC m=+1093.240690629" lastFinishedPulling="2026-03-13 10:22:37.478863886 +0000 UTC m=+1131.501394019" observedRunningTime="2026-03-13 10:22:38.501107125 +0000 UTC m=+1132.523637258" watchObservedRunningTime="2026-03-13 10:22:38.504872785 +0000 UTC m=+1132.527402918" Mar 13 10:22:38 crc kubenswrapper[4632]: I0313 10:22:38.584082 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" Mar 13 10:22:38 crc kubenswrapper[4632]: I0313 10:22:38.766887 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-677c674df7-qbfg2" Mar 13 10:22:38 crc kubenswrapper[4632]: I0313 10:22:38.864755 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" Mar 13 10:22:39 crc kubenswrapper[4632]: I0313 10:22:39.067082 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" Mar 13 10:22:39 crc kubenswrapper[4632]: I0313 10:22:39.110992 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" Mar 13 10:22:40 crc kubenswrapper[4632]: I0313 10:22:40.484685 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2lzt8" event={"ID":"daba1153-3b28-4234-8dd0-ec20160abbfe","Type":"ContainerStarted","Data":"28d68b5d61b4f271fabf0662eb66dc8da8eb38d3aab1b5660194b1a0aa44a4b3"} Mar 13 10:22:40 crc kubenswrapper[4632]: I0313 10:22:40.504024 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-2lzt8" podStartSLOduration=3.042150061 podStartE2EDuration="42.504008397s" podCreationTimestamp="2026-03-13 10:21:58 +0000 UTC" firstStartedPulling="2026-03-13 10:22:00.192875129 +0000 UTC m=+1094.215405262" lastFinishedPulling="2026-03-13 10:22:39.654733465 +0000 UTC m=+1133.677263598" observedRunningTime="2026-03-13 10:22:40.500041843 +0000 UTC m=+1134.522572006" watchObservedRunningTime="2026-03-13 10:22:40.504008397 +0000 UTC m=+1134.526538530" Mar 13 10:22:41 crc kubenswrapper[4632]: I0313 10:22:41.099914 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" Mar 13 10:22:43 crc kubenswrapper[4632]: I0313 10:22:43.676382 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" Mar 13 10:22:44 crc kubenswrapper[4632]: I0313 10:22:44.284107 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" Mar 13 10:22:47 crc kubenswrapper[4632]: I0313 10:22:47.602832 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" Mar 13 10:22:47 crc kubenswrapper[4632]: I0313 10:22:47.606015 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" Mar 13 10:22:47 crc kubenswrapper[4632]: I0313 10:22:47.705179 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" Mar 13 10:22:47 crc kubenswrapper[4632]: I0313 10:22:47.771475 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" Mar 13 10:22:47 crc kubenswrapper[4632]: I0313 10:22:47.821227 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" Mar 13 10:22:48 crc kubenswrapper[4632]: I0313 10:22:48.287199 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn" Mar 13 10:22:48 crc kubenswrapper[4632]: I0313 10:22:48.388353 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" Mar 13 10:22:48 crc kubenswrapper[4632]: I0313 10:22:48.504118 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" Mar 13 10:22:58 crc kubenswrapper[4632]: I0313 10:22:58.663182 4632 scope.go:117] "RemoveContainer" containerID="6a34c241348123944aa499915ed71c016789c868e3e563c2a1cb71763ed56ad8" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.189633 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b95c5c449-th4fn"] Mar 13 10:23:06 crc kubenswrapper[4632]: E0313 10:23:06.190439 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bedc1d17-f5c4-4a62-ab0c-f20a002e859b" containerName="oc" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.190453 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bedc1d17-f5c4-4a62-ab0c-f20a002e859b" containerName="oc" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.190587 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="bedc1d17-f5c4-4a62-ab0c-f20a002e859b" containerName="oc" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.191245 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b95c5c449-th4fn" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.204722 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-6rt6f" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.204956 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.205149 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.205289 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.222650 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b95c5c449-th4fn"] Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.240373 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c54n6\" (UniqueName: \"kubernetes.io/projected/dcb4500d-7a53-4091-b3af-394eb0f49130-kube-api-access-c54n6\") pod \"dnsmasq-dns-7b95c5c449-th4fn\" (UID: \"dcb4500d-7a53-4091-b3af-394eb0f49130\") " pod="openstack/dnsmasq-dns-7b95c5c449-th4fn" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.240442 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcb4500d-7a53-4091-b3af-394eb0f49130-config\") pod \"dnsmasq-dns-7b95c5c449-th4fn\" (UID: \"dcb4500d-7a53-4091-b3af-394eb0f49130\") " pod="openstack/dnsmasq-dns-7b95c5c449-th4fn" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.276725 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bd9cf7445-frlmw"] Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.277920 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.280520 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.294052 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bd9cf7445-frlmw"] Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.342186 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4mmhg\" (UniqueName: \"kubernetes.io/projected/da131e9a-8968-4569-a970-3aa95b2a830b-kube-api-access-4mmhg\") pod \"dnsmasq-dns-bd9cf7445-frlmw\" (UID: \"da131e9a-8968-4569-a970-3aa95b2a830b\") " pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.342263 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da131e9a-8968-4569-a970-3aa95b2a830b-config\") pod \"dnsmasq-dns-bd9cf7445-frlmw\" (UID: \"da131e9a-8968-4569-a970-3aa95b2a830b\") " pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.342301 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c54n6\" (UniqueName: \"kubernetes.io/projected/dcb4500d-7a53-4091-b3af-394eb0f49130-kube-api-access-c54n6\") pod \"dnsmasq-dns-7b95c5c449-th4fn\" (UID: \"dcb4500d-7a53-4091-b3af-394eb0f49130\") " pod="openstack/dnsmasq-dns-7b95c5c449-th4fn" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.342368 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da131e9a-8968-4569-a970-3aa95b2a830b-dns-svc\") pod \"dnsmasq-dns-bd9cf7445-frlmw\" (UID: \"da131e9a-8968-4569-a970-3aa95b2a830b\") " pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.342407 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcb4500d-7a53-4091-b3af-394eb0f49130-config\") pod \"dnsmasq-dns-7b95c5c449-th4fn\" (UID: \"dcb4500d-7a53-4091-b3af-394eb0f49130\") " pod="openstack/dnsmasq-dns-7b95c5c449-th4fn" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.343788 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcb4500d-7a53-4091-b3af-394eb0f49130-config\") pod \"dnsmasq-dns-7b95c5c449-th4fn\" (UID: \"dcb4500d-7a53-4091-b3af-394eb0f49130\") " pod="openstack/dnsmasq-dns-7b95c5c449-th4fn" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.372354 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c54n6\" (UniqueName: \"kubernetes.io/projected/dcb4500d-7a53-4091-b3af-394eb0f49130-kube-api-access-c54n6\") pod \"dnsmasq-dns-7b95c5c449-th4fn\" (UID: \"dcb4500d-7a53-4091-b3af-394eb0f49130\") " pod="openstack/dnsmasq-dns-7b95c5c449-th4fn" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.443824 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da131e9a-8968-4569-a970-3aa95b2a830b-dns-svc\") pod \"dnsmasq-dns-bd9cf7445-frlmw\" (UID: \"da131e9a-8968-4569-a970-3aa95b2a830b\") " pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.444341 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4mmhg\" (UniqueName: \"kubernetes.io/projected/da131e9a-8968-4569-a970-3aa95b2a830b-kube-api-access-4mmhg\") pod \"dnsmasq-dns-bd9cf7445-frlmw\" (UID: \"da131e9a-8968-4569-a970-3aa95b2a830b\") " pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.444393 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da131e9a-8968-4569-a970-3aa95b2a830b-config\") pod \"dnsmasq-dns-bd9cf7445-frlmw\" (UID: \"da131e9a-8968-4569-a970-3aa95b2a830b\") " pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.444932 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da131e9a-8968-4569-a970-3aa95b2a830b-dns-svc\") pod \"dnsmasq-dns-bd9cf7445-frlmw\" (UID: \"da131e9a-8968-4569-a970-3aa95b2a830b\") " pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.445487 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da131e9a-8968-4569-a970-3aa95b2a830b-config\") pod \"dnsmasq-dns-bd9cf7445-frlmw\" (UID: \"da131e9a-8968-4569-a970-3aa95b2a830b\") " pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.469661 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4mmhg\" (UniqueName: \"kubernetes.io/projected/da131e9a-8968-4569-a970-3aa95b2a830b-kube-api-access-4mmhg\") pod \"dnsmasq-dns-bd9cf7445-frlmw\" (UID: \"da131e9a-8968-4569-a970-3aa95b2a830b\") " pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.540683 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b95c5c449-th4fn" Mar 13 10:23:06 crc kubenswrapper[4632]: I0313 10:23:06.600871 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" Mar 13 10:23:07 crc kubenswrapper[4632]: I0313 10:23:07.075682 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b95c5c449-th4fn"] Mar 13 10:23:07 crc kubenswrapper[4632]: I0313 10:23:07.158883 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bd9cf7445-frlmw"] Mar 13 10:23:07 crc kubenswrapper[4632]: W0313 10:23:07.161305 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda131e9a_8968_4569_a970_3aa95b2a830b.slice/crio-55fcbfaa372a43ed729e32593f5c37c0b24f03c99a93bb7fe4dbc16e4c07e1b0 WatchSource:0}: Error finding container 55fcbfaa372a43ed729e32593f5c37c0b24f03c99a93bb7fe4dbc16e4c07e1b0: Status 404 returned error can't find the container with id 55fcbfaa372a43ed729e32593f5c37c0b24f03c99a93bb7fe4dbc16e4c07e1b0 Mar 13 10:23:07 crc kubenswrapper[4632]: I0313 10:23:07.669547 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b95c5c449-th4fn" event={"ID":"dcb4500d-7a53-4091-b3af-394eb0f49130","Type":"ContainerStarted","Data":"795a9fce6e217788c208128d366941b235a94e772dfaab8892b09ba717741562"} Mar 13 10:23:07 crc kubenswrapper[4632]: I0313 10:23:07.675141 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" event={"ID":"da131e9a-8968-4569-a970-3aa95b2a830b","Type":"ContainerStarted","Data":"55fcbfaa372a43ed729e32593f5c37c0b24f03c99a93bb7fe4dbc16e4c07e1b0"} Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.101803 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b95c5c449-th4fn"] Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.130770 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-dcf85566c-59l8m"] Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.134135 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcf85566c-59l8m" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.146843 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dcf85566c-59l8m"] Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.204180 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea62b75b-fe31-433d-9ff1-a7333aacb383-dns-svc\") pod \"dnsmasq-dns-dcf85566c-59l8m\" (UID: \"ea62b75b-fe31-433d-9ff1-a7333aacb383\") " pod="openstack/dnsmasq-dns-dcf85566c-59l8m" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.204263 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-868r4\" (UniqueName: \"kubernetes.io/projected/ea62b75b-fe31-433d-9ff1-a7333aacb383-kube-api-access-868r4\") pod \"dnsmasq-dns-dcf85566c-59l8m\" (UID: \"ea62b75b-fe31-433d-9ff1-a7333aacb383\") " pod="openstack/dnsmasq-dns-dcf85566c-59l8m" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.204298 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea62b75b-fe31-433d-9ff1-a7333aacb383-config\") pod \"dnsmasq-dns-dcf85566c-59l8m\" (UID: \"ea62b75b-fe31-433d-9ff1-a7333aacb383\") " pod="openstack/dnsmasq-dns-dcf85566c-59l8m" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.305190 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-868r4\" (UniqueName: \"kubernetes.io/projected/ea62b75b-fe31-433d-9ff1-a7333aacb383-kube-api-access-868r4\") pod \"dnsmasq-dns-dcf85566c-59l8m\" (UID: \"ea62b75b-fe31-433d-9ff1-a7333aacb383\") " pod="openstack/dnsmasq-dns-dcf85566c-59l8m" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.305247 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea62b75b-fe31-433d-9ff1-a7333aacb383-config\") pod \"dnsmasq-dns-dcf85566c-59l8m\" (UID: \"ea62b75b-fe31-433d-9ff1-a7333aacb383\") " pod="openstack/dnsmasq-dns-dcf85566c-59l8m" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.305324 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea62b75b-fe31-433d-9ff1-a7333aacb383-dns-svc\") pod \"dnsmasq-dns-dcf85566c-59l8m\" (UID: \"ea62b75b-fe31-433d-9ff1-a7333aacb383\") " pod="openstack/dnsmasq-dns-dcf85566c-59l8m" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.306405 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea62b75b-fe31-433d-9ff1-a7333aacb383-dns-svc\") pod \"dnsmasq-dns-dcf85566c-59l8m\" (UID: \"ea62b75b-fe31-433d-9ff1-a7333aacb383\") " pod="openstack/dnsmasq-dns-dcf85566c-59l8m" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.306405 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea62b75b-fe31-433d-9ff1-a7333aacb383-config\") pod \"dnsmasq-dns-dcf85566c-59l8m\" (UID: \"ea62b75b-fe31-433d-9ff1-a7333aacb383\") " pod="openstack/dnsmasq-dns-dcf85566c-59l8m" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.346993 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-868r4\" (UniqueName: \"kubernetes.io/projected/ea62b75b-fe31-433d-9ff1-a7333aacb383-kube-api-access-868r4\") pod \"dnsmasq-dns-dcf85566c-59l8m\" (UID: \"ea62b75b-fe31-433d-9ff1-a7333aacb383\") " pod="openstack/dnsmasq-dns-dcf85566c-59l8m" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.467850 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bd9cf7445-frlmw"] Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.472510 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcf85566c-59l8m" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.525415 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86545856d7-fkxhx"] Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.527422 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86545856d7-fkxhx" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.552637 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86545856d7-fkxhx"] Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.610172 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7afad0f9-c29c-40e6-8605-1df67a505a82-config\") pod \"dnsmasq-dns-86545856d7-fkxhx\" (UID: \"7afad0f9-c29c-40e6-8605-1df67a505a82\") " pod="openstack/dnsmasq-dns-86545856d7-fkxhx" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.610224 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdf8h\" (UniqueName: \"kubernetes.io/projected/7afad0f9-c29c-40e6-8605-1df67a505a82-kube-api-access-qdf8h\") pod \"dnsmasq-dns-86545856d7-fkxhx\" (UID: \"7afad0f9-c29c-40e6-8605-1df67a505a82\") " pod="openstack/dnsmasq-dns-86545856d7-fkxhx" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.610258 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7afad0f9-c29c-40e6-8605-1df67a505a82-dns-svc\") pod \"dnsmasq-dns-86545856d7-fkxhx\" (UID: \"7afad0f9-c29c-40e6-8605-1df67a505a82\") " pod="openstack/dnsmasq-dns-86545856d7-fkxhx" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.711103 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7afad0f9-c29c-40e6-8605-1df67a505a82-config\") pod \"dnsmasq-dns-86545856d7-fkxhx\" (UID: \"7afad0f9-c29c-40e6-8605-1df67a505a82\") " pod="openstack/dnsmasq-dns-86545856d7-fkxhx" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.711159 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdf8h\" (UniqueName: \"kubernetes.io/projected/7afad0f9-c29c-40e6-8605-1df67a505a82-kube-api-access-qdf8h\") pod \"dnsmasq-dns-86545856d7-fkxhx\" (UID: \"7afad0f9-c29c-40e6-8605-1df67a505a82\") " pod="openstack/dnsmasq-dns-86545856d7-fkxhx" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.711200 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7afad0f9-c29c-40e6-8605-1df67a505a82-dns-svc\") pod \"dnsmasq-dns-86545856d7-fkxhx\" (UID: \"7afad0f9-c29c-40e6-8605-1df67a505a82\") " pod="openstack/dnsmasq-dns-86545856d7-fkxhx" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.712185 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7afad0f9-c29c-40e6-8605-1df67a505a82-dns-svc\") pod \"dnsmasq-dns-86545856d7-fkxhx\" (UID: \"7afad0f9-c29c-40e6-8605-1df67a505a82\") " pod="openstack/dnsmasq-dns-86545856d7-fkxhx" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.712477 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7afad0f9-c29c-40e6-8605-1df67a505a82-config\") pod \"dnsmasq-dns-86545856d7-fkxhx\" (UID: \"7afad0f9-c29c-40e6-8605-1df67a505a82\") " pod="openstack/dnsmasq-dns-86545856d7-fkxhx" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.747858 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdf8h\" (UniqueName: \"kubernetes.io/projected/7afad0f9-c29c-40e6-8605-1df67a505a82-kube-api-access-qdf8h\") pod \"dnsmasq-dns-86545856d7-fkxhx\" (UID: \"7afad0f9-c29c-40e6-8605-1df67a505a82\") " pod="openstack/dnsmasq-dns-86545856d7-fkxhx" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.910529 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86545856d7-fkxhx" Mar 13 10:23:09 crc kubenswrapper[4632]: I0313 10:23:09.936868 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-dcf85566c-59l8m"] Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.289620 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.292208 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.294041 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.294202 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.296823 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.297065 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.301175 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.301513 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.301718 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-m5r4h" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.303846 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.321748 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.321800 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/159c6cee-c82b-4725-82d6-dbd27216f53c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.321825 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.321876 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.321909 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.321930 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.321980 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.322305 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/159c6cee-c82b-4725-82d6-dbd27216f53c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.322331 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8hmx\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-kube-api-access-k8hmx\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.322354 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.322379 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.424147 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.424219 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.424249 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.424283 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.424361 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/159c6cee-c82b-4725-82d6-dbd27216f53c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.424382 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8hmx\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-kube-api-access-k8hmx\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.424405 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.424431 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.424485 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.424511 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/159c6cee-c82b-4725-82d6-dbd27216f53c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.424559 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.424860 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.426312 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.426631 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.426804 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.424880 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.427883 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.434248 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86545856d7-fkxhx"] Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.434986 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/159c6cee-c82b-4725-82d6-dbd27216f53c-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.448190 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/159c6cee-c82b-4725-82d6-dbd27216f53c-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.449847 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.452955 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8hmx\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-kube-api-access-k8hmx\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.464724 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.475678 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.632528 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.695846 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.701229 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.705531 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-x424t" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.705657 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.707981 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.706003 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.706132 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.706175 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.706258 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.706296 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.832861 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86545856d7-fkxhx" event={"ID":"7afad0f9-c29c-40e6-8605-1df67a505a82","Type":"ContainerStarted","Data":"99dcdc47d7e36bada4f3fc23414bc9bbd494a02612477871943ec84558819c97"} Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.835319 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcf85566c-59l8m" event={"ID":"ea62b75b-fe31-433d-9ff1-a7333aacb383","Type":"ContainerStarted","Data":"8c2e1684ce51b6e904615df8d377b14786ec45f198df25a296fb0708834a1826"} Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.839025 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.839066 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.839140 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kfgh\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-kube-api-access-6kfgh\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.839713 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.839737 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/211718f0-f29c-457b-bc2b-487bb76d4801-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.839795 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.839849 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.839905 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-config-data\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.839982 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.840031 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-server-conf\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.840060 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/211718f0-f29c-457b-bc2b-487bb76d4801-pod-info\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.940876 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.940930 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.940969 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kfgh\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-kube-api-access-6kfgh\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.941078 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.941107 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/211718f0-f29c-457b-bc2b-487bb76d4801-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.941147 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.941173 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.941197 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-config-data\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.941226 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.941441 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-server-conf\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.941458 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/211718f0-f29c-457b-bc2b-487bb76d4801-pod-info\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.941698 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.942069 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.945499 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-config-data\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.945528 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.945732 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.946092 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/211718f0-f29c-457b-bc2b-487bb76d4801-pod-info\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.946695 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-server-conf\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.949118 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.950976 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.952788 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/211718f0-f29c-457b-bc2b-487bb76d4801-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.967486 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kfgh\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-kube-api-access-6kfgh\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:10 crc kubenswrapper[4632]: I0313 10:23:10.991895 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " pod="openstack/rabbitmq-server-0" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.063241 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.502955 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 10:23:11 crc kubenswrapper[4632]: W0313 10:23:11.700009 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod159c6cee_c82b_4725_82d6_dbd27216f53c.slice/crio-06613fdc2799f04ea62de7d5a6995bb48161830d28a55edb1ede1542c640e10e WatchSource:0}: Error finding container 06613fdc2799f04ea62de7d5a6995bb48161830d28a55edb1ede1542c640e10e: Status 404 returned error can't find the container with id 06613fdc2799f04ea62de7d5a6995bb48161830d28a55edb1ede1542c640e10e Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.847279 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.852599 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.859442 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.860140 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-j7sp7" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.864850 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.865183 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.870122 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"159c6cee-c82b-4725-82d6-dbd27216f53c","Type":"ContainerStarted","Data":"06613fdc2799f04ea62de7d5a6995bb48161830d28a55edb1ede1542c640e10e"} Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.892190 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.902713 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.965482 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cb2f546-c8c5-4ec9-aba8-d3782431de10-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.965550 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2cb2f546-c8c5-4ec9-aba8-d3782431de10-kolla-config\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.965578 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2cb2f546-c8c5-4ec9-aba8-d3782431de10-config-data-default\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.965602 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qs454\" (UniqueName: \"kubernetes.io/projected/2cb2f546-c8c5-4ec9-aba8-d3782431de10-kube-api-access-qs454\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.965657 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2cb2f546-c8c5-4ec9-aba8-d3782431de10-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.965691 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cb2f546-c8c5-4ec9-aba8-d3782431de10-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.965877 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cb2f546-c8c5-4ec9-aba8-d3782431de10-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:11 crc kubenswrapper[4632]: I0313 10:23:11.965976 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.016738 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.077817 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cb2f546-c8c5-4ec9-aba8-d3782431de10-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.077873 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2cb2f546-c8c5-4ec9-aba8-d3782431de10-kolla-config\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.077898 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2cb2f546-c8c5-4ec9-aba8-d3782431de10-config-data-default\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.077924 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qs454\" (UniqueName: \"kubernetes.io/projected/2cb2f546-c8c5-4ec9-aba8-d3782431de10-kube-api-access-qs454\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.078001 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2cb2f546-c8c5-4ec9-aba8-d3782431de10-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.078040 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cb2f546-c8c5-4ec9-aba8-d3782431de10-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.078063 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cb2f546-c8c5-4ec9-aba8-d3782431de10-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.078093 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.078892 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2cb2f546-c8c5-4ec9-aba8-d3782431de10-kolla-config\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.078988 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.080286 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2cb2f546-c8c5-4ec9-aba8-d3782431de10-operator-scripts\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.080616 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2cb2f546-c8c5-4ec9-aba8-d3782431de10-config-data-generated\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.081470 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2cb2f546-c8c5-4ec9-aba8-d3782431de10-config-data-default\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.134423 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cb2f546-c8c5-4ec9-aba8-d3782431de10-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.134783 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cb2f546-c8c5-4ec9-aba8-d3782431de10-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.180342 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qs454\" (UniqueName: \"kubernetes.io/projected/2cb2f546-c8c5-4ec9-aba8-d3782431de10-kube-api-access-qs454\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.191790 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"2cb2f546-c8c5-4ec9-aba8-d3782431de10\") " pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.476713 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 10:23:12 crc kubenswrapper[4632]: I0313 10:23:12.900973 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"211718f0-f29c-457b-bc2b-487bb76d4801","Type":"ContainerStarted","Data":"fd0dcad1534e2c23d238622a824c4e32c97444e16220054d2406cb0e89183756"} Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.056655 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.058205 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.072188 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.078898 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.079187 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-rqbfd" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.079374 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.079601 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.089593 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.228876 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1761ca69-46fd-4375-af60-22b3e77c19a2-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.229300 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65kkf\" (UniqueName: \"kubernetes.io/projected/1761ca69-46fd-4375-af60-22b3e77c19a2-kube-api-access-65kkf\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.229329 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1761ca69-46fd-4375-af60-22b3e77c19a2-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.229432 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1761ca69-46fd-4375-af60-22b3e77c19a2-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.229477 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1761ca69-46fd-4375-af60-22b3e77c19a2-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.229502 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1761ca69-46fd-4375-af60-22b3e77c19a2-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.229543 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1761ca69-46fd-4375-af60-22b3e77c19a2-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.229595 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.331782 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1761ca69-46fd-4375-af60-22b3e77c19a2-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.331896 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.331961 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1761ca69-46fd-4375-af60-22b3e77c19a2-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.331996 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65kkf\" (UniqueName: \"kubernetes.io/projected/1761ca69-46fd-4375-af60-22b3e77c19a2-kube-api-access-65kkf\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.332017 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1761ca69-46fd-4375-af60-22b3e77c19a2-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.332092 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1761ca69-46fd-4375-af60-22b3e77c19a2-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.332131 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1761ca69-46fd-4375-af60-22b3e77c19a2-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.332160 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1761ca69-46fd-4375-af60-22b3e77c19a2-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.332244 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.332765 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/1761ca69-46fd-4375-af60-22b3e77c19a2-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.333318 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1761ca69-46fd-4375-af60-22b3e77c19a2-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.333643 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/1761ca69-46fd-4375-af60-22b3e77c19a2-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.333977 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/1761ca69-46fd-4375-af60-22b3e77c19a2-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.355366 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1761ca69-46fd-4375-af60-22b3e77c19a2-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.358136 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/1761ca69-46fd-4375-af60-22b3e77c19a2-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.369799 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65kkf\" (UniqueName: \"kubernetes.io/projected/1761ca69-46fd-4375-af60-22b3e77c19a2-kube-api-access-65kkf\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.421090 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.424047 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.435487 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.435681 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-vcv78" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.435785 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.458183 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.466212 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"openstack-cell1-galera-0\" (UID: \"1761ca69-46fd-4375-af60-22b3e77c19a2\") " pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.538001 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9100748-6b15-4ccf-b961-aab1135f08d1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.538054 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9100748-6b15-4ccf-b961-aab1135f08d1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.538086 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d9100748-6b15-4ccf-b961-aab1135f08d1-config-data\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.538140 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvr2h\" (UniqueName: \"kubernetes.io/projected/d9100748-6b15-4ccf-b961-aab1135f08d1-kube-api-access-pvr2h\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.538179 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d9100748-6b15-4ccf-b961-aab1135f08d1-kolla-config\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.639660 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9100748-6b15-4ccf-b961-aab1135f08d1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.640249 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9100748-6b15-4ccf-b961-aab1135f08d1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.640296 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d9100748-6b15-4ccf-b961-aab1135f08d1-config-data\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.640464 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvr2h\" (UniqueName: \"kubernetes.io/projected/d9100748-6b15-4ccf-b961-aab1135f08d1-kube-api-access-pvr2h\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.640500 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d9100748-6b15-4ccf-b961-aab1135f08d1-kolla-config\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.641555 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d9100748-6b15-4ccf-b961-aab1135f08d1-config-data\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.642691 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d9100748-6b15-4ccf-b961-aab1135f08d1-kolla-config\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.648529 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d9100748-6b15-4ccf-b961-aab1135f08d1-combined-ca-bundle\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.660479 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/d9100748-6b15-4ccf-b961-aab1135f08d1-memcached-tls-certs\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.665028 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvr2h\" (UniqueName: \"kubernetes.io/projected/d9100748-6b15-4ccf-b961-aab1135f08d1-kube-api-access-pvr2h\") pod \"memcached-0\" (UID: \"d9100748-6b15-4ccf-b961-aab1135f08d1\") " pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.732334 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.795957 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 10:23:13 crc kubenswrapper[4632]: I0313 10:23:13.925241 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2cb2f546-c8c5-4ec9-aba8-d3782431de10","Type":"ContainerStarted","Data":"d523bed9fc8debb17e0b795ccb0624f1e9b1ee50c6bf87e25c1af5e36718ba64"} Mar 13 10:23:14 crc kubenswrapper[4632]: I0313 10:23:14.501200 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 10:23:14 crc kubenswrapper[4632]: W0313 10:23:14.591303 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1761ca69_46fd_4375_af60_22b3e77c19a2.slice/crio-6eda8f9cca63eda6a98a0d70fe660cd76972356b76ae16d0745a8444a016580f WatchSource:0}: Error finding container 6eda8f9cca63eda6a98a0d70fe660cd76972356b76ae16d0745a8444a016580f: Status 404 returned error can't find the container with id 6eda8f9cca63eda6a98a0d70fe660cd76972356b76ae16d0745a8444a016580f Mar 13 10:23:14 crc kubenswrapper[4632]: I0313 10:23:14.640522 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 13 10:23:14 crc kubenswrapper[4632]: I0313 10:23:14.989962 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1761ca69-46fd-4375-af60-22b3e77c19a2","Type":"ContainerStarted","Data":"6eda8f9cca63eda6a98a0d70fe660cd76972356b76ae16d0745a8444a016580f"} Mar 13 10:23:15 crc kubenswrapper[4632]: I0313 10:23:15.014115 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"d9100748-6b15-4ccf-b961-aab1135f08d1","Type":"ContainerStarted","Data":"1cc9e1cc023d82b95b1d6c1c240824e68c8964912c3afae22861ac002424add2"} Mar 13 10:23:16 crc kubenswrapper[4632]: I0313 10:23:16.074851 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 10:23:16 crc kubenswrapper[4632]: I0313 10:23:16.076189 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 10:23:16 crc kubenswrapper[4632]: I0313 10:23:16.084971 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 10:23:16 crc kubenswrapper[4632]: I0313 10:23:16.088931 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-nrwtz" Mar 13 10:23:16 crc kubenswrapper[4632]: I0313 10:23:16.206488 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfstv\" (UniqueName: \"kubernetes.io/projected/21ce0311-ff05-4626-9663-a373ae31eb56-kube-api-access-hfstv\") pod \"kube-state-metrics-0\" (UID: \"21ce0311-ff05-4626-9663-a373ae31eb56\") " pod="openstack/kube-state-metrics-0" Mar 13 10:23:16 crc kubenswrapper[4632]: I0313 10:23:16.308406 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfstv\" (UniqueName: \"kubernetes.io/projected/21ce0311-ff05-4626-9663-a373ae31eb56-kube-api-access-hfstv\") pod \"kube-state-metrics-0\" (UID: \"21ce0311-ff05-4626-9663-a373ae31eb56\") " pod="openstack/kube-state-metrics-0" Mar 13 10:23:16 crc kubenswrapper[4632]: I0313 10:23:16.353850 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfstv\" (UniqueName: \"kubernetes.io/projected/21ce0311-ff05-4626-9663-a373ae31eb56-kube-api-access-hfstv\") pod \"kube-state-metrics-0\" (UID: \"21ce0311-ff05-4626-9663-a373ae31eb56\") " pod="openstack/kube-state-metrics-0" Mar 13 10:23:16 crc kubenswrapper[4632]: I0313 10:23:16.447109 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 10:23:16 crc kubenswrapper[4632]: I0313 10:23:16.918239 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.133192 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9kd7r"] Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.134795 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.141065 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-sgqtl" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.141289 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.141411 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.206661 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9kd7r"] Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.245585 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-c5xnp"] Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.250405 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.265042 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-c5xnp"] Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.274895 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eab798dd-482a-4c66-983b-908966cd1f94-var-run\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.274964 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/eab798dd-482a-4c66-983b-908966cd1f94-ovn-controller-tls-certs\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.274987 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlzh8\" (UniqueName: \"kubernetes.io/projected/eab798dd-482a-4c66-983b-908966cd1f94-kube-api-access-tlzh8\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.275013 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eab798dd-482a-4c66-983b-908966cd1f94-scripts\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.275034 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eab798dd-482a-4c66-983b-908966cd1f94-combined-ca-bundle\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.277182 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eab798dd-482a-4c66-983b-908966cd1f94-var-log-ovn\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.279054 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eab798dd-482a-4c66-983b-908966cd1f94-var-run-ovn\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.386391 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgmn6\" (UniqueName: \"kubernetes.io/projected/d2677b19-4860-497e-a473-6d52d4901d8c-kube-api-access-kgmn6\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.386440 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eab798dd-482a-4c66-983b-908966cd1f94-var-run\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.386468 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d2677b19-4860-497e-a473-6d52d4901d8c-var-run\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.386490 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/eab798dd-482a-4c66-983b-908966cd1f94-ovn-controller-tls-certs\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.386506 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlzh8\" (UniqueName: \"kubernetes.io/projected/eab798dd-482a-4c66-983b-908966cd1f94-kube-api-access-tlzh8\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.386527 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d2677b19-4860-497e-a473-6d52d4901d8c-var-lib\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.386544 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eab798dd-482a-4c66-983b-908966cd1f94-scripts\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.386567 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eab798dd-482a-4c66-983b-908966cd1f94-combined-ca-bundle\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.386581 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d2677b19-4860-497e-a473-6d52d4901d8c-etc-ovs\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.386602 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2677b19-4860-497e-a473-6d52d4901d8c-scripts\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.386627 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eab798dd-482a-4c66-983b-908966cd1f94-var-log-ovn\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.386673 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eab798dd-482a-4c66-983b-908966cd1f94-var-run-ovn\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.386696 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d2677b19-4860-497e-a473-6d52d4901d8c-var-log\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.387274 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eab798dd-482a-4c66-983b-908966cd1f94-var-run\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.389177 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eab798dd-482a-4c66-983b-908966cd1f94-var-log-ovn\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.389279 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eab798dd-482a-4c66-983b-908966cd1f94-var-run-ovn\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.394618 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eab798dd-482a-4c66-983b-908966cd1f94-scripts\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.403579 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/eab798dd-482a-4c66-983b-908966cd1f94-ovn-controller-tls-certs\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.404958 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eab798dd-482a-4c66-983b-908966cd1f94-combined-ca-bundle\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.417211 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlzh8\" (UniqueName: \"kubernetes.io/projected/eab798dd-482a-4c66-983b-908966cd1f94-kube-api-access-tlzh8\") pod \"ovn-controller-9kd7r\" (UID: \"eab798dd-482a-4c66-983b-908966cd1f94\") " pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.488660 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d2677b19-4860-497e-a473-6d52d4901d8c-var-log\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.489020 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/d2677b19-4860-497e-a473-6d52d4901d8c-var-log\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.489072 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgmn6\" (UniqueName: \"kubernetes.io/projected/d2677b19-4860-497e-a473-6d52d4901d8c-kube-api-access-kgmn6\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.489125 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d2677b19-4860-497e-a473-6d52d4901d8c-var-run\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.489166 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d2677b19-4860-497e-a473-6d52d4901d8c-var-lib\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.489319 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d2677b19-4860-497e-a473-6d52d4901d8c-var-run\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.489455 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/d2677b19-4860-497e-a473-6d52d4901d8c-var-lib\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.489199 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d2677b19-4860-497e-a473-6d52d4901d8c-etc-ovs\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.489529 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2677b19-4860-497e-a473-6d52d4901d8c-scripts\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.489623 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/d2677b19-4860-497e-a473-6d52d4901d8c-etc-ovs\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.491216 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.492911 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2677b19-4860-497e-a473-6d52d4901d8c-scripts\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.524838 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgmn6\" (UniqueName: \"kubernetes.io/projected/d2677b19-4860-497e-a473-6d52d4901d8c-kube-api-access-kgmn6\") pod \"ovn-controller-ovs-c5xnp\" (UID: \"d2677b19-4860-497e-a473-6d52d4901d8c\") " pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:19 crc kubenswrapper[4632]: I0313 10:23:19.583155 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.033461 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.034984 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.040759 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.042327 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-ws7xr" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.043402 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.043483 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.043995 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.067422 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.099307 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-config\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.099402 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.099496 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.099551 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.099581 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.099618 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.099673 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mqrx\" (UniqueName: \"kubernetes.io/projected/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-kube-api-access-9mqrx\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.099730 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.201957 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.202029 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-config\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.202061 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.202086 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.202130 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.202150 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.202178 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.202224 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mqrx\" (UniqueName: \"kubernetes.io/projected/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-kube-api-access-9mqrx\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.204695 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-config\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.206547 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.212453 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.212677 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.214086 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.214638 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.215025 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.234882 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mqrx\" (UniqueName: \"kubernetes.io/projected/4ee148f1-cc66-4aa0-b603-c8a70f3554f5-kube-api-access-9mqrx\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.238386 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"4ee148f1-cc66-4aa0-b603-c8a70f3554f5\") " pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:20 crc kubenswrapper[4632]: I0313 10:23:20.371859 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.673118 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.674922 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.677294 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.677687 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.678631 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.680118 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-k56t5" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.699461 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.776964 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5529a725-48d8-4a60-91cd-775a4b520c20-config\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.777043 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5529a725-48d8-4a60-91cd-775a4b520c20-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.777183 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5529a725-48d8-4a60-91cd-775a4b520c20-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.777258 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5529a725-48d8-4a60-91cd-775a4b520c20-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.777292 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5529a725-48d8-4a60-91cd-775a4b520c20-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.777321 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5529a725-48d8-4a60-91cd-775a4b520c20-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.777350 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.777384 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gpcj7\" (UniqueName: \"kubernetes.io/projected/5529a725-48d8-4a60-91cd-775a4b520c20-kube-api-access-gpcj7\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.879302 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.879373 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gpcj7\" (UniqueName: \"kubernetes.io/projected/5529a725-48d8-4a60-91cd-775a4b520c20-kube-api-access-gpcj7\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.879402 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5529a725-48d8-4a60-91cd-775a4b520c20-config\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.879449 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5529a725-48d8-4a60-91cd-775a4b520c20-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.879520 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5529a725-48d8-4a60-91cd-775a4b520c20-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.879569 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5529a725-48d8-4a60-91cd-775a4b520c20-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.879610 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5529a725-48d8-4a60-91cd-775a4b520c20-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.879631 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5529a725-48d8-4a60-91cd-775a4b520c20-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.881074 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5529a725-48d8-4a60-91cd-775a4b520c20-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.881356 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/5529a725-48d8-4a60-91cd-775a4b520c20-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.881475 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5529a725-48d8-4a60-91cd-775a4b520c20-config\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.881728 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.886618 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5529a725-48d8-4a60-91cd-775a4b520c20-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.887352 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5529a725-48d8-4a60-91cd-775a4b520c20-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.889960 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/5529a725-48d8-4a60-91cd-775a4b520c20-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.898524 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gpcj7\" (UniqueName: \"kubernetes.io/projected/5529a725-48d8-4a60-91cd-775a4b520c20-kube-api-access-gpcj7\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:22 crc kubenswrapper[4632]: I0313 10:23:22.908794 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"ovsdbserver-sb-0\" (UID: \"5529a725-48d8-4a60-91cd-775a4b520c20\") " pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:23 crc kubenswrapper[4632]: I0313 10:23:23.014082 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:24 crc kubenswrapper[4632]: W0313 10:23:24.789249 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21ce0311_ff05_4626_9663_a373ae31eb56.slice/crio-96915bb97645358a6555ca60c9308596dd68c9b71a65da098dd5679653d9f202 WatchSource:0}: Error finding container 96915bb97645358a6555ca60c9308596dd68c9b71a65da098dd5679653d9f202: Status 404 returned error can't find the container with id 96915bb97645358a6555ca60c9308596dd68c9b71a65da098dd5679653d9f202 Mar 13 10:23:25 crc kubenswrapper[4632]: I0313 10:23:25.299866 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"21ce0311-ff05-4626-9663-a373ae31eb56","Type":"ContainerStarted","Data":"96915bb97645358a6555ca60c9308596dd68c9b71a65da098dd5679653d9f202"} Mar 13 10:23:35 crc kubenswrapper[4632]: E0313 10:23:35.985716 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:23:35 crc kubenswrapper[4632]: E0313 10:23:35.987242 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:23:35 crc kubenswrapper[4632]: E0313 10:23:35.987374 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:e43235cb19da04699a53f42b6a75afe9,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4mmhg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-bd9cf7445-frlmw_openstack(da131e9a-8968-4569-a970-3aa95b2a830b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:23:35 crc kubenswrapper[4632]: E0313 10:23:35.988535 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" podUID="da131e9a-8968-4569-a970-3aa95b2a830b" Mar 13 10:23:37 crc kubenswrapper[4632]: E0313 10:23:37.172257 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:23:37 crc kubenswrapper[4632]: E0313 10:23:37.172529 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:23:37 crc kubenswrapper[4632]: E0313 10:23:37.172675 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:e43235cb19da04699a53f42b6a75afe9,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-k8hmx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(159c6cee-c82b-4725-82d6-dbd27216f53c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:23:37 crc kubenswrapper[4632]: E0313 10:23:37.174348 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="159c6cee-c82b-4725-82d6-dbd27216f53c" Mar 13 10:23:37 crc kubenswrapper[4632]: E0313 10:23:37.420424 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:e43235cb19da04699a53f42b6a75afe9\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="159c6cee-c82b-4725-82d6-dbd27216f53c" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.066018 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-mariadb:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.066359 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-mariadb:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.071151 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:mysql-bootstrap,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-mariadb:e43235cb19da04699a53f42b6a75afe9,Command:[bash /var/lib/operator-scripts/mysql_bootstrap.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:True,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:mysql-db,ReadOnly:false,MountPath:/var/lib/mysql,SubPath:mysql,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-default,ReadOnly:true,MountPath:/var/lib/config-data/default,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data-generated,ReadOnly:false,MountPath:/var/lib/config-data/generated,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:operator-scripts,ReadOnly:true,MountPath:/var/lib/operator-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qs454,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-galera-0_openstack(2cb2f546-c8c5-4ec9-aba8-d3782431de10): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.073696 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.099358 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.099415 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.099534 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:e43235cb19da04699a53f42b6a75afe9,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-868r4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-dcf85566c-59l8m_openstack(ea62b75b-fe31-433d-9ff1-a7333aacb383): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.100728 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-dcf85566c-59l8m" podUID="ea62b75b-fe31-433d-9ff1-a7333aacb383" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.437670 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:e43235cb19da04699a53f42b6a75afe9\\\"\"" pod="openstack/dnsmasq-dns-dcf85566c-59l8m" podUID="ea62b75b-fe31-433d-9ff1-a7333aacb383" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.438374 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"mysql-bootstrap\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-mariadb:e43235cb19da04699a53f42b6a75afe9\\\"\"" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.811727 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-memcached:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.811788 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-memcached:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.812296 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-memcached:e43235cb19da04699a53f42b6a75afe9,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n697h566h558h688hd9h59bh5f7h64bhc9h85h94hbdh557h9bh56ch5b4h5fdhd9h5cbh9ch56h64dh55fh694h5cbh54dh65h66ch65ch87h688h59cq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pvr2h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(d9100748-6b15-4ccf-b961-aab1135f08d1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.814200 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="d9100748-6b15-4ccf-b961-aab1135f08d1" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.857233 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.857307 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.857451 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:e43235cb19da04699a53f42b6a75afe9,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6kfgh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(211718f0-f29c-457b-bc2b-487bb76d4801): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.858754 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="211718f0-f29c-457b-bc2b-487bb76d4801" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.911304 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.911564 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.911669 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-neutron-server:e43235cb19da04699a53f42b6a75afe9,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-c54n6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-7b95c5c449-th4fn_openstack(dcb4500d-7a53-4091-b3af-394eb0f49130): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:23:39 crc kubenswrapper[4632]: E0313 10:23:39.913023 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-7b95c5c449-th4fn" podUID="dcb4500d-7a53-4091-b3af-394eb0f49130" Mar 13 10:23:39 crc kubenswrapper[4632]: I0313 10:23:39.980398 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.101365 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4mmhg\" (UniqueName: \"kubernetes.io/projected/da131e9a-8968-4569-a970-3aa95b2a830b-kube-api-access-4mmhg\") pod \"da131e9a-8968-4569-a970-3aa95b2a830b\" (UID: \"da131e9a-8968-4569-a970-3aa95b2a830b\") " Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.101466 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da131e9a-8968-4569-a970-3aa95b2a830b-dns-svc\") pod \"da131e9a-8968-4569-a970-3aa95b2a830b\" (UID: \"da131e9a-8968-4569-a970-3aa95b2a830b\") " Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.101559 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da131e9a-8968-4569-a970-3aa95b2a830b-config\") pod \"da131e9a-8968-4569-a970-3aa95b2a830b\" (UID: \"da131e9a-8968-4569-a970-3aa95b2a830b\") " Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.103057 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da131e9a-8968-4569-a970-3aa95b2a830b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "da131e9a-8968-4569-a970-3aa95b2a830b" (UID: "da131e9a-8968-4569-a970-3aa95b2a830b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.103257 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/da131e9a-8968-4569-a970-3aa95b2a830b-config" (OuterVolumeSpecName: "config") pod "da131e9a-8968-4569-a970-3aa95b2a830b" (UID: "da131e9a-8968-4569-a970-3aa95b2a830b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.112187 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da131e9a-8968-4569-a970-3aa95b2a830b-kube-api-access-4mmhg" (OuterVolumeSpecName: "kube-api-access-4mmhg") pod "da131e9a-8968-4569-a970-3aa95b2a830b" (UID: "da131e9a-8968-4569-a970-3aa95b2a830b"). InnerVolumeSpecName "kube-api-access-4mmhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.206469 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4mmhg\" (UniqueName: \"kubernetes.io/projected/da131e9a-8968-4569-a970-3aa95b2a830b-kube-api-access-4mmhg\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.206498 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/da131e9a-8968-4569-a970-3aa95b2a830b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.206509 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/da131e9a-8968-4569-a970-3aa95b2a830b-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.450123 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.451226 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bd9cf7445-frlmw" event={"ID":"da131e9a-8968-4569-a970-3aa95b2a830b","Type":"ContainerDied","Data":"55fcbfaa372a43ed729e32593f5c37c0b24f03c99a93bb7fe4dbc16e4c07e1b0"} Mar 13 10:23:40 crc kubenswrapper[4632]: E0313 10:23:40.457033 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-rabbitmq:e43235cb19da04699a53f42b6a75afe9\\\"\"" pod="openstack/rabbitmq-server-0" podUID="211718f0-f29c-457b-bc2b-487bb76d4801" Mar 13 10:23:40 crc kubenswrapper[4632]: E0313 10:23:40.457042 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-memcached:e43235cb19da04699a53f42b6a75afe9\\\"\"" pod="openstack/memcached-0" podUID="d9100748-6b15-4ccf-b961-aab1135f08d1" Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.461206 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.461269 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.489173 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9kd7r"] Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.632157 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bd9cf7445-frlmw"] Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.646779 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bd9cf7445-frlmw"] Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.724665 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 10:23:40 crc kubenswrapper[4632]: W0313 10:23:40.886779 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ee148f1_cc66_4aa0_b603_c8a70f3554f5.slice/crio-2fdfb53a72f64e053dd71b35245437359614c655b2ac8c95f14508dcc5053577 WatchSource:0}: Error finding container 2fdfb53a72f64e053dd71b35245437359614c655b2ac8c95f14508dcc5053577: Status 404 returned error can't find the container with id 2fdfb53a72f64e053dd71b35245437359614c655b2ac8c95f14508dcc5053577 Mar 13 10:23:40 crc kubenswrapper[4632]: W0313 10:23:40.891812 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeab798dd_482a_4c66_983b_908966cd1f94.slice/crio-83331c2a9a3d5614a4cf49eb307ac4e28725cc613d758576db2b95cc3de7bb84 WatchSource:0}: Error finding container 83331c2a9a3d5614a4cf49eb307ac4e28725cc613d758576db2b95cc3de7bb84: Status 404 returned error can't find the container with id 83331c2a9a3d5614a4cf49eb307ac4e28725cc613d758576db2b95cc3de7bb84 Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.904761 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b95c5c449-th4fn" Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.944420 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcb4500d-7a53-4091-b3af-394eb0f49130-config\") pod \"dcb4500d-7a53-4091-b3af-394eb0f49130\" (UID: \"dcb4500d-7a53-4091-b3af-394eb0f49130\") " Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.944667 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c54n6\" (UniqueName: \"kubernetes.io/projected/dcb4500d-7a53-4091-b3af-394eb0f49130-kube-api-access-c54n6\") pod \"dcb4500d-7a53-4091-b3af-394eb0f49130\" (UID: \"dcb4500d-7a53-4091-b3af-394eb0f49130\") " Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.946314 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dcb4500d-7a53-4091-b3af-394eb0f49130-config" (OuterVolumeSpecName: "config") pod "dcb4500d-7a53-4091-b3af-394eb0f49130" (UID: "dcb4500d-7a53-4091-b3af-394eb0f49130"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:23:40 crc kubenswrapper[4632]: I0313 10:23:40.949976 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcb4500d-7a53-4091-b3af-394eb0f49130-kube-api-access-c54n6" (OuterVolumeSpecName: "kube-api-access-c54n6") pod "dcb4500d-7a53-4091-b3af-394eb0f49130" (UID: "dcb4500d-7a53-4091-b3af-394eb0f49130"). InnerVolumeSpecName "kube-api-access-c54n6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:23:41 crc kubenswrapper[4632]: I0313 10:23:41.046050 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c54n6\" (UniqueName: \"kubernetes.io/projected/dcb4500d-7a53-4091-b3af-394eb0f49130-kube-api-access-c54n6\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:41 crc kubenswrapper[4632]: I0313 10:23:41.046084 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dcb4500d-7a53-4091-b3af-394eb0f49130-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:41 crc kubenswrapper[4632]: I0313 10:23:41.474489 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1761ca69-46fd-4375-af60-22b3e77c19a2","Type":"ContainerStarted","Data":"38fabecc1af392ae2500911da4ea37a128aafd15dfb148ef01201cadf3cfb5e8"} Mar 13 10:23:41 crc kubenswrapper[4632]: I0313 10:23:41.477618 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4ee148f1-cc66-4aa0-b603-c8a70f3554f5","Type":"ContainerStarted","Data":"2fdfb53a72f64e053dd71b35245437359614c655b2ac8c95f14508dcc5053577"} Mar 13 10:23:41 crc kubenswrapper[4632]: I0313 10:23:41.480006 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b95c5c449-th4fn" event={"ID":"dcb4500d-7a53-4091-b3af-394eb0f49130","Type":"ContainerDied","Data":"795a9fce6e217788c208128d366941b235a94e772dfaab8892b09ba717741562"} Mar 13 10:23:41 crc kubenswrapper[4632]: I0313 10:23:41.480046 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b95c5c449-th4fn" Mar 13 10:23:41 crc kubenswrapper[4632]: I0313 10:23:41.482721 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9kd7r" event={"ID":"eab798dd-482a-4c66-983b-908966cd1f94","Type":"ContainerStarted","Data":"83331c2a9a3d5614a4cf49eb307ac4e28725cc613d758576db2b95cc3de7bb84"} Mar 13 10:23:41 crc kubenswrapper[4632]: I0313 10:23:41.572768 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-c5xnp"] Mar 13 10:23:41 crc kubenswrapper[4632]: I0313 10:23:41.590106 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7b95c5c449-th4fn"] Mar 13 10:23:41 crc kubenswrapper[4632]: I0313 10:23:41.607582 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7b95c5c449-th4fn"] Mar 13 10:23:41 crc kubenswrapper[4632]: I0313 10:23:41.722174 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 10:23:41 crc kubenswrapper[4632]: W0313 10:23:41.765094 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5529a725_48d8_4a60_91cd_775a4b520c20.slice/crio-b6992dcf3b7705827a5dd88fc12a83e39cc06f6db59b355289a33a8b11826133 WatchSource:0}: Error finding container b6992dcf3b7705827a5dd88fc12a83e39cc06f6db59b355289a33a8b11826133: Status 404 returned error can't find the container with id b6992dcf3b7705827a5dd88fc12a83e39cc06f6db59b355289a33a8b11826133 Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.056786 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da131e9a-8968-4569-a970-3aa95b2a830b" path="/var/lib/kubelet/pods/da131e9a-8968-4569-a970-3aa95b2a830b/volumes" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.057402 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcb4500d-7a53-4091-b3af-394eb0f49130" path="/var/lib/kubelet/pods/dcb4500d-7a53-4091-b3af-394eb0f49130/volumes" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.371999 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-798sf"] Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.373288 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.380590 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.387108 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-798sf"] Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.509525 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9246fc4f-3716-4a8b-9854-52137cf04e9a-ovs-rundir\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.509826 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9246fc4f-3716-4a8b-9854-52137cf04e9a-ovn-rundir\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.509983 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9246fc4f-3716-4a8b-9854-52137cf04e9a-config\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.510050 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jstns\" (UniqueName: \"kubernetes.io/projected/9246fc4f-3716-4a8b-9854-52137cf04e9a-kube-api-access-jstns\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.510075 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9246fc4f-3716-4a8b-9854-52137cf04e9a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.510160 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9246fc4f-3716-4a8b-9854-52137cf04e9a-combined-ca-bundle\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.559719 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5529a725-48d8-4a60-91cd-775a4b520c20","Type":"ContainerStarted","Data":"b6992dcf3b7705827a5dd88fc12a83e39cc06f6db59b355289a33a8b11826133"} Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.562227 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-c5xnp" event={"ID":"d2677b19-4860-497e-a473-6d52d4901d8c","Type":"ContainerStarted","Data":"0a2653ff7d965ff6780efa77e0c16b65c2fa0483036255f3de2799ed24f8a7a4"} Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.612062 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9246fc4f-3716-4a8b-9854-52137cf04e9a-ovn-rundir\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.612163 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9246fc4f-3716-4a8b-9854-52137cf04e9a-config\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.612208 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jstns\" (UniqueName: \"kubernetes.io/projected/9246fc4f-3716-4a8b-9854-52137cf04e9a-kube-api-access-jstns\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.612237 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9246fc4f-3716-4a8b-9854-52137cf04e9a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.612298 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9246fc4f-3716-4a8b-9854-52137cf04e9a-combined-ca-bundle\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.612325 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9246fc4f-3716-4a8b-9854-52137cf04e9a-ovs-rundir\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.612714 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/9246fc4f-3716-4a8b-9854-52137cf04e9a-ovs-rundir\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.612796 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/9246fc4f-3716-4a8b-9854-52137cf04e9a-ovn-rundir\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.613817 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9246fc4f-3716-4a8b-9854-52137cf04e9a-config\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.630280 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9246fc4f-3716-4a8b-9854-52137cf04e9a-combined-ca-bundle\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.631592 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9246fc4f-3716-4a8b-9854-52137cf04e9a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.654470 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jstns\" (UniqueName: \"kubernetes.io/projected/9246fc4f-3716-4a8b-9854-52137cf04e9a-kube-api-access-jstns\") pod \"ovn-controller-metrics-798sf\" (UID: \"9246fc4f-3716-4a8b-9854-52137cf04e9a\") " pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.693265 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86545856d7-fkxhx"] Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.719925 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7df696cbbf-4tc7r"] Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.721535 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.737798 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.739410 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-config\") pod \"dnsmasq-dns-7df696cbbf-4tc7r\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.739473 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7phzq\" (UniqueName: \"kubernetes.io/projected/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-kube-api-access-7phzq\") pod \"dnsmasq-dns-7df696cbbf-4tc7r\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.739693 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-dns-svc\") pod \"dnsmasq-dns-7df696cbbf-4tc7r\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.739720 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-ovsdbserver-nb\") pod \"dnsmasq-dns-7df696cbbf-4tc7r\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.754866 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7df696cbbf-4tc7r"] Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.841916 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-dns-svc\") pod \"dnsmasq-dns-7df696cbbf-4tc7r\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.841990 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-ovsdbserver-nb\") pod \"dnsmasq-dns-7df696cbbf-4tc7r\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.842070 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-config\") pod \"dnsmasq-dns-7df696cbbf-4tc7r\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.842102 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7phzq\" (UniqueName: \"kubernetes.io/projected/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-kube-api-access-7phzq\") pod \"dnsmasq-dns-7df696cbbf-4tc7r\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.843721 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-ovsdbserver-nb\") pod \"dnsmasq-dns-7df696cbbf-4tc7r\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.844443 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-config\") pod \"dnsmasq-dns-7df696cbbf-4tc7r\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.844512 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-798sf" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.848215 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-dns-svc\") pod \"dnsmasq-dns-7df696cbbf-4tc7r\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:42 crc kubenswrapper[4632]: I0313 10:23:42.874015 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7phzq\" (UniqueName: \"kubernetes.io/projected/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-kube-api-access-7phzq\") pod \"dnsmasq-dns-7df696cbbf-4tc7r\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.086826 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dcf85566c-59l8m"] Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.133699 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.229487 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-79dfb79747-jv5m6"] Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.231319 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.240065 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.254222 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79dfb79747-jv5m6"] Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.357175 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-config\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.357849 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-ovsdbserver-sb\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.358055 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-dns-svc\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.358212 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-ovsdbserver-nb\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.358299 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pv67\" (UniqueName: \"kubernetes.io/projected/a362448e-8daa-4bf4-958f-f3ca135be228-kube-api-access-5pv67\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.461648 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-config\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.461703 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-ovsdbserver-sb\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.461776 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-dns-svc\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.461808 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-ovsdbserver-nb\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.461835 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pv67\" (UniqueName: \"kubernetes.io/projected/a362448e-8daa-4bf4-958f-f3ca135be228-kube-api-access-5pv67\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.463177 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-dns-svc\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.463275 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-ovsdbserver-sb\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.463753 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-ovsdbserver-nb\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.464013 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-config\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.511607 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pv67\" (UniqueName: \"kubernetes.io/projected/a362448e-8daa-4bf4-958f-f3ca135be228-kube-api-access-5pv67\") pod \"dnsmasq-dns-79dfb79747-jv5m6\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.610191 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"21ce0311-ff05-4626-9663-a373ae31eb56","Type":"ContainerStarted","Data":"1a16adc836bd27406a32b6f9e9672d40ce7e70e9caf414e0a9334fb34a8ec7ab"} Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.610499 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.617294 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.621466 4632 generic.go:334] "Generic (PLEG): container finished" podID="7afad0f9-c29c-40e6-8605-1df67a505a82" containerID="3265e919f0b844473557cc692c4854b5d839d0a7869d7f08dc68a3ae52955bc2" exitCode=0 Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.621495 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86545856d7-fkxhx" event={"ID":"7afad0f9-c29c-40e6-8605-1df67a505a82","Type":"ContainerDied","Data":"3265e919f0b844473557cc692c4854b5d839d0a7869d7f08dc68a3ae52955bc2"} Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.654633 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=9.923342576 podStartE2EDuration="27.654611941s" podCreationTimestamp="2026-03-13 10:23:16 +0000 UTC" firstStartedPulling="2026-03-13 10:23:24.795462405 +0000 UTC m=+1178.817992538" lastFinishedPulling="2026-03-13 10:23:42.52673177 +0000 UTC m=+1196.549261903" observedRunningTime="2026-03-13 10:23:43.634721873 +0000 UTC m=+1197.657252016" watchObservedRunningTime="2026-03-13 10:23:43.654611941 +0000 UTC m=+1197.677142074" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.794193 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcf85566c-59l8m" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.878844 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea62b75b-fe31-433d-9ff1-a7333aacb383-dns-svc\") pod \"ea62b75b-fe31-433d-9ff1-a7333aacb383\" (UID: \"ea62b75b-fe31-433d-9ff1-a7333aacb383\") " Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.884430 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea62b75b-fe31-433d-9ff1-a7333aacb383-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea62b75b-fe31-433d-9ff1-a7333aacb383" (UID: "ea62b75b-fe31-433d-9ff1-a7333aacb383"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.886521 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-798sf"] Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.892716 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-868r4\" (UniqueName: \"kubernetes.io/projected/ea62b75b-fe31-433d-9ff1-a7333aacb383-kube-api-access-868r4\") pod \"ea62b75b-fe31-433d-9ff1-a7333aacb383\" (UID: \"ea62b75b-fe31-433d-9ff1-a7333aacb383\") " Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.892772 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea62b75b-fe31-433d-9ff1-a7333aacb383-config\") pod \"ea62b75b-fe31-433d-9ff1-a7333aacb383\" (UID: \"ea62b75b-fe31-433d-9ff1-a7333aacb383\") " Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.893354 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea62b75b-fe31-433d-9ff1-a7333aacb383-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.893783 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea62b75b-fe31-433d-9ff1-a7333aacb383-config" (OuterVolumeSpecName: "config") pod "ea62b75b-fe31-433d-9ff1-a7333aacb383" (UID: "ea62b75b-fe31-433d-9ff1-a7333aacb383"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.895007 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7df696cbbf-4tc7r"] Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.911899 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea62b75b-fe31-433d-9ff1-a7333aacb383-kube-api-access-868r4" (OuterVolumeSpecName: "kube-api-access-868r4") pod "ea62b75b-fe31-433d-9ff1-a7333aacb383" (UID: "ea62b75b-fe31-433d-9ff1-a7333aacb383"). InnerVolumeSpecName "kube-api-access-868r4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.995483 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-868r4\" (UniqueName: \"kubernetes.io/projected/ea62b75b-fe31-433d-9ff1-a7333aacb383-kube-api-access-868r4\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:43 crc kubenswrapper[4632]: I0313 10:23:43.995521 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea62b75b-fe31-433d-9ff1-a7333aacb383-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.061782 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86545856d7-fkxhx" Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.096110 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7afad0f9-c29c-40e6-8605-1df67a505a82-config\") pod \"7afad0f9-c29c-40e6-8605-1df67a505a82\" (UID: \"7afad0f9-c29c-40e6-8605-1df67a505a82\") " Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.096314 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7afad0f9-c29c-40e6-8605-1df67a505a82-dns-svc\") pod \"7afad0f9-c29c-40e6-8605-1df67a505a82\" (UID: \"7afad0f9-c29c-40e6-8605-1df67a505a82\") " Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.096404 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qdf8h\" (UniqueName: \"kubernetes.io/projected/7afad0f9-c29c-40e6-8605-1df67a505a82-kube-api-access-qdf8h\") pod \"7afad0f9-c29c-40e6-8605-1df67a505a82\" (UID: \"7afad0f9-c29c-40e6-8605-1df67a505a82\") " Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.101694 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7afad0f9-c29c-40e6-8605-1df67a505a82-kube-api-access-qdf8h" (OuterVolumeSpecName: "kube-api-access-qdf8h") pod "7afad0f9-c29c-40e6-8605-1df67a505a82" (UID: "7afad0f9-c29c-40e6-8605-1df67a505a82"). InnerVolumeSpecName "kube-api-access-qdf8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.119299 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7afad0f9-c29c-40e6-8605-1df67a505a82-config" (OuterVolumeSpecName: "config") pod "7afad0f9-c29c-40e6-8605-1df67a505a82" (UID: "7afad0f9-c29c-40e6-8605-1df67a505a82"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.120743 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7afad0f9-c29c-40e6-8605-1df67a505a82-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7afad0f9-c29c-40e6-8605-1df67a505a82" (UID: "7afad0f9-c29c-40e6-8605-1df67a505a82"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.198875 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7afad0f9-c29c-40e6-8605-1df67a505a82-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.198910 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qdf8h\" (UniqueName: \"kubernetes.io/projected/7afad0f9-c29c-40e6-8605-1df67a505a82-kube-api-access-qdf8h\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.198921 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7afad0f9-c29c-40e6-8605-1df67a505a82-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.300548 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-79dfb79747-jv5m6"] Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.648291 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86545856d7-fkxhx" Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.648293 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86545856d7-fkxhx" event={"ID":"7afad0f9-c29c-40e6-8605-1df67a505a82","Type":"ContainerDied","Data":"99dcdc47d7e36bada4f3fc23414bc9bbd494a02612477871943ec84558819c97"} Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.648822 4632 scope.go:117] "RemoveContainer" containerID="3265e919f0b844473557cc692c4854b5d839d0a7869d7f08dc68a3ae52955bc2" Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.651797 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-798sf" event={"ID":"9246fc4f-3716-4a8b-9854-52137cf04e9a","Type":"ContainerStarted","Data":"baeb1292c403de7ee01a776c7dd912c73407e5f4c64078e55174c0cb15e01bae"} Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.654311 4632 generic.go:334] "Generic (PLEG): container finished" podID="df11136a-b7d7-4a5a-a0cc-d0ebbc069b34" containerID="2915f2c9d1176c44121e26cd23d0bb33a0c1b7ffaf310d2e0081ce7eb76b0909" exitCode=0 Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.654464 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" event={"ID":"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34","Type":"ContainerDied","Data":"2915f2c9d1176c44121e26cd23d0bb33a0c1b7ffaf310d2e0081ce7eb76b0909"} Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.654542 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" event={"ID":"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34","Type":"ContainerStarted","Data":"c5cd8fa4d9fefd8144c63785679d7e91126ba9c71436b6d101b2ee0cc8ea3019"} Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.656038 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-dcf85566c-59l8m" event={"ID":"ea62b75b-fe31-433d-9ff1-a7333aacb383","Type":"ContainerDied","Data":"8c2e1684ce51b6e904615df8d377b14786ec45f198df25a296fb0708834a1826"} Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.656139 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-dcf85566c-59l8m" Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.661918 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" event={"ID":"a362448e-8daa-4bf4-958f-f3ca135be228","Type":"ContainerStarted","Data":"353c97e7f46060d145aa3be9824f787e63d6eed5891607427c9311023caa0833"} Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.762873 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-dcf85566c-59l8m"] Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.766661 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-dcf85566c-59l8m"] Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.780752 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86545856d7-fkxhx"] Mar 13 10:23:44 crc kubenswrapper[4632]: I0313 10:23:44.786174 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86545856d7-fkxhx"] Mar 13 10:23:45 crc kubenswrapper[4632]: I0313 10:23:45.682599 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" event={"ID":"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34","Type":"ContainerStarted","Data":"e8191df8902be4a4da6f0e247d64bea9567a78aed70ff1b1918dad0f09a75382"} Mar 13 10:23:45 crc kubenswrapper[4632]: I0313 10:23:45.682917 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:45 crc kubenswrapper[4632]: I0313 10:23:45.687008 4632 generic.go:334] "Generic (PLEG): container finished" podID="a362448e-8daa-4bf4-958f-f3ca135be228" containerID="c9a88952f81b62d419132fe9a18256ffde7daf30602bf205173e43b0963b20c3" exitCode=0 Mar 13 10:23:45 crc kubenswrapper[4632]: I0313 10:23:45.687142 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" event={"ID":"a362448e-8daa-4bf4-958f-f3ca135be228","Type":"ContainerDied","Data":"c9a88952f81b62d419132fe9a18256ffde7daf30602bf205173e43b0963b20c3"} Mar 13 10:23:45 crc kubenswrapper[4632]: I0313 10:23:45.718346 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" podStartSLOduration=3.718328062 podStartE2EDuration="3.718328062s" podCreationTimestamp="2026-03-13 10:23:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:23:45.711590857 +0000 UTC m=+1199.734121010" watchObservedRunningTime="2026-03-13 10:23:45.718328062 +0000 UTC m=+1199.740858195" Mar 13 10:23:46 crc kubenswrapper[4632]: I0313 10:23:46.056630 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7afad0f9-c29c-40e6-8605-1df67a505a82" path="/var/lib/kubelet/pods/7afad0f9-c29c-40e6-8605-1df67a505a82/volumes" Mar 13 10:23:46 crc kubenswrapper[4632]: I0313 10:23:46.057401 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea62b75b-fe31-433d-9ff1-a7333aacb383" path="/var/lib/kubelet/pods/ea62b75b-fe31-433d-9ff1-a7333aacb383/volumes" Mar 13 10:23:46 crc kubenswrapper[4632]: I0313 10:23:46.698278 4632 generic.go:334] "Generic (PLEG): container finished" podID="1761ca69-46fd-4375-af60-22b3e77c19a2" containerID="38fabecc1af392ae2500911da4ea37a128aafd15dfb148ef01201cadf3cfb5e8" exitCode=0 Mar 13 10:23:46 crc kubenswrapper[4632]: I0313 10:23:46.699366 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1761ca69-46fd-4375-af60-22b3e77c19a2","Type":"ContainerDied","Data":"38fabecc1af392ae2500911da4ea37a128aafd15dfb148ef01201cadf3cfb5e8"} Mar 13 10:23:49 crc kubenswrapper[4632]: I0313 10:23:49.720794 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" event={"ID":"a362448e-8daa-4bf4-958f-f3ca135be228","Type":"ContainerStarted","Data":"6e1c032b958be8592683422ee06f119be07d42e5fc24c06ebfd10193412b1ccc"} Mar 13 10:23:49 crc kubenswrapper[4632]: I0313 10:23:49.721399 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:49 crc kubenswrapper[4632]: I0313 10:23:49.742486 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" podStartSLOduration=6.742470529 podStartE2EDuration="6.742470529s" podCreationTimestamp="2026-03-13 10:23:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:23:49.739329452 +0000 UTC m=+1203.761859615" watchObservedRunningTime="2026-03-13 10:23:49.742470529 +0000 UTC m=+1203.765000662" Mar 13 10:23:53 crc kubenswrapper[4632]: I0313 10:23:53.136012 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:55 crc kubenswrapper[4632]: I0313 10:23:55.772627 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"d9100748-6b15-4ccf-b961-aab1135f08d1","Type":"ContainerStarted","Data":"1b263ebf136efa3203186a62eec927f384736d1b2e1990d4a68dbaa9f6638698"} Mar 13 10:23:55 crc kubenswrapper[4632]: I0313 10:23:55.774039 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 13 10:23:55 crc kubenswrapper[4632]: I0313 10:23:55.787171 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2cb2f546-c8c5-4ec9-aba8-d3782431de10","Type":"ContainerStarted","Data":"aa55e67bc4a499dbbd5e317a3b00bbfa4846877aed563132955d84afe5164371"} Mar 13 10:23:55 crc kubenswrapper[4632]: I0313 10:23:55.797339 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-798sf" event={"ID":"9246fc4f-3716-4a8b-9854-52137cf04e9a","Type":"ContainerStarted","Data":"c9098f2c3a39988551490caac37c53effbd8dc1e78c733223e6d49d975e2886d"} Mar 13 10:23:55 crc kubenswrapper[4632]: I0313 10:23:55.810299 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.605021047 podStartE2EDuration="42.810277034s" podCreationTimestamp="2026-03-13 10:23:13 +0000 UTC" firstStartedPulling="2026-03-13 10:23:14.649528326 +0000 UTC m=+1168.672058459" lastFinishedPulling="2026-03-13 10:23:54.854784313 +0000 UTC m=+1208.877314446" observedRunningTime="2026-03-13 10:23:55.792818577 +0000 UTC m=+1209.815348720" watchObservedRunningTime="2026-03-13 10:23:55.810277034 +0000 UTC m=+1209.832807167" Mar 13 10:23:55 crc kubenswrapper[4632]: I0313 10:23:55.820664 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1761ca69-46fd-4375-af60-22b3e77c19a2","Type":"ContainerStarted","Data":"8d6e3c3bf2cc94b4f346233606a1c5c55a2993e7644d8a78c77dc12972c98a9e"} Mar 13 10:23:55 crc kubenswrapper[4632]: I0313 10:23:55.837659 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4ee148f1-cc66-4aa0-b603-c8a70f3554f5","Type":"ContainerStarted","Data":"7846dbbc8d612576b942082c32786f4118590cc8dbfca6330420b7f7e10d85e9"} Mar 13 10:23:55 crc kubenswrapper[4632]: I0313 10:23:55.850560 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5529a725-48d8-4a60-91cd-775a4b520c20","Type":"ContainerStarted","Data":"484b3992ae7deb9ebbf37851b58541e4ffa16a07a4275161ef47376ec6250e84"} Mar 13 10:23:55 crc kubenswrapper[4632]: I0313 10:23:55.853575 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=18.275293193 podStartE2EDuration="43.853558736s" podCreationTimestamp="2026-03-13 10:23:12 +0000 UTC" firstStartedPulling="2026-03-13 10:23:14.600191815 +0000 UTC m=+1168.622721958" lastFinishedPulling="2026-03-13 10:23:40.178457368 +0000 UTC m=+1194.200987501" observedRunningTime="2026-03-13 10:23:55.849962878 +0000 UTC m=+1209.872493021" watchObservedRunningTime="2026-03-13 10:23:55.853558736 +0000 UTC m=+1209.876088869" Mar 13 10:23:55 crc kubenswrapper[4632]: I0313 10:23:55.863077 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9kd7r" event={"ID":"eab798dd-482a-4c66-983b-908966cd1f94","Type":"ContainerStarted","Data":"80f6e66214e6fd7a5ba2fbf97142f0099cd2a2a9c54e115d0323772b3f5f702b"} Mar 13 10:23:55 crc kubenswrapper[4632]: I0313 10:23:55.863178 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-9kd7r" Mar 13 10:23:55 crc kubenswrapper[4632]: I0313 10:23:55.869408 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-c5xnp" event={"ID":"d2677b19-4860-497e-a473-6d52d4901d8c","Type":"ContainerStarted","Data":"11b6da1f5ac161c5fa2ede8304d7459c7fa06380905e5f947dc2d54c0f26b4cf"} Mar 13 10:23:55 crc kubenswrapper[4632]: I0313 10:23:55.923024 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-798sf" podStartSLOduration=2.98828617 podStartE2EDuration="13.92300546s" podCreationTimestamp="2026-03-13 10:23:42 +0000 UTC" firstStartedPulling="2026-03-13 10:23:43.920092574 +0000 UTC m=+1197.942622707" lastFinishedPulling="2026-03-13 10:23:54.854811864 +0000 UTC m=+1208.877341997" observedRunningTime="2026-03-13 10:23:55.883403148 +0000 UTC m=+1209.905933281" watchObservedRunningTime="2026-03-13 10:23:55.92300546 +0000 UTC m=+1209.945535593" Mar 13 10:23:55 crc kubenswrapper[4632]: I0313 10:23:55.923283 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-9kd7r" podStartSLOduration=22.989840578 podStartE2EDuration="36.923277587s" podCreationTimestamp="2026-03-13 10:23:19 +0000 UTC" firstStartedPulling="2026-03-13 10:23:40.89673919 +0000 UTC m=+1194.919269323" lastFinishedPulling="2026-03-13 10:23:54.830176199 +0000 UTC m=+1208.852706332" observedRunningTime="2026-03-13 10:23:55.919170566 +0000 UTC m=+1209.941700709" watchObservedRunningTime="2026-03-13 10:23:55.923277587 +0000 UTC m=+1209.945807730" Mar 13 10:23:56 crc kubenswrapper[4632]: I0313 10:23:56.465053 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 13 10:23:56 crc kubenswrapper[4632]: I0313 10:23:56.877546 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"5529a725-48d8-4a60-91cd-775a4b520c20","Type":"ContainerStarted","Data":"72cf388b074df1070acb7f0728aa38efe39290a683180db18cbd4deeba546b7f"} Mar 13 10:23:56 crc kubenswrapper[4632]: I0313 10:23:56.887593 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"159c6cee-c82b-4725-82d6-dbd27216f53c","Type":"ContainerStarted","Data":"d5bd67d741203861cfd1afa23ec3f20fd6236a99625563ac3c10816dbb2a6677"} Mar 13 10:23:56 crc kubenswrapper[4632]: I0313 10:23:56.900908 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=22.83559854 podStartE2EDuration="35.900892461s" podCreationTimestamp="2026-03-13 10:23:21 +0000 UTC" firstStartedPulling="2026-03-13 10:23:41.778961734 +0000 UTC m=+1195.801491867" lastFinishedPulling="2026-03-13 10:23:54.844255655 +0000 UTC m=+1208.866785788" observedRunningTime="2026-03-13 10:23:56.899235691 +0000 UTC m=+1210.921765834" watchObservedRunningTime="2026-03-13 10:23:56.900892461 +0000 UTC m=+1210.923422594" Mar 13 10:23:56 crc kubenswrapper[4632]: I0313 10:23:56.907245 4632 generic.go:334] "Generic (PLEG): container finished" podID="d2677b19-4860-497e-a473-6d52d4901d8c" containerID="11b6da1f5ac161c5fa2ede8304d7459c7fa06380905e5f947dc2d54c0f26b4cf" exitCode=0 Mar 13 10:23:56 crc kubenswrapper[4632]: I0313 10:23:56.907338 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-c5xnp" event={"ID":"d2677b19-4860-497e-a473-6d52d4901d8c","Type":"ContainerDied","Data":"11b6da1f5ac161c5fa2ede8304d7459c7fa06380905e5f947dc2d54c0f26b4cf"} Mar 13 10:23:56 crc kubenswrapper[4632]: I0313 10:23:56.913556 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"211718f0-f29c-457b-bc2b-487bb76d4801","Type":"ContainerStarted","Data":"92d546a480b1e583e7b11dc48ab2d570a4a8d7af0616de2352d72ca175520f17"} Mar 13 10:23:56 crc kubenswrapper[4632]: I0313 10:23:56.918496 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"4ee148f1-cc66-4aa0-b603-c8a70f3554f5","Type":"ContainerStarted","Data":"a71593ee348ef6c77356c7339c1905f6ce2c2c09c03e7e4db8c667a05c46e720"} Mar 13 10:23:57 crc kubenswrapper[4632]: I0313 10:23:57.091736 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=25.153601378 podStartE2EDuration="39.091702342s" podCreationTimestamp="2026-03-13 10:23:18 +0000 UTC" firstStartedPulling="2026-03-13 10:23:40.892078515 +0000 UTC m=+1194.914608648" lastFinishedPulling="2026-03-13 10:23:54.830179479 +0000 UTC m=+1208.852709612" observedRunningTime="2026-03-13 10:23:57.085690275 +0000 UTC m=+1211.108220418" watchObservedRunningTime="2026-03-13 10:23:57.091702342 +0000 UTC m=+1211.114232485" Mar 13 10:23:57 crc kubenswrapper[4632]: I0313 10:23:57.929965 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-c5xnp" event={"ID":"d2677b19-4860-497e-a473-6d52d4901d8c","Type":"ContainerStarted","Data":"ae34fea57a4127d6113c858d3b0966d7368d212d3cd1d556e2036d7d77bdfa0f"} Mar 13 10:23:57 crc kubenswrapper[4632]: I0313 10:23:57.930324 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-c5xnp" event={"ID":"d2677b19-4860-497e-a473-6d52d4901d8c","Type":"ContainerStarted","Data":"808a50dadd2432be26181ff455295e45b780e27c923ac2d92d7852cd6644117a"} Mar 13 10:23:57 crc kubenswrapper[4632]: I0313 10:23:57.952154 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-c5xnp" podStartSLOduration=26.003174455 podStartE2EDuration="38.952098261s" podCreationTimestamp="2026-03-13 10:23:19 +0000 UTC" firstStartedPulling="2026-03-13 10:23:41.757455526 +0000 UTC m=+1195.779985659" lastFinishedPulling="2026-03-13 10:23:54.706379332 +0000 UTC m=+1208.728909465" observedRunningTime="2026-03-13 10:23:57.950933782 +0000 UTC m=+1211.973463915" watchObservedRunningTime="2026-03-13 10:23:57.952098261 +0000 UTC m=+1211.974628394" Mar 13 10:23:58 crc kubenswrapper[4632]: I0313 10:23:58.014966 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:58 crc kubenswrapper[4632]: I0313 10:23:58.623154 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:23:58 crc kubenswrapper[4632]: I0313 10:23:58.698598 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7df696cbbf-4tc7r"] Mar 13 10:23:58 crc kubenswrapper[4632]: I0313 10:23:58.698917 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" podUID="df11136a-b7d7-4a5a-a0cc-d0ebbc069b34" containerName="dnsmasq-dns" containerID="cri-o://e8191df8902be4a4da6f0e247d64bea9567a78aed70ff1b1918dad0f09a75382" gracePeriod=10 Mar 13 10:23:58 crc kubenswrapper[4632]: I0313 10:23:58.944925 4632 generic.go:334] "Generic (PLEG): container finished" podID="df11136a-b7d7-4a5a-a0cc-d0ebbc069b34" containerID="e8191df8902be4a4da6f0e247d64bea9567a78aed70ff1b1918dad0f09a75382" exitCode=0 Mar 13 10:23:58 crc kubenswrapper[4632]: I0313 10:23:58.945622 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" event={"ID":"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34","Type":"ContainerDied","Data":"e8191df8902be4a4da6f0e247d64bea9567a78aed70ff1b1918dad0f09a75382"} Mar 13 10:23:58 crc kubenswrapper[4632]: I0313 10:23:58.946395 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:58 crc kubenswrapper[4632]: I0313 10:23:58.946482 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.016030 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.067043 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.211833 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.295519 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-config\") pod \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.295637 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-ovsdbserver-nb\") pod \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.295696 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7phzq\" (UniqueName: \"kubernetes.io/projected/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-kube-api-access-7phzq\") pod \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.295715 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-dns-svc\") pod \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\" (UID: \"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34\") " Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.309354 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-kube-api-access-7phzq" (OuterVolumeSpecName: "kube-api-access-7phzq") pod "df11136a-b7d7-4a5a-a0cc-d0ebbc069b34" (UID: "df11136a-b7d7-4a5a-a0cc-d0ebbc069b34"). InnerVolumeSpecName "kube-api-access-7phzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.334477 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "df11136a-b7d7-4a5a-a0cc-d0ebbc069b34" (UID: "df11136a-b7d7-4a5a-a0cc-d0ebbc069b34"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.349516 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "df11136a-b7d7-4a5a-a0cc-d0ebbc069b34" (UID: "df11136a-b7d7-4a5a-a0cc-d0ebbc069b34"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.352598 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-config" (OuterVolumeSpecName: "config") pod "df11136a-b7d7-4a5a-a0cc-d0ebbc069b34" (UID: "df11136a-b7d7-4a5a-a0cc-d0ebbc069b34"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.373545 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.397815 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.397853 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.397871 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7phzq\" (UniqueName: \"kubernetes.io/projected/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-kube-api-access-7phzq\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.397883 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.416133 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.955112 4632 generic.go:334] "Generic (PLEG): container finished" podID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerID="aa55e67bc4a499dbbd5e317a3b00bbfa4846877aed563132955d84afe5164371" exitCode=0 Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.955221 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2cb2f546-c8c5-4ec9-aba8-d3782431de10","Type":"ContainerDied","Data":"aa55e67bc4a499dbbd5e317a3b00bbfa4846877aed563132955d84afe5164371"} Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.958483 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" event={"ID":"df11136a-b7d7-4a5a-a0cc-d0ebbc069b34","Type":"ContainerDied","Data":"c5cd8fa4d9fefd8144c63785679d7e91126ba9c71436b6d101b2ee0cc8ea3019"} Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.958550 4632 scope.go:117] "RemoveContainer" containerID="e8191df8902be4a4da6f0e247d64bea9567a78aed70ff1b1918dad0f09a75382" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.958674 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7df696cbbf-4tc7r" Mar 13 10:23:59 crc kubenswrapper[4632]: I0313 10:23:59.959183 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.097101 4632 scope.go:117] "RemoveContainer" containerID="2915f2c9d1176c44121e26cd23d0bb33a0c1b7ffaf310d2e0081ce7eb76b0909" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.097790 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7df696cbbf-4tc7r"] Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.110493 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7df696cbbf-4tc7r"] Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.161728 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556624-tnl4c"] Mar 13 10:24:00 crc kubenswrapper[4632]: E0313 10:24:00.162189 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df11136a-b7d7-4a5a-a0cc-d0ebbc069b34" containerName="init" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.162211 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="df11136a-b7d7-4a5a-a0cc-d0ebbc069b34" containerName="init" Mar 13 10:24:00 crc kubenswrapper[4632]: E0313 10:24:00.162243 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df11136a-b7d7-4a5a-a0cc-d0ebbc069b34" containerName="dnsmasq-dns" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.162251 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="df11136a-b7d7-4a5a-a0cc-d0ebbc069b34" containerName="dnsmasq-dns" Mar 13 10:24:00 crc kubenswrapper[4632]: E0313 10:24:00.162274 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7afad0f9-c29c-40e6-8605-1df67a505a82" containerName="init" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.162282 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7afad0f9-c29c-40e6-8605-1df67a505a82" containerName="init" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.162478 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="7afad0f9-c29c-40e6-8605-1df67a505a82" containerName="init" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.162500 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="df11136a-b7d7-4a5a-a0cc-d0ebbc069b34" containerName="dnsmasq-dns" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.163174 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556624-tnl4c" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.166345 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.166564 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.167134 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.202184 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556624-tnl4c"] Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.318008 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52gwq\" (UniqueName: \"kubernetes.io/projected/5b2374d5-8d19-4837-8d91-79df0e65fc1f-kube-api-access-52gwq\") pod \"auto-csr-approver-29556624-tnl4c\" (UID: \"5b2374d5-8d19-4837-8d91-79df0e65fc1f\") " pod="openshift-infra/auto-csr-approver-29556624-tnl4c" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.419059 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.419518 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52gwq\" (UniqueName: \"kubernetes.io/projected/5b2374d5-8d19-4837-8d91-79df0e65fc1f-kube-api-access-52gwq\") pod \"auto-csr-approver-29556624-tnl4c\" (UID: \"5b2374d5-8d19-4837-8d91-79df0e65fc1f\") " pod="openshift-infra/auto-csr-approver-29556624-tnl4c" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.444258 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52gwq\" (UniqueName: \"kubernetes.io/projected/5b2374d5-8d19-4837-8d91-79df0e65fc1f-kube-api-access-52gwq\") pod \"auto-csr-approver-29556624-tnl4c\" (UID: \"5b2374d5-8d19-4837-8d91-79df0e65fc1f\") " pod="openshift-infra/auto-csr-approver-29556624-tnl4c" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.493283 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556624-tnl4c" Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.938490 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556624-tnl4c"] Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.972058 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2cb2f546-c8c5-4ec9-aba8-d3782431de10","Type":"ContainerStarted","Data":"cdf295326a62a01129c4f9b5741f57b3d80d103e3c5a6bf64f5cc1951034264f"} Mar 13 10:24:00 crc kubenswrapper[4632]: I0313 10:24:00.974459 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556624-tnl4c" event={"ID":"5b2374d5-8d19-4837-8d91-79df0e65fc1f","Type":"ContainerStarted","Data":"c1fe076f0e8c292e4cfd8e3aab67afc57f3617492b69247c418c8f293cfd491f"} Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.007434 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=-9223371985.847382 podStartE2EDuration="51.007393999s" podCreationTimestamp="2026-03-13 10:23:10 +0000 UTC" firstStartedPulling="2026-03-13 10:23:13.194067668 +0000 UTC m=+1167.216597801" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:00.999040724 +0000 UTC m=+1215.021570847" watchObservedRunningTime="2026-03-13 10:24:01.007393999 +0000 UTC m=+1215.029924132" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.027830 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.329575 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.331125 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.336576 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.337654 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.338015 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.338205 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-6rcwx" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.364471 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.443850 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a169306-9d47-41ae-8667-1efb89c43d82-config\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.444077 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a169306-9d47-41ae-8667-1efb89c43d82-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.444117 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9a169306-9d47-41ae-8667-1efb89c43d82-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.444142 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znvdq\" (UniqueName: \"kubernetes.io/projected/9a169306-9d47-41ae-8667-1efb89c43d82-kube-api-access-znvdq\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.444342 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a169306-9d47-41ae-8667-1efb89c43d82-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.444398 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a169306-9d47-41ae-8667-1efb89c43d82-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.444432 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a169306-9d47-41ae-8667-1efb89c43d82-scripts\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.546189 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a169306-9d47-41ae-8667-1efb89c43d82-config\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.546273 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a169306-9d47-41ae-8667-1efb89c43d82-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.546311 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9a169306-9d47-41ae-8667-1efb89c43d82-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.546343 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znvdq\" (UniqueName: \"kubernetes.io/projected/9a169306-9d47-41ae-8667-1efb89c43d82-kube-api-access-znvdq\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.546867 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a169306-9d47-41ae-8667-1efb89c43d82-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.547010 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/9a169306-9d47-41ae-8667-1efb89c43d82-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.547284 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a169306-9d47-41ae-8667-1efb89c43d82-config\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.547551 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a169306-9d47-41ae-8667-1efb89c43d82-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.547581 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a169306-9d47-41ae-8667-1efb89c43d82-scripts\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.548537 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/9a169306-9d47-41ae-8667-1efb89c43d82-scripts\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.552329 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a169306-9d47-41ae-8667-1efb89c43d82-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.556238 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9a169306-9d47-41ae-8667-1efb89c43d82-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.564884 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/9a169306-9d47-41ae-8667-1efb89c43d82-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.575911 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znvdq\" (UniqueName: \"kubernetes.io/projected/9a169306-9d47-41ae-8667-1efb89c43d82-kube-api-access-znvdq\") pod \"ovn-northd-0\" (UID: \"9a169306-9d47-41ae-8667-1efb89c43d82\") " pod="openstack/ovn-northd-0" Mar 13 10:24:01 crc kubenswrapper[4632]: I0313 10:24:01.650929 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 10:24:02 crc kubenswrapper[4632]: I0313 10:24:02.055670 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df11136a-b7d7-4a5a-a0cc-d0ebbc069b34" path="/var/lib/kubelet/pods/df11136a-b7d7-4a5a-a0cc-d0ebbc069b34/volumes" Mar 13 10:24:02 crc kubenswrapper[4632]: I0313 10:24:02.231863 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 10:24:02 crc kubenswrapper[4632]: W0313 10:24:02.232372 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a169306_9d47_41ae_8667_1efb89c43d82.slice/crio-1761d9710c7607181cd1ab28cdd44ec208e3e218f4f1cf1283e7637db0199561 WatchSource:0}: Error finding container 1761d9710c7607181cd1ab28cdd44ec208e3e218f4f1cf1283e7637db0199561: Status 404 returned error can't find the container with id 1761d9710c7607181cd1ab28cdd44ec208e3e218f4f1cf1283e7637db0199561 Mar 13 10:24:02 crc kubenswrapper[4632]: I0313 10:24:02.477240 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 13 10:24:02 crc kubenswrapper[4632]: I0313 10:24:02.477845 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 13 10:24:03 crc kubenswrapper[4632]: I0313 10:24:03.016891 4632 generic.go:334] "Generic (PLEG): container finished" podID="5b2374d5-8d19-4837-8d91-79df0e65fc1f" containerID="6257821be47ec7e5943095f3b1d29a6e6fd0a1190515cb74642f7cb762d806d1" exitCode=0 Mar 13 10:24:03 crc kubenswrapper[4632]: I0313 10:24:03.016983 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556624-tnl4c" event={"ID":"5b2374d5-8d19-4837-8d91-79df0e65fc1f","Type":"ContainerDied","Data":"6257821be47ec7e5943095f3b1d29a6e6fd0a1190515cb74642f7cb762d806d1"} Mar 13 10:24:03 crc kubenswrapper[4632]: I0313 10:24:03.018630 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9a169306-9d47-41ae-8667-1efb89c43d82","Type":"ContainerStarted","Data":"1761d9710c7607181cd1ab28cdd44ec208e3e218f4f1cf1283e7637db0199561"} Mar 13 10:24:03 crc kubenswrapper[4632]: I0313 10:24:03.734849 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 13 10:24:03 crc kubenswrapper[4632]: I0313 10:24:03.735854 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 13 10:24:03 crc kubenswrapper[4632]: I0313 10:24:03.799145 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 13 10:24:03 crc kubenswrapper[4632]: I0313 10:24:03.846817 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 13 10:24:04 crc kubenswrapper[4632]: I0313 10:24:04.043608 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9a169306-9d47-41ae-8667-1efb89c43d82","Type":"ContainerStarted","Data":"31e7d5bdbc4af9e778f12b375dde292167a77cf66ef9c81c12db83b3be575a88"} Mar 13 10:24:04 crc kubenswrapper[4632]: I0313 10:24:04.043709 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"9a169306-9d47-41ae-8667-1efb89c43d82","Type":"ContainerStarted","Data":"37e8b2b92022c84ec3213a38ce8f6e0c39ef2af8b108102efff96d21fcecf14a"} Mar 13 10:24:04 crc kubenswrapper[4632]: I0313 10:24:04.056026 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 13 10:24:04 crc kubenswrapper[4632]: I0313 10:24:04.080887 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=2.297183455 podStartE2EDuration="3.080863182s" podCreationTimestamp="2026-03-13 10:24:01 +0000 UTC" firstStartedPulling="2026-03-13 10:24:02.234511104 +0000 UTC m=+1216.257041237" lastFinishedPulling="2026-03-13 10:24:03.018190831 +0000 UTC m=+1217.040720964" observedRunningTime="2026-03-13 10:24:04.077281175 +0000 UTC m=+1218.099811328" watchObservedRunningTime="2026-03-13 10:24:04.080863182 +0000 UTC m=+1218.103393325" Mar 13 10:24:04 crc kubenswrapper[4632]: I0313 10:24:04.174136 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 13 10:24:04 crc kubenswrapper[4632]: I0313 10:24:04.441162 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556624-tnl4c" Mar 13 10:24:04 crc kubenswrapper[4632]: I0313 10:24:04.636295 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52gwq\" (UniqueName: \"kubernetes.io/projected/5b2374d5-8d19-4837-8d91-79df0e65fc1f-kube-api-access-52gwq\") pod \"5b2374d5-8d19-4837-8d91-79df0e65fc1f\" (UID: \"5b2374d5-8d19-4837-8d91-79df0e65fc1f\") " Mar 13 10:24:04 crc kubenswrapper[4632]: I0313 10:24:04.644774 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b2374d5-8d19-4837-8d91-79df0e65fc1f-kube-api-access-52gwq" (OuterVolumeSpecName: "kube-api-access-52gwq") pod "5b2374d5-8d19-4837-8d91-79df0e65fc1f" (UID: "5b2374d5-8d19-4837-8d91-79df0e65fc1f"). InnerVolumeSpecName "kube-api-access-52gwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:04 crc kubenswrapper[4632]: I0313 10:24:04.738491 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52gwq\" (UniqueName: \"kubernetes.io/projected/5b2374d5-8d19-4837-8d91-79df0e65fc1f-kube-api-access-52gwq\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:05 crc kubenswrapper[4632]: I0313 10:24:05.047061 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556624-tnl4c" Mar 13 10:24:05 crc kubenswrapper[4632]: I0313 10:24:05.048032 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556624-tnl4c" event={"ID":"5b2374d5-8d19-4837-8d91-79df0e65fc1f","Type":"ContainerDied","Data":"c1fe076f0e8c292e4cfd8e3aab67afc57f3617492b69247c418c8f293cfd491f"} Mar 13 10:24:05 crc kubenswrapper[4632]: I0313 10:24:05.048056 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1fe076f0e8c292e4cfd8e3aab67afc57f3617492b69247c418c8f293cfd491f" Mar 13 10:24:05 crc kubenswrapper[4632]: I0313 10:24:05.090205 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 13 10:24:05 crc kubenswrapper[4632]: I0313 10:24:05.184522 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 13 10:24:05 crc kubenswrapper[4632]: I0313 10:24:05.510000 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556618-ngbmk"] Mar 13 10:24:05 crc kubenswrapper[4632]: I0313 10:24:05.518552 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556618-ngbmk"] Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.056367 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b93f1106-edf9-4cde-9acb-e265d8e07191" path="/var/lib/kubelet/pods/b93f1106-edf9-4cde-9acb-e265d8e07191/volumes" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.396013 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b59dbc87f-7zwrj"] Mar 13 10:24:06 crc kubenswrapper[4632]: E0313 10:24:06.396333 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b2374d5-8d19-4837-8d91-79df0e65fc1f" containerName="oc" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.396349 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b2374d5-8d19-4837-8d91-79df0e65fc1f" containerName="oc" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.396520 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b2374d5-8d19-4837-8d91-79df0e65fc1f" containerName="oc" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.397398 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.430503 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b59dbc87f-7zwrj"] Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.572045 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-ovsdbserver-nb\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.572113 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-dns-svc\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.572156 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r95hn\" (UniqueName: \"kubernetes.io/projected/7203640d-964c-4c28-8cc2-6a7ae27cdab3-kube-api-access-r95hn\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.572207 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-config\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.572240 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-ovsdbserver-sb\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.673441 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-config\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.673527 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-ovsdbserver-sb\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.673568 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-ovsdbserver-nb\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.673639 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-dns-svc\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.673707 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r95hn\" (UniqueName: \"kubernetes.io/projected/7203640d-964c-4c28-8cc2-6a7ae27cdab3-kube-api-access-r95hn\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.674733 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-config\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.674782 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-ovsdbserver-nb\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.675643 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-ovsdbserver-sb\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.675658 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-dns-svc\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.694265 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r95hn\" (UniqueName: \"kubernetes.io/projected/7203640d-964c-4c28-8cc2-6a7ae27cdab3-kube-api-access-r95hn\") pod \"dnsmasq-dns-5b59dbc87f-7zwrj\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:06 crc kubenswrapper[4632]: I0313 10:24:06.718051 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.231457 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b59dbc87f-7zwrj"] Mar 13 10:24:07 crc kubenswrapper[4632]: W0313 10:24:07.238344 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7203640d_964c_4c28_8cc2_6a7ae27cdab3.slice/crio-501cdcd9d1f38a4b8b82ad7d76e2b6765f391cfadd65ee750e8254d78d76de84 WatchSource:0}: Error finding container 501cdcd9d1f38a4b8b82ad7d76e2b6765f391cfadd65ee750e8254d78d76de84: Status 404 returned error can't find the container with id 501cdcd9d1f38a4b8b82ad7d76e2b6765f391cfadd65ee750e8254d78d76de84 Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.662370 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.672409 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.678342 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.678340 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.678572 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-nmkc7" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.678817 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.734214 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.793422 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-lock\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.793655 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.793735 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65nwv\" (UniqueName: \"kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-kube-api-access-65nwv\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.793843 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.793873 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.794103 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-cache\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.895333 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-cache\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.895683 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-lock\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.895821 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.895918 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65nwv\" (UniqueName: \"kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-kube-api-access-65nwv\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.895992 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-lock\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.895863 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-cache\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: E0313 10:24:07.896153 4632 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 10:24:07 crc kubenswrapper[4632]: E0313 10:24:07.896250 4632 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 10:24:07 crc kubenswrapper[4632]: E0313 10:24:07.896350 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift podName:e37b3d77-de2e-4be9-9984-550d4ba0f2f0 nodeName:}" failed. No retries permitted until 2026-03-13 10:24:08.39631989 +0000 UTC m=+1222.418850023 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift") pod "swift-storage-0" (UID: "e37b3d77-de2e-4be9-9984-550d4ba0f2f0") : configmap "swift-ring-files" not found Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.896676 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.896730 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.897434 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.904029 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.921770 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65nwv\" (UniqueName: \"kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-kube-api-access-65nwv\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:07 crc kubenswrapper[4632]: I0313 10:24:07.926979 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:08 crc kubenswrapper[4632]: I0313 10:24:08.071596 4632 generic.go:334] "Generic (PLEG): container finished" podID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerID="f74cf11731f4fec2422112ef6bdd1e43cc133692a8363ef95d5bb5847ffb0fd1" exitCode=0 Mar 13 10:24:08 crc kubenswrapper[4632]: I0313 10:24:08.071660 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" event={"ID":"7203640d-964c-4c28-8cc2-6a7ae27cdab3","Type":"ContainerDied","Data":"f74cf11731f4fec2422112ef6bdd1e43cc133692a8363ef95d5bb5847ffb0fd1"} Mar 13 10:24:08 crc kubenswrapper[4632]: I0313 10:24:08.071692 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" event={"ID":"7203640d-964c-4c28-8cc2-6a7ae27cdab3","Type":"ContainerStarted","Data":"501cdcd9d1f38a4b8b82ad7d76e2b6765f391cfadd65ee750e8254d78d76de84"} Mar 13 10:24:08 crc kubenswrapper[4632]: I0313 10:24:08.404209 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:08 crc kubenswrapper[4632]: E0313 10:24:08.404425 4632 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 10:24:08 crc kubenswrapper[4632]: E0313 10:24:08.404728 4632 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 10:24:08 crc kubenswrapper[4632]: E0313 10:24:08.404799 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift podName:e37b3d77-de2e-4be9-9984-550d4ba0f2f0 nodeName:}" failed. No retries permitted until 2026-03-13 10:24:09.404777604 +0000 UTC m=+1223.427307737 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift") pod "swift-storage-0" (UID: "e37b3d77-de2e-4be9-9984-550d4ba0f2f0") : configmap "swift-ring-files" not found Mar 13 10:24:09 crc kubenswrapper[4632]: I0313 10:24:09.081082 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" event={"ID":"7203640d-964c-4c28-8cc2-6a7ae27cdab3","Type":"ContainerStarted","Data":"f1255f2b0d97d7bcc13a7045fc5d8e4778eece89f9f6f1d468ae8c05e428c6f7"} Mar 13 10:24:09 crc kubenswrapper[4632]: I0313 10:24:09.081458 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:09 crc kubenswrapper[4632]: I0313 10:24:09.102172 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" podStartSLOduration=3.102152103 podStartE2EDuration="3.102152103s" podCreationTimestamp="2026-03-13 10:24:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:09.098136646 +0000 UTC m=+1223.120666789" watchObservedRunningTime="2026-03-13 10:24:09.102152103 +0000 UTC m=+1223.124682236" Mar 13 10:24:09 crc kubenswrapper[4632]: I0313 10:24:09.422813 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:09 crc kubenswrapper[4632]: E0313 10:24:09.423000 4632 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 10:24:09 crc kubenswrapper[4632]: E0313 10:24:09.423713 4632 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 10:24:09 crc kubenswrapper[4632]: E0313 10:24:09.423776 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift podName:e37b3d77-de2e-4be9-9984-550d4ba0f2f0 nodeName:}" failed. No retries permitted until 2026-03-13 10:24:11.423756293 +0000 UTC m=+1225.446286426 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift") pod "swift-storage-0" (UID: "e37b3d77-de2e-4be9-9984-550d4ba0f2f0") : configmap "swift-ring-files" not found Mar 13 10:24:10 crc kubenswrapper[4632]: I0313 10:24:10.461165 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:24:10 crc kubenswrapper[4632]: I0313 10:24:10.461241 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:24:10 crc kubenswrapper[4632]: I0313 10:24:10.827676 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-c8lh5"] Mar 13 10:24:10 crc kubenswrapper[4632]: I0313 10:24:10.828816 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c8lh5" Mar 13 10:24:10 crc kubenswrapper[4632]: I0313 10:24:10.832055 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 13 10:24:10 crc kubenswrapper[4632]: I0313 10:24:10.865339 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-c8lh5"] Mar 13 10:24:10 crc kubenswrapper[4632]: I0313 10:24:10.951306 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc41c555-17e5-4785-a003-3f8e9f10d799-operator-scripts\") pod \"root-account-create-update-c8lh5\" (UID: \"cc41c555-17e5-4785-a003-3f8e9f10d799\") " pod="openstack/root-account-create-update-c8lh5" Mar 13 10:24:10 crc kubenswrapper[4632]: I0313 10:24:10.952793 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvft2\" (UniqueName: \"kubernetes.io/projected/cc41c555-17e5-4785-a003-3f8e9f10d799-kube-api-access-gvft2\") pod \"root-account-create-update-c8lh5\" (UID: \"cc41c555-17e5-4785-a003-3f8e9f10d799\") " pod="openstack/root-account-create-update-c8lh5" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.055321 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvft2\" (UniqueName: \"kubernetes.io/projected/cc41c555-17e5-4785-a003-3f8e9f10d799-kube-api-access-gvft2\") pod \"root-account-create-update-c8lh5\" (UID: \"cc41c555-17e5-4785-a003-3f8e9f10d799\") " pod="openstack/root-account-create-update-c8lh5" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.055706 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc41c555-17e5-4785-a003-3f8e9f10d799-operator-scripts\") pod \"root-account-create-update-c8lh5\" (UID: \"cc41c555-17e5-4785-a003-3f8e9f10d799\") " pod="openstack/root-account-create-update-c8lh5" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.056667 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc41c555-17e5-4785-a003-3f8e9f10d799-operator-scripts\") pod \"root-account-create-update-c8lh5\" (UID: \"cc41c555-17e5-4785-a003-3f8e9f10d799\") " pod="openstack/root-account-create-update-c8lh5" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.078918 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvft2\" (UniqueName: \"kubernetes.io/projected/cc41c555-17e5-4785-a003-3f8e9f10d799-kube-api-access-gvft2\") pod \"root-account-create-update-c8lh5\" (UID: \"cc41c555-17e5-4785-a003-3f8e9f10d799\") " pod="openstack/root-account-create-update-c8lh5" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.161392 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c8lh5" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.451274 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-mkdcg"] Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.455725 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.460108 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.460848 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.463683 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-mkdcg"] Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.464013 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.464015 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:11 crc kubenswrapper[4632]: E0313 10:24:11.464285 4632 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 10:24:11 crc kubenswrapper[4632]: E0313 10:24:11.464306 4632 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 10:24:11 crc kubenswrapper[4632]: E0313 10:24:11.464363 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift podName:e37b3d77-de2e-4be9-9984-550d4ba0f2f0 nodeName:}" failed. No retries permitted until 2026-03-13 10:24:15.464346907 +0000 UTC m=+1229.486877040 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift") pod "swift-storage-0" (UID: "e37b3d77-de2e-4be9-9984-550d4ba0f2f0") : configmap "swift-ring-files" not found Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.566327 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bc39c52e-008f-40c1-b93b-532707127fcd-ring-data-devices\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.566581 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bc39c52e-008f-40c1-b93b-532707127fcd-etc-swift\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.566622 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-dispersionconf\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.566721 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-swiftconf\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.566749 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7f9b\" (UniqueName: \"kubernetes.io/projected/bc39c52e-008f-40c1-b93b-532707127fcd-kube-api-access-v7f9b\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.566816 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc39c52e-008f-40c1-b93b-532707127fcd-scripts\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.566836 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-combined-ca-bundle\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.615154 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-c8lh5"] Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.671827 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bc39c52e-008f-40c1-b93b-532707127fcd-ring-data-devices\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.671876 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bc39c52e-008f-40c1-b93b-532707127fcd-etc-swift\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.671895 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-dispersionconf\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.672854 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bc39c52e-008f-40c1-b93b-532707127fcd-ring-data-devices\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.673099 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bc39c52e-008f-40c1-b93b-532707127fcd-etc-swift\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.674783 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-swiftconf\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.674847 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7f9b\" (UniqueName: \"kubernetes.io/projected/bc39c52e-008f-40c1-b93b-532707127fcd-kube-api-access-v7f9b\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.674929 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc39c52e-008f-40c1-b93b-532707127fcd-scripts\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.674969 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-combined-ca-bundle\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.676205 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc39c52e-008f-40c1-b93b-532707127fcd-scripts\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.688450 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-dispersionconf\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.688860 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-combined-ca-bundle\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.689846 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-swiftconf\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.697761 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7f9b\" (UniqueName: \"kubernetes.io/projected/bc39c52e-008f-40c1-b93b-532707127fcd-kube-api-access-v7f9b\") pod \"swift-ring-rebalance-mkdcg\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:11 crc kubenswrapper[4632]: I0313 10:24:11.778784 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:12 crc kubenswrapper[4632]: I0313 10:24:12.107878 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-c8lh5" event={"ID":"cc41c555-17e5-4785-a003-3f8e9f10d799","Type":"ContainerStarted","Data":"9011fe3e8ff19daa76b8d8bddf336d224d69f10272938404d994caa9a1a4d6ee"} Mar 13 10:24:12 crc kubenswrapper[4632]: I0313 10:24:12.108392 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-c8lh5" event={"ID":"cc41c555-17e5-4785-a003-3f8e9f10d799","Type":"ContainerStarted","Data":"0b853f73f9460c789c867abbd82e3ae379a221406d460cfe9764bd3d4e71050b"} Mar 13 10:24:12 crc kubenswrapper[4632]: I0313 10:24:12.217809 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-c8lh5" podStartSLOduration=2.217788002 podStartE2EDuration="2.217788002s" podCreationTimestamp="2026-03-13 10:24:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:12.13333445 +0000 UTC m=+1226.155864593" watchObservedRunningTime="2026-03-13 10:24:12.217788002 +0000 UTC m=+1226.240318135" Mar 13 10:24:12 crc kubenswrapper[4632]: I0313 10:24:12.227542 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-mkdcg"] Mar 13 10:24:13 crc kubenswrapper[4632]: I0313 10:24:13.116177 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mkdcg" event={"ID":"bc39c52e-008f-40c1-b93b-532707127fcd","Type":"ContainerStarted","Data":"b9c30e7d71115b424270718b1169d7d7c69bba98c01559c66a08b1a331e3ccdd"} Mar 13 10:24:13 crc kubenswrapper[4632]: I0313 10:24:13.119176 4632 generic.go:334] "Generic (PLEG): container finished" podID="cc41c555-17e5-4785-a003-3f8e9f10d799" containerID="9011fe3e8ff19daa76b8d8bddf336d224d69f10272938404d994caa9a1a4d6ee" exitCode=0 Mar 13 10:24:13 crc kubenswrapper[4632]: I0313 10:24:13.119244 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-c8lh5" event={"ID":"cc41c555-17e5-4785-a003-3f8e9f10d799","Type":"ContainerDied","Data":"9011fe3e8ff19daa76b8d8bddf336d224d69f10272938404d994caa9a1a4d6ee"} Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.001113 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-7hqpw"] Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.002671 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7hqpw" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.010649 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-7hqpw"] Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.130774 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-9698-account-create-update-9kfhv"] Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.137816 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9698-account-create-update-9kfhv" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.142992 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wszzh\" (UniqueName: \"kubernetes.io/projected/584d2818-4b22-468f-b296-bd1850c7915b-kube-api-access-wszzh\") pod \"glance-db-create-7hqpw\" (UID: \"584d2818-4b22-468f-b296-bd1850c7915b\") " pod="openstack/glance-db-create-7hqpw" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.143116 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/584d2818-4b22-468f-b296-bd1850c7915b-operator-scripts\") pod \"glance-db-create-7hqpw\" (UID: \"584d2818-4b22-468f-b296-bd1850c7915b\") " pod="openstack/glance-db-create-7hqpw" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.143570 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.150668 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9698-account-create-update-9kfhv"] Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.245541 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wszzh\" (UniqueName: \"kubernetes.io/projected/584d2818-4b22-468f-b296-bd1850c7915b-kube-api-access-wszzh\") pod \"glance-db-create-7hqpw\" (UID: \"584d2818-4b22-468f-b296-bd1850c7915b\") " pod="openstack/glance-db-create-7hqpw" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.245623 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/584d2818-4b22-468f-b296-bd1850c7915b-operator-scripts\") pod \"glance-db-create-7hqpw\" (UID: \"584d2818-4b22-468f-b296-bd1850c7915b\") " pod="openstack/glance-db-create-7hqpw" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.245696 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2610abab-1da4-4912-9ca7-f2aa2d7c0486-operator-scripts\") pod \"glance-9698-account-create-update-9kfhv\" (UID: \"2610abab-1da4-4912-9ca7-f2aa2d7c0486\") " pod="openstack/glance-9698-account-create-update-9kfhv" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.245858 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29xj8\" (UniqueName: \"kubernetes.io/projected/2610abab-1da4-4912-9ca7-f2aa2d7c0486-kube-api-access-29xj8\") pod \"glance-9698-account-create-update-9kfhv\" (UID: \"2610abab-1da4-4912-9ca7-f2aa2d7c0486\") " pod="openstack/glance-9698-account-create-update-9kfhv" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.247873 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/584d2818-4b22-468f-b296-bd1850c7915b-operator-scripts\") pod \"glance-db-create-7hqpw\" (UID: \"584d2818-4b22-468f-b296-bd1850c7915b\") " pod="openstack/glance-db-create-7hqpw" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.283661 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wszzh\" (UniqueName: \"kubernetes.io/projected/584d2818-4b22-468f-b296-bd1850c7915b-kube-api-access-wszzh\") pod \"glance-db-create-7hqpw\" (UID: \"584d2818-4b22-468f-b296-bd1850c7915b\") " pod="openstack/glance-db-create-7hqpw" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.347883 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2610abab-1da4-4912-9ca7-f2aa2d7c0486-operator-scripts\") pod \"glance-9698-account-create-update-9kfhv\" (UID: \"2610abab-1da4-4912-9ca7-f2aa2d7c0486\") " pod="openstack/glance-9698-account-create-update-9kfhv" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.348085 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29xj8\" (UniqueName: \"kubernetes.io/projected/2610abab-1da4-4912-9ca7-f2aa2d7c0486-kube-api-access-29xj8\") pod \"glance-9698-account-create-update-9kfhv\" (UID: \"2610abab-1da4-4912-9ca7-f2aa2d7c0486\") " pod="openstack/glance-9698-account-create-update-9kfhv" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.349306 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2610abab-1da4-4912-9ca7-f2aa2d7c0486-operator-scripts\") pod \"glance-9698-account-create-update-9kfhv\" (UID: \"2610abab-1da4-4912-9ca7-f2aa2d7c0486\") " pod="openstack/glance-9698-account-create-update-9kfhv" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.362448 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7hqpw" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.373659 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29xj8\" (UniqueName: \"kubernetes.io/projected/2610abab-1da4-4912-9ca7-f2aa2d7c0486-kube-api-access-29xj8\") pod \"glance-9698-account-create-update-9kfhv\" (UID: \"2610abab-1da4-4912-9ca7-f2aa2d7c0486\") " pod="openstack/glance-9698-account-create-update-9kfhv" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.476449 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9698-account-create-update-9kfhv" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.760874 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-bfb6b"] Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.761989 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bfb6b" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.782684 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bfb6b"] Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.828121 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c8lh5" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.863484 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e353045-e09b-4cd2-b659-1954485ec8db-operator-scripts\") pod \"keystone-db-create-bfb6b\" (UID: \"8e353045-e09b-4cd2-b659-1954485ec8db\") " pod="openstack/keystone-db-create-bfb6b" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.863727 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjh6b\" (UniqueName: \"kubernetes.io/projected/8e353045-e09b-4cd2-b659-1954485ec8db-kube-api-access-fjh6b\") pod \"keystone-db-create-bfb6b\" (UID: \"8e353045-e09b-4cd2-b659-1954485ec8db\") " pod="openstack/keystone-db-create-bfb6b" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.901559 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-ab0c-account-create-update-tr7hx"] Mar 13 10:24:14 crc kubenswrapper[4632]: E0313 10:24:14.901962 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc41c555-17e5-4785-a003-3f8e9f10d799" containerName="mariadb-account-create-update" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.901987 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc41c555-17e5-4785-a003-3f8e9f10d799" containerName="mariadb-account-create-update" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.902369 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc41c555-17e5-4785-a003-3f8e9f10d799" containerName="mariadb-account-create-update" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.903175 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ab0c-account-create-update-tr7hx" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.907649 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.922893 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ab0c-account-create-update-tr7hx"] Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.965767 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc41c555-17e5-4785-a003-3f8e9f10d799-operator-scripts\") pod \"cc41c555-17e5-4785-a003-3f8e9f10d799\" (UID: \"cc41c555-17e5-4785-a003-3f8e9f10d799\") " Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.965921 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gvft2\" (UniqueName: \"kubernetes.io/projected/cc41c555-17e5-4785-a003-3f8e9f10d799-kube-api-access-gvft2\") pod \"cc41c555-17e5-4785-a003-3f8e9f10d799\" (UID: \"cc41c555-17e5-4785-a003-3f8e9f10d799\") " Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.966252 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjh6b\" (UniqueName: \"kubernetes.io/projected/8e353045-e09b-4cd2-b659-1954485ec8db-kube-api-access-fjh6b\") pod \"keystone-db-create-bfb6b\" (UID: \"8e353045-e09b-4cd2-b659-1954485ec8db\") " pod="openstack/keystone-db-create-bfb6b" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.966338 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e353045-e09b-4cd2-b659-1954485ec8db-operator-scripts\") pod \"keystone-db-create-bfb6b\" (UID: \"8e353045-e09b-4cd2-b659-1954485ec8db\") " pod="openstack/keystone-db-create-bfb6b" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.969709 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc41c555-17e5-4785-a003-3f8e9f10d799-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cc41c555-17e5-4785-a003-3f8e9f10d799" (UID: "cc41c555-17e5-4785-a003-3f8e9f10d799"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.970692 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e353045-e09b-4cd2-b659-1954485ec8db-operator-scripts\") pod \"keystone-db-create-bfb6b\" (UID: \"8e353045-e09b-4cd2-b659-1954485ec8db\") " pod="openstack/keystone-db-create-bfb6b" Mar 13 10:24:14 crc kubenswrapper[4632]: I0313 10:24:14.989419 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc41c555-17e5-4785-a003-3f8e9f10d799-kube-api-access-gvft2" (OuterVolumeSpecName: "kube-api-access-gvft2") pod "cc41c555-17e5-4785-a003-3f8e9f10d799" (UID: "cc41c555-17e5-4785-a003-3f8e9f10d799"). InnerVolumeSpecName "kube-api-access-gvft2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.013969 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjh6b\" (UniqueName: \"kubernetes.io/projected/8e353045-e09b-4cd2-b659-1954485ec8db-kube-api-access-fjh6b\") pod \"keystone-db-create-bfb6b\" (UID: \"8e353045-e09b-4cd2-b659-1954485ec8db\") " pod="openstack/keystone-db-create-bfb6b" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.014084 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-64xvf"] Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.015706 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-64xvf" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.025400 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-64xvf"] Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.072384 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpvp9\" (UniqueName: \"kubernetes.io/projected/6c84aa49-2900-4a14-b81b-bb03e925d1b7-kube-api-access-wpvp9\") pod \"keystone-ab0c-account-create-update-tr7hx\" (UID: \"6c84aa49-2900-4a14-b81b-bb03e925d1b7\") " pod="openstack/keystone-ab0c-account-create-update-tr7hx" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.077174 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c84aa49-2900-4a14-b81b-bb03e925d1b7-operator-scripts\") pod \"keystone-ab0c-account-create-update-tr7hx\" (UID: \"6c84aa49-2900-4a14-b81b-bb03e925d1b7\") " pod="openstack/keystone-ab0c-account-create-update-tr7hx" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.077673 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc41c555-17e5-4785-a003-3f8e9f10d799-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.077701 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gvft2\" (UniqueName: \"kubernetes.io/projected/cc41c555-17e5-4785-a003-3f8e9f10d799-kube-api-access-gvft2\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.102510 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-a750-account-create-update-7wk26"] Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.104426 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a750-account-create-update-7wk26" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.108911 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.114829 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a750-account-create-update-7wk26"] Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.151562 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bfb6b" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.155233 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-c8lh5" event={"ID":"cc41c555-17e5-4785-a003-3f8e9f10d799","Type":"ContainerDied","Data":"0b853f73f9460c789c867abbd82e3ae379a221406d460cfe9764bd3d4e71050b"} Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.155283 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0b853f73f9460c789c867abbd82e3ae379a221406d460cfe9764bd3d4e71050b" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.155339 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-c8lh5" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.180226 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlcdq\" (UniqueName: \"kubernetes.io/projected/5f09e2f4-4f82-4388-9b5a-a9e890d3a950-kube-api-access-mlcdq\") pod \"placement-db-create-64xvf\" (UID: \"5f09e2f4-4f82-4388-9b5a-a9e890d3a950\") " pod="openstack/placement-db-create-64xvf" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.180348 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c84aa49-2900-4a14-b81b-bb03e925d1b7-operator-scripts\") pod \"keystone-ab0c-account-create-update-tr7hx\" (UID: \"6c84aa49-2900-4a14-b81b-bb03e925d1b7\") " pod="openstack/keystone-ab0c-account-create-update-tr7hx" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.180599 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f09e2f4-4f82-4388-9b5a-a9e890d3a950-operator-scripts\") pod \"placement-db-create-64xvf\" (UID: \"5f09e2f4-4f82-4388-9b5a-a9e890d3a950\") " pod="openstack/placement-db-create-64xvf" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.181312 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpvp9\" (UniqueName: \"kubernetes.io/projected/6c84aa49-2900-4a14-b81b-bb03e925d1b7-kube-api-access-wpvp9\") pod \"keystone-ab0c-account-create-update-tr7hx\" (UID: \"6c84aa49-2900-4a14-b81b-bb03e925d1b7\") " pod="openstack/keystone-ab0c-account-create-update-tr7hx" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.182203 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c84aa49-2900-4a14-b81b-bb03e925d1b7-operator-scripts\") pod \"keystone-ab0c-account-create-update-tr7hx\" (UID: \"6c84aa49-2900-4a14-b81b-bb03e925d1b7\") " pod="openstack/keystone-ab0c-account-create-update-tr7hx" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.198087 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpvp9\" (UniqueName: \"kubernetes.io/projected/6c84aa49-2900-4a14-b81b-bb03e925d1b7-kube-api-access-wpvp9\") pod \"keystone-ab0c-account-create-update-tr7hx\" (UID: \"6c84aa49-2900-4a14-b81b-bb03e925d1b7\") " pod="openstack/keystone-ab0c-account-create-update-tr7hx" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.225080 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ab0c-account-create-update-tr7hx" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.290316 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlcdq\" (UniqueName: \"kubernetes.io/projected/5f09e2f4-4f82-4388-9b5a-a9e890d3a950-kube-api-access-mlcdq\") pod \"placement-db-create-64xvf\" (UID: \"5f09e2f4-4f82-4388-9b5a-a9e890d3a950\") " pod="openstack/placement-db-create-64xvf" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.290706 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wds8g\" (UniqueName: \"kubernetes.io/projected/c4f6b362-7670-4867-b8fa-1f4c6170389f-kube-api-access-wds8g\") pod \"placement-a750-account-create-update-7wk26\" (UID: \"c4f6b362-7670-4867-b8fa-1f4c6170389f\") " pod="openstack/placement-a750-account-create-update-7wk26" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.291216 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f09e2f4-4f82-4388-9b5a-a9e890d3a950-operator-scripts\") pod \"placement-db-create-64xvf\" (UID: \"5f09e2f4-4f82-4388-9b5a-a9e890d3a950\") " pod="openstack/placement-db-create-64xvf" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.291290 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4f6b362-7670-4867-b8fa-1f4c6170389f-operator-scripts\") pod \"placement-a750-account-create-update-7wk26\" (UID: \"c4f6b362-7670-4867-b8fa-1f4c6170389f\") " pod="openstack/placement-a750-account-create-update-7wk26" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.293156 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f09e2f4-4f82-4388-9b5a-a9e890d3a950-operator-scripts\") pod \"placement-db-create-64xvf\" (UID: \"5f09e2f4-4f82-4388-9b5a-a9e890d3a950\") " pod="openstack/placement-db-create-64xvf" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.314000 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlcdq\" (UniqueName: \"kubernetes.io/projected/5f09e2f4-4f82-4388-9b5a-a9e890d3a950-kube-api-access-mlcdq\") pod \"placement-db-create-64xvf\" (UID: \"5f09e2f4-4f82-4388-9b5a-a9e890d3a950\") " pod="openstack/placement-db-create-64xvf" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.348531 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-64xvf" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.394142 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wds8g\" (UniqueName: \"kubernetes.io/projected/c4f6b362-7670-4867-b8fa-1f4c6170389f-kube-api-access-wds8g\") pod \"placement-a750-account-create-update-7wk26\" (UID: \"c4f6b362-7670-4867-b8fa-1f4c6170389f\") " pod="openstack/placement-a750-account-create-update-7wk26" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.394355 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4f6b362-7670-4867-b8fa-1f4c6170389f-operator-scripts\") pod \"placement-a750-account-create-update-7wk26\" (UID: \"c4f6b362-7670-4867-b8fa-1f4c6170389f\") " pod="openstack/placement-a750-account-create-update-7wk26" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.395234 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4f6b362-7670-4867-b8fa-1f4c6170389f-operator-scripts\") pod \"placement-a750-account-create-update-7wk26\" (UID: \"c4f6b362-7670-4867-b8fa-1f4c6170389f\") " pod="openstack/placement-a750-account-create-update-7wk26" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.416924 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wds8g\" (UniqueName: \"kubernetes.io/projected/c4f6b362-7670-4867-b8fa-1f4c6170389f-kube-api-access-wds8g\") pod \"placement-a750-account-create-update-7wk26\" (UID: \"c4f6b362-7670-4867-b8fa-1f4c6170389f\") " pod="openstack/placement-a750-account-create-update-7wk26" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.427082 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a750-account-create-update-7wk26" Mar 13 10:24:15 crc kubenswrapper[4632]: I0313 10:24:15.495010 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:15 crc kubenswrapper[4632]: E0313 10:24:15.495234 4632 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 10:24:15 crc kubenswrapper[4632]: E0313 10:24:15.495249 4632 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 10:24:15 crc kubenswrapper[4632]: E0313 10:24:15.495291 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift podName:e37b3d77-de2e-4be9-9984-550d4ba0f2f0 nodeName:}" failed. No retries permitted until 2026-03-13 10:24:23.495277311 +0000 UTC m=+1237.517807444 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift") pod "swift-storage-0" (UID: "e37b3d77-de2e-4be9-9984-550d4ba0f2f0") : configmap "swift-ring-files" not found Mar 13 10:24:16 crc kubenswrapper[4632]: W0313 10:24:16.703184 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2610abab_1da4_4912_9ca7_f2aa2d7c0486.slice/crio-93ffdbe4b1ae6dbe85c9cdaa72f589075fb53d7d372bcaee83be3fb573b289c4 WatchSource:0}: Error finding container 93ffdbe4b1ae6dbe85c9cdaa72f589075fb53d7d372bcaee83be3fb573b289c4: Status 404 returned error can't find the container with id 93ffdbe4b1ae6dbe85c9cdaa72f589075fb53d7d372bcaee83be3fb573b289c4 Mar 13 10:24:16 crc kubenswrapper[4632]: I0313 10:24:16.710594 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-9698-account-create-update-9kfhv"] Mar 13 10:24:16 crc kubenswrapper[4632]: I0313 10:24:16.720718 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:24:16 crc kubenswrapper[4632]: I0313 10:24:16.799281 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79dfb79747-jv5m6"] Mar 13 10:24:16 crc kubenswrapper[4632]: I0313 10:24:16.799555 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" podUID="a362448e-8daa-4bf4-958f-f3ca135be228" containerName="dnsmasq-dns" containerID="cri-o://6e1c032b958be8592683422ee06f119be07d42e5fc24c06ebfd10193412b1ccc" gracePeriod=10 Mar 13 10:24:16 crc kubenswrapper[4632]: I0313 10:24:16.889118 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-bfb6b"] Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.070068 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-c8lh5"] Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.091310 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-c8lh5"] Mar 13 10:24:17 crc kubenswrapper[4632]: W0313 10:24:17.170582 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4f6b362_7670_4867_b8fa_1f4c6170389f.slice/crio-8da340b9259c4e3950c1e2300189828fc046a0a3f7a7546b25adfdd801a1d232 WatchSource:0}: Error finding container 8da340b9259c4e3950c1e2300189828fc046a0a3f7a7546b25adfdd801a1d232: Status 404 returned error can't find the container with id 8da340b9259c4e3950c1e2300189828fc046a0a3f7a7546b25adfdd801a1d232 Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.178593 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a750-account-create-update-7wk26"] Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.194577 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-7hqpw"] Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.196467 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9698-account-create-update-9kfhv" event={"ID":"2610abab-1da4-4912-9ca7-f2aa2d7c0486","Type":"ContainerStarted","Data":"c0ed44d952b9a10d8f17f6b274d11ae8079f72b678bca2ec969eb44a14c0f18e"} Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.196505 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9698-account-create-update-9kfhv" event={"ID":"2610abab-1da4-4912-9ca7-f2aa2d7c0486","Type":"ContainerStarted","Data":"93ffdbe4b1ae6dbe85c9cdaa72f589075fb53d7d372bcaee83be3fb573b289c4"} Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.206278 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ab0c-account-create-update-tr7hx"] Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.207218 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bfb6b" event={"ID":"8e353045-e09b-4cd2-b659-1954485ec8db","Type":"ContainerStarted","Data":"e12bb579655132c65f7afaf171587507463b77c9b73d0902f8981397a2c342cd"} Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.207343 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bfb6b" event={"ID":"8e353045-e09b-4cd2-b659-1954485ec8db","Type":"ContainerStarted","Data":"535f1a0d62083c8f1779cf07cdf7b6b338543f49b074ea77690edaf35b0bd836"} Mar 13 10:24:17 crc kubenswrapper[4632]: W0313 10:24:17.208400 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod584d2818_4b22_468f_b296_bd1850c7915b.slice/crio-4a12c0a38209ede38a672765a10b8d7b9b038d2da32f90548d5d8030e926a912 WatchSource:0}: Error finding container 4a12c0a38209ede38a672765a10b8d7b9b038d2da32f90548d5d8030e926a912: Status 404 returned error can't find the container with id 4a12c0a38209ede38a672765a10b8d7b9b038d2da32f90548d5d8030e926a912 Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.218261 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-64xvf"] Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.224442 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-9698-account-create-update-9kfhv" podStartSLOduration=3.224418293 podStartE2EDuration="3.224418293s" podCreationTimestamp="2026-03-13 10:24:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:17.214923801 +0000 UTC m=+1231.237453934" watchObservedRunningTime="2026-03-13 10:24:17.224418293 +0000 UTC m=+1231.246948416" Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.231261 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mkdcg" event={"ID":"bc39c52e-008f-40c1-b93b-532707127fcd","Type":"ContainerStarted","Data":"c1060b701ea818cb4c5d1e5e94618270eed048e8bd16d50775b43c9b34c6b1b9"} Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.248046 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-bfb6b" podStartSLOduration=3.247987741 podStartE2EDuration="3.247987741s" podCreationTimestamp="2026-03-13 10:24:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:17.23324607 +0000 UTC m=+1231.255776203" watchObservedRunningTime="2026-03-13 10:24:17.247987741 +0000 UTC m=+1231.270517874" Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.265442 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-mkdcg" podStartSLOduration=2.2120488050000002 podStartE2EDuration="6.2654154s" podCreationTimestamp="2026-03-13 10:24:11 +0000 UTC" firstStartedPulling="2026-03-13 10:24:12.227328045 +0000 UTC m=+1226.249858178" lastFinishedPulling="2026-03-13 10:24:16.28069464 +0000 UTC m=+1230.303224773" observedRunningTime="2026-03-13 10:24:17.260723324 +0000 UTC m=+1231.283253477" watchObservedRunningTime="2026-03-13 10:24:17.2654154 +0000 UTC m=+1231.287945533" Mar 13 10:24:17 crc kubenswrapper[4632]: W0313 10:24:17.266027 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c84aa49_2900_4a14_b81b_bb03e925d1b7.slice/crio-c64cf5189ab7a5f49e61bd270ad953bb57c962bfc193a6a4df7169ba439b5084 WatchSource:0}: Error finding container c64cf5189ab7a5f49e61bd270ad953bb57c962bfc193a6a4df7169ba439b5084: Status 404 returned error can't find the container with id c64cf5189ab7a5f49e61bd270ad953bb57c962bfc193a6a4df7169ba439b5084 Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.266190 4632 generic.go:334] "Generic (PLEG): container finished" podID="a362448e-8daa-4bf4-958f-f3ca135be228" containerID="6e1c032b958be8592683422ee06f119be07d42e5fc24c06ebfd10193412b1ccc" exitCode=0 Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.266266 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" event={"ID":"a362448e-8daa-4bf4-958f-f3ca135be228","Type":"ContainerDied","Data":"6e1c032b958be8592683422ee06f119be07d42e5fc24c06ebfd10193412b1ccc"} Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.714243 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.841737 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-ovsdbserver-sb\") pod \"a362448e-8daa-4bf4-958f-f3ca135be228\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.841990 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pv67\" (UniqueName: \"kubernetes.io/projected/a362448e-8daa-4bf4-958f-f3ca135be228-kube-api-access-5pv67\") pod \"a362448e-8daa-4bf4-958f-f3ca135be228\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.842183 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-dns-svc\") pod \"a362448e-8daa-4bf4-958f-f3ca135be228\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.842276 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-config\") pod \"a362448e-8daa-4bf4-958f-f3ca135be228\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.842411 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-ovsdbserver-nb\") pod \"a362448e-8daa-4bf4-958f-f3ca135be228\" (UID: \"a362448e-8daa-4bf4-958f-f3ca135be228\") " Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.852307 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a362448e-8daa-4bf4-958f-f3ca135be228-kube-api-access-5pv67" (OuterVolumeSpecName: "kube-api-access-5pv67") pod "a362448e-8daa-4bf4-958f-f3ca135be228" (UID: "a362448e-8daa-4bf4-958f-f3ca135be228"). InnerVolumeSpecName "kube-api-access-5pv67". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:17 crc kubenswrapper[4632]: I0313 10:24:17.946035 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pv67\" (UniqueName: \"kubernetes.io/projected/a362448e-8daa-4bf4-958f-f3ca135be228-kube-api-access-5pv67\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.045802 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a362448e-8daa-4bf4-958f-f3ca135be228" (UID: "a362448e-8daa-4bf4-958f-f3ca135be228"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.047076 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.072805 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc41c555-17e5-4785-a003-3f8e9f10d799" path="/var/lib/kubelet/pods/cc41c555-17e5-4785-a003-3f8e9f10d799/volumes" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.111237 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a362448e-8daa-4bf4-958f-f3ca135be228" (UID: "a362448e-8daa-4bf4-958f-f3ca135be228"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.149805 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.150882 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-config" (OuterVolumeSpecName: "config") pod "a362448e-8daa-4bf4-958f-f3ca135be228" (UID: "a362448e-8daa-4bf4-958f-f3ca135be228"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.153703 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a362448e-8daa-4bf4-958f-f3ca135be228" (UID: "a362448e-8daa-4bf4-958f-f3ca135be228"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.251460 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.251842 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a362448e-8daa-4bf4-958f-f3ca135be228-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.292146 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.293794 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-79dfb79747-jv5m6" event={"ID":"a362448e-8daa-4bf4-958f-f3ca135be228","Type":"ContainerDied","Data":"353c97e7f46060d145aa3be9824f787e63d6eed5891607427c9311023caa0833"} Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.294031 4632 scope.go:117] "RemoveContainer" containerID="6e1c032b958be8592683422ee06f119be07d42e5fc24c06ebfd10193412b1ccc" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.298656 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a750-account-create-update-7wk26" event={"ID":"c4f6b362-7670-4867-b8fa-1f4c6170389f","Type":"ContainerStarted","Data":"10bcedf0effae05b832e3793407fcf2703d9df4f7136a8211c78de6b0a99c17b"} Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.298862 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a750-account-create-update-7wk26" event={"ID":"c4f6b362-7670-4867-b8fa-1f4c6170389f","Type":"ContainerStarted","Data":"8da340b9259c4e3950c1e2300189828fc046a0a3f7a7546b25adfdd801a1d232"} Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.304581 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ab0c-account-create-update-tr7hx" event={"ID":"6c84aa49-2900-4a14-b81b-bb03e925d1b7","Type":"ContainerStarted","Data":"40127d251d4cb7407ae0ce8a1705cd5210171fb2a750df3289fa3b2b9a54b055"} Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.304866 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ab0c-account-create-update-tr7hx" event={"ID":"6c84aa49-2900-4a14-b81b-bb03e925d1b7","Type":"ContainerStarted","Data":"c64cf5189ab7a5f49e61bd270ad953bb57c962bfc193a6a4df7169ba439b5084"} Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.318179 4632 generic.go:334] "Generic (PLEG): container finished" podID="2610abab-1da4-4912-9ca7-f2aa2d7c0486" containerID="c0ed44d952b9a10d8f17f6b274d11ae8079f72b678bca2ec969eb44a14c0f18e" exitCode=0 Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.318245 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9698-account-create-update-9kfhv" event={"ID":"2610abab-1da4-4912-9ca7-f2aa2d7c0486","Type":"ContainerDied","Data":"c0ed44d952b9a10d8f17f6b274d11ae8079f72b678bca2ec969eb44a14c0f18e"} Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.332500 4632 generic.go:334] "Generic (PLEG): container finished" podID="8e353045-e09b-4cd2-b659-1954485ec8db" containerID="e12bb579655132c65f7afaf171587507463b77c9b73d0902f8981397a2c342cd" exitCode=0 Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.332602 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bfb6b" event={"ID":"8e353045-e09b-4cd2-b659-1954485ec8db","Type":"ContainerDied","Data":"e12bb579655132c65f7afaf171587507463b77c9b73d0902f8981397a2c342cd"} Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.336502 4632 generic.go:334] "Generic (PLEG): container finished" podID="5f09e2f4-4f82-4388-9b5a-a9e890d3a950" containerID="05358506b7b8a5602da80aa6b4985f897c7b0818d4a2f70ed84421563493ee78" exitCode=0 Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.336567 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-64xvf" event={"ID":"5f09e2f4-4f82-4388-9b5a-a9e890d3a950","Type":"ContainerDied","Data":"05358506b7b8a5602da80aa6b4985f897c7b0818d4a2f70ed84421563493ee78"} Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.336594 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-64xvf" event={"ID":"5f09e2f4-4f82-4388-9b5a-a9e890d3a950","Type":"ContainerStarted","Data":"fa3bd3847d87a827b95ad3e0b17ca932be91f863713ca115297ee8fc7b29e228"} Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.342066 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-a750-account-create-update-7wk26" podStartSLOduration=3.342040163 podStartE2EDuration="3.342040163s" podCreationTimestamp="2026-03-13 10:24:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:18.326183234 +0000 UTC m=+1232.348713387" watchObservedRunningTime="2026-03-13 10:24:18.342040163 +0000 UTC m=+1232.364570316" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.346314 4632 generic.go:334] "Generic (PLEG): container finished" podID="584d2818-4b22-468f-b296-bd1850c7915b" containerID="2fd6ae14a44d07bfe626dada3603473befbf9326ca83648414737abd80e0ce5e" exitCode=0 Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.346470 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-7hqpw" event={"ID":"584d2818-4b22-468f-b296-bd1850c7915b","Type":"ContainerDied","Data":"2fd6ae14a44d07bfe626dada3603473befbf9326ca83648414737abd80e0ce5e"} Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.346517 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-7hqpw" event={"ID":"584d2818-4b22-468f-b296-bd1850c7915b","Type":"ContainerStarted","Data":"4a12c0a38209ede38a672765a10b8d7b9b038d2da32f90548d5d8030e926a912"} Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.369108 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-ab0c-account-create-update-tr7hx" podStartSLOduration=4.369091057 podStartE2EDuration="4.369091057s" podCreationTimestamp="2026-03-13 10:24:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:18.346268037 +0000 UTC m=+1232.368798170" watchObservedRunningTime="2026-03-13 10:24:18.369091057 +0000 UTC m=+1232.391621190" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.375345 4632 scope.go:117] "RemoveContainer" containerID="c9a88952f81b62d419132fe9a18256ffde7daf30602bf205173e43b0963b20c3" Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.432654 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-79dfb79747-jv5m6"] Mar 13 10:24:18 crc kubenswrapper[4632]: I0313 10:24:18.441916 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-79dfb79747-jv5m6"] Mar 13 10:24:19 crc kubenswrapper[4632]: I0313 10:24:19.357243 4632 generic.go:334] "Generic (PLEG): container finished" podID="6c84aa49-2900-4a14-b81b-bb03e925d1b7" containerID="40127d251d4cb7407ae0ce8a1705cd5210171fb2a750df3289fa3b2b9a54b055" exitCode=0 Mar 13 10:24:19 crc kubenswrapper[4632]: I0313 10:24:19.357312 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ab0c-account-create-update-tr7hx" event={"ID":"6c84aa49-2900-4a14-b81b-bb03e925d1b7","Type":"ContainerDied","Data":"40127d251d4cb7407ae0ce8a1705cd5210171fb2a750df3289fa3b2b9a54b055"} Mar 13 10:24:19 crc kubenswrapper[4632]: I0313 10:24:19.360539 4632 generic.go:334] "Generic (PLEG): container finished" podID="c4f6b362-7670-4867-b8fa-1f4c6170389f" containerID="10bcedf0effae05b832e3793407fcf2703d9df4f7136a8211c78de6b0a99c17b" exitCode=0 Mar 13 10:24:19 crc kubenswrapper[4632]: I0313 10:24:19.360766 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a750-account-create-update-7wk26" event={"ID":"c4f6b362-7670-4867-b8fa-1f4c6170389f","Type":"ContainerDied","Data":"10bcedf0effae05b832e3793407fcf2703d9df4f7136a8211c78de6b0a99c17b"} Mar 13 10:24:19 crc kubenswrapper[4632]: I0313 10:24:19.888199 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bfb6b" Mar 13 10:24:19 crc kubenswrapper[4632]: I0313 10:24:19.980480 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjh6b\" (UniqueName: \"kubernetes.io/projected/8e353045-e09b-4cd2-b659-1954485ec8db-kube-api-access-fjh6b\") pod \"8e353045-e09b-4cd2-b659-1954485ec8db\" (UID: \"8e353045-e09b-4cd2-b659-1954485ec8db\") " Mar 13 10:24:19 crc kubenswrapper[4632]: I0313 10:24:19.980563 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e353045-e09b-4cd2-b659-1954485ec8db-operator-scripts\") pod \"8e353045-e09b-4cd2-b659-1954485ec8db\" (UID: \"8e353045-e09b-4cd2-b659-1954485ec8db\") " Mar 13 10:24:19 crc kubenswrapper[4632]: I0313 10:24:19.981074 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e353045-e09b-4cd2-b659-1954485ec8db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8e353045-e09b-4cd2-b659-1954485ec8db" (UID: "8e353045-e09b-4cd2-b659-1954485ec8db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:19 crc kubenswrapper[4632]: I0313 10:24:19.986506 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e353045-e09b-4cd2-b659-1954485ec8db-kube-api-access-fjh6b" (OuterVolumeSpecName: "kube-api-access-fjh6b") pod "8e353045-e09b-4cd2-b659-1954485ec8db" (UID: "8e353045-e09b-4cd2-b659-1954485ec8db"). InnerVolumeSpecName "kube-api-access-fjh6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.034474 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7hqpw" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.050998 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-64xvf" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.058190 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9698-account-create-update-9kfhv" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.070486 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a362448e-8daa-4bf4-958f-f3ca135be228" path="/var/lib/kubelet/pods/a362448e-8daa-4bf4-958f-f3ca135be228/volumes" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.082744 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjh6b\" (UniqueName: \"kubernetes.io/projected/8e353045-e09b-4cd2-b659-1954485ec8db-kube-api-access-fjh6b\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.082765 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e353045-e09b-4cd2-b659-1954485ec8db-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.183905 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2610abab-1da4-4912-9ca7-f2aa2d7c0486-operator-scripts\") pod \"2610abab-1da4-4912-9ca7-f2aa2d7c0486\" (UID: \"2610abab-1da4-4912-9ca7-f2aa2d7c0486\") " Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.184005 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlcdq\" (UniqueName: \"kubernetes.io/projected/5f09e2f4-4f82-4388-9b5a-a9e890d3a950-kube-api-access-mlcdq\") pod \"5f09e2f4-4f82-4388-9b5a-a9e890d3a950\" (UID: \"5f09e2f4-4f82-4388-9b5a-a9e890d3a950\") " Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.184109 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wszzh\" (UniqueName: \"kubernetes.io/projected/584d2818-4b22-468f-b296-bd1850c7915b-kube-api-access-wszzh\") pod \"584d2818-4b22-468f-b296-bd1850c7915b\" (UID: \"584d2818-4b22-468f-b296-bd1850c7915b\") " Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.184149 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29xj8\" (UniqueName: \"kubernetes.io/projected/2610abab-1da4-4912-9ca7-f2aa2d7c0486-kube-api-access-29xj8\") pod \"2610abab-1da4-4912-9ca7-f2aa2d7c0486\" (UID: \"2610abab-1da4-4912-9ca7-f2aa2d7c0486\") " Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.184219 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/584d2818-4b22-468f-b296-bd1850c7915b-operator-scripts\") pod \"584d2818-4b22-468f-b296-bd1850c7915b\" (UID: \"584d2818-4b22-468f-b296-bd1850c7915b\") " Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.184255 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f09e2f4-4f82-4388-9b5a-a9e890d3a950-operator-scripts\") pod \"5f09e2f4-4f82-4388-9b5a-a9e890d3a950\" (UID: \"5f09e2f4-4f82-4388-9b5a-a9e890d3a950\") " Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.184979 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2610abab-1da4-4912-9ca7-f2aa2d7c0486-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2610abab-1da4-4912-9ca7-f2aa2d7c0486" (UID: "2610abab-1da4-4912-9ca7-f2aa2d7c0486"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.185098 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/584d2818-4b22-468f-b296-bd1850c7915b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "584d2818-4b22-468f-b296-bd1850c7915b" (UID: "584d2818-4b22-468f-b296-bd1850c7915b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.185236 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5f09e2f4-4f82-4388-9b5a-a9e890d3a950-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5f09e2f4-4f82-4388-9b5a-a9e890d3a950" (UID: "5f09e2f4-4f82-4388-9b5a-a9e890d3a950"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.187015 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/584d2818-4b22-468f-b296-bd1850c7915b-kube-api-access-wszzh" (OuterVolumeSpecName: "kube-api-access-wszzh") pod "584d2818-4b22-468f-b296-bd1850c7915b" (UID: "584d2818-4b22-468f-b296-bd1850c7915b"). InnerVolumeSpecName "kube-api-access-wszzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.187142 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f09e2f4-4f82-4388-9b5a-a9e890d3a950-kube-api-access-mlcdq" (OuterVolumeSpecName: "kube-api-access-mlcdq") pod "5f09e2f4-4f82-4388-9b5a-a9e890d3a950" (UID: "5f09e2f4-4f82-4388-9b5a-a9e890d3a950"). InnerVolumeSpecName "kube-api-access-mlcdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.187352 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2610abab-1da4-4912-9ca7-f2aa2d7c0486-kube-api-access-29xj8" (OuterVolumeSpecName: "kube-api-access-29xj8") pod "2610abab-1da4-4912-9ca7-f2aa2d7c0486" (UID: "2610abab-1da4-4912-9ca7-f2aa2d7c0486"). InnerVolumeSpecName "kube-api-access-29xj8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.285829 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wszzh\" (UniqueName: \"kubernetes.io/projected/584d2818-4b22-468f-b296-bd1850c7915b-kube-api-access-wszzh\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.285883 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29xj8\" (UniqueName: \"kubernetes.io/projected/2610abab-1da4-4912-9ca7-f2aa2d7c0486-kube-api-access-29xj8\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.285912 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/584d2818-4b22-468f-b296-bd1850c7915b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.285930 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5f09e2f4-4f82-4388-9b5a-a9e890d3a950-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.285976 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2610abab-1da4-4912-9ca7-f2aa2d7c0486-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.285994 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlcdq\" (UniqueName: \"kubernetes.io/projected/5f09e2f4-4f82-4388-9b5a-a9e890d3a950-kube-api-access-mlcdq\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.368878 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-9698-account-create-update-9kfhv" event={"ID":"2610abab-1da4-4912-9ca7-f2aa2d7c0486","Type":"ContainerDied","Data":"93ffdbe4b1ae6dbe85c9cdaa72f589075fb53d7d372bcaee83be3fb573b289c4"} Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.368919 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93ffdbe4b1ae6dbe85c9cdaa72f589075fb53d7d372bcaee83be3fb573b289c4" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.368957 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-9698-account-create-update-9kfhv" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.370588 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-bfb6b" event={"ID":"8e353045-e09b-4cd2-b659-1954485ec8db","Type":"ContainerDied","Data":"535f1a0d62083c8f1779cf07cdf7b6b338543f49b074ea77690edaf35b0bd836"} Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.370628 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="535f1a0d62083c8f1779cf07cdf7b6b338543f49b074ea77690edaf35b0bd836" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.370681 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-bfb6b" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.372381 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-64xvf" event={"ID":"5f09e2f4-4f82-4388-9b5a-a9e890d3a950","Type":"ContainerDied","Data":"fa3bd3847d87a827b95ad3e0b17ca932be91f863713ca115297ee8fc7b29e228"} Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.372408 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa3bd3847d87a827b95ad3e0b17ca932be91f863713ca115297ee8fc7b29e228" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.372390 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-64xvf" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.376386 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-7hqpw" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.376562 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-7hqpw" event={"ID":"584d2818-4b22-468f-b296-bd1850c7915b","Type":"ContainerDied","Data":"4a12c0a38209ede38a672765a10b8d7b9b038d2da32f90548d5d8030e926a912"} Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.376599 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a12c0a38209ede38a672765a10b8d7b9b038d2da32f90548d5d8030e926a912" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.725969 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ab0c-account-create-update-tr7hx" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.771270 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a750-account-create-update-7wk26" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.796758 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c84aa49-2900-4a14-b81b-bb03e925d1b7-operator-scripts\") pod \"6c84aa49-2900-4a14-b81b-bb03e925d1b7\" (UID: \"6c84aa49-2900-4a14-b81b-bb03e925d1b7\") " Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.796908 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpvp9\" (UniqueName: \"kubernetes.io/projected/6c84aa49-2900-4a14-b81b-bb03e925d1b7-kube-api-access-wpvp9\") pod \"6c84aa49-2900-4a14-b81b-bb03e925d1b7\" (UID: \"6c84aa49-2900-4a14-b81b-bb03e925d1b7\") " Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.798318 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c84aa49-2900-4a14-b81b-bb03e925d1b7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6c84aa49-2900-4a14-b81b-bb03e925d1b7" (UID: "6c84aa49-2900-4a14-b81b-bb03e925d1b7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.801588 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c84aa49-2900-4a14-b81b-bb03e925d1b7-kube-api-access-wpvp9" (OuterVolumeSpecName: "kube-api-access-wpvp9") pod "6c84aa49-2900-4a14-b81b-bb03e925d1b7" (UID: "6c84aa49-2900-4a14-b81b-bb03e925d1b7"). InnerVolumeSpecName "kube-api-access-wpvp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.900581 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wds8g\" (UniqueName: \"kubernetes.io/projected/c4f6b362-7670-4867-b8fa-1f4c6170389f-kube-api-access-wds8g\") pod \"c4f6b362-7670-4867-b8fa-1f4c6170389f\" (UID: \"c4f6b362-7670-4867-b8fa-1f4c6170389f\") " Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.903334 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4f6b362-7670-4867-b8fa-1f4c6170389f-operator-scripts\") pod \"c4f6b362-7670-4867-b8fa-1f4c6170389f\" (UID: \"c4f6b362-7670-4867-b8fa-1f4c6170389f\") " Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.903772 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4f6b362-7670-4867-b8fa-1f4c6170389f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c4f6b362-7670-4867-b8fa-1f4c6170389f" (UID: "c4f6b362-7670-4867-b8fa-1f4c6170389f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.904161 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpvp9\" (UniqueName: \"kubernetes.io/projected/6c84aa49-2900-4a14-b81b-bb03e925d1b7-kube-api-access-wpvp9\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.904191 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c4f6b362-7670-4867-b8fa-1f4c6170389f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.904204 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c84aa49-2900-4a14-b81b-bb03e925d1b7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:20 crc kubenswrapper[4632]: I0313 10:24:20.905761 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4f6b362-7670-4867-b8fa-1f4c6170389f-kube-api-access-wds8g" (OuterVolumeSpecName: "kube-api-access-wds8g") pod "c4f6b362-7670-4867-b8fa-1f4c6170389f" (UID: "c4f6b362-7670-4867-b8fa-1f4c6170389f"). InnerVolumeSpecName "kube-api-access-wds8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:21 crc kubenswrapper[4632]: I0313 10:24:21.005920 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wds8g\" (UniqueName: \"kubernetes.io/projected/c4f6b362-7670-4867-b8fa-1f4c6170389f-kube-api-access-wds8g\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:21 crc kubenswrapper[4632]: I0313 10:24:21.386169 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ab0c-account-create-update-tr7hx" event={"ID":"6c84aa49-2900-4a14-b81b-bb03e925d1b7","Type":"ContainerDied","Data":"c64cf5189ab7a5f49e61bd270ad953bb57c962bfc193a6a4df7169ba439b5084"} Mar 13 10:24:21 crc kubenswrapper[4632]: I0313 10:24:21.386828 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c64cf5189ab7a5f49e61bd270ad953bb57c962bfc193a6a4df7169ba439b5084" Mar 13 10:24:21 crc kubenswrapper[4632]: I0313 10:24:21.386331 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ab0c-account-create-update-tr7hx" Mar 13 10:24:21 crc kubenswrapper[4632]: I0313 10:24:21.387897 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a750-account-create-update-7wk26" event={"ID":"c4f6b362-7670-4867-b8fa-1f4c6170389f","Type":"ContainerDied","Data":"8da340b9259c4e3950c1e2300189828fc046a0a3f7a7546b25adfdd801a1d232"} Mar 13 10:24:21 crc kubenswrapper[4632]: I0313 10:24:21.387985 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8da340b9259c4e3950c1e2300189828fc046a0a3f7a7546b25adfdd801a1d232" Mar 13 10:24:21 crc kubenswrapper[4632]: I0313 10:24:21.388039 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a750-account-create-update-7wk26" Mar 13 10:24:21 crc kubenswrapper[4632]: I0313 10:24:21.713072 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.067984 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-lrjmj"] Mar 13 10:24:22 crc kubenswrapper[4632]: E0313 10:24:22.068391 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e353045-e09b-4cd2-b659-1954485ec8db" containerName="mariadb-database-create" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068411 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e353045-e09b-4cd2-b659-1954485ec8db" containerName="mariadb-database-create" Mar 13 10:24:22 crc kubenswrapper[4632]: E0313 10:24:22.068436 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c84aa49-2900-4a14-b81b-bb03e925d1b7" containerName="mariadb-account-create-update" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068444 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c84aa49-2900-4a14-b81b-bb03e925d1b7" containerName="mariadb-account-create-update" Mar 13 10:24:22 crc kubenswrapper[4632]: E0313 10:24:22.068476 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f09e2f4-4f82-4388-9b5a-a9e890d3a950" containerName="mariadb-database-create" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068485 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f09e2f4-4f82-4388-9b5a-a9e890d3a950" containerName="mariadb-database-create" Mar 13 10:24:22 crc kubenswrapper[4632]: E0313 10:24:22.068498 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="584d2818-4b22-468f-b296-bd1850c7915b" containerName="mariadb-database-create" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068508 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="584d2818-4b22-468f-b296-bd1850c7915b" containerName="mariadb-database-create" Mar 13 10:24:22 crc kubenswrapper[4632]: E0313 10:24:22.068525 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a362448e-8daa-4bf4-958f-f3ca135be228" containerName="dnsmasq-dns" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068535 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a362448e-8daa-4bf4-958f-f3ca135be228" containerName="dnsmasq-dns" Mar 13 10:24:22 crc kubenswrapper[4632]: E0313 10:24:22.068549 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a362448e-8daa-4bf4-958f-f3ca135be228" containerName="init" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068559 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a362448e-8daa-4bf4-958f-f3ca135be228" containerName="init" Mar 13 10:24:22 crc kubenswrapper[4632]: E0313 10:24:22.068569 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4f6b362-7670-4867-b8fa-1f4c6170389f" containerName="mariadb-account-create-update" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068578 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4f6b362-7670-4867-b8fa-1f4c6170389f" containerName="mariadb-account-create-update" Mar 13 10:24:22 crc kubenswrapper[4632]: E0313 10:24:22.068597 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2610abab-1da4-4912-9ca7-f2aa2d7c0486" containerName="mariadb-account-create-update" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068605 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2610abab-1da4-4912-9ca7-f2aa2d7c0486" containerName="mariadb-account-create-update" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068811 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e353045-e09b-4cd2-b659-1954485ec8db" containerName="mariadb-database-create" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068830 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c84aa49-2900-4a14-b81b-bb03e925d1b7" containerName="mariadb-account-create-update" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068841 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="a362448e-8daa-4bf4-958f-f3ca135be228" containerName="dnsmasq-dns" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068849 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4f6b362-7670-4867-b8fa-1f4c6170389f" containerName="mariadb-account-create-update" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068861 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="584d2818-4b22-468f-b296-bd1850c7915b" containerName="mariadb-database-create" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068872 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="2610abab-1da4-4912-9ca7-f2aa2d7c0486" containerName="mariadb-account-create-update" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.068884 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f09e2f4-4f82-4388-9b5a-a9e890d3a950" containerName="mariadb-database-create" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.069558 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-lrjmj" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.074987 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.081255 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-lrjmj"] Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.225217 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d670715-74f3-46a6-974c-b6953af9fdb7-operator-scripts\") pod \"root-account-create-update-lrjmj\" (UID: \"2d670715-74f3-46a6-974c-b6953af9fdb7\") " pod="openstack/root-account-create-update-lrjmj" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.225301 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qhtfx\" (UniqueName: \"kubernetes.io/projected/2d670715-74f3-46a6-974c-b6953af9fdb7-kube-api-access-qhtfx\") pod \"root-account-create-update-lrjmj\" (UID: \"2d670715-74f3-46a6-974c-b6953af9fdb7\") " pod="openstack/root-account-create-update-lrjmj" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.327456 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d670715-74f3-46a6-974c-b6953af9fdb7-operator-scripts\") pod \"root-account-create-update-lrjmj\" (UID: \"2d670715-74f3-46a6-974c-b6953af9fdb7\") " pod="openstack/root-account-create-update-lrjmj" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.328258 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qhtfx\" (UniqueName: \"kubernetes.io/projected/2d670715-74f3-46a6-974c-b6953af9fdb7-kube-api-access-qhtfx\") pod \"root-account-create-update-lrjmj\" (UID: \"2d670715-74f3-46a6-974c-b6953af9fdb7\") " pod="openstack/root-account-create-update-lrjmj" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.328404 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d670715-74f3-46a6-974c-b6953af9fdb7-operator-scripts\") pod \"root-account-create-update-lrjmj\" (UID: \"2d670715-74f3-46a6-974c-b6953af9fdb7\") " pod="openstack/root-account-create-update-lrjmj" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.353165 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qhtfx\" (UniqueName: \"kubernetes.io/projected/2d670715-74f3-46a6-974c-b6953af9fdb7-kube-api-access-qhtfx\") pod \"root-account-create-update-lrjmj\" (UID: \"2d670715-74f3-46a6-974c-b6953af9fdb7\") " pod="openstack/root-account-create-update-lrjmj" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.392035 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-lrjmj" Mar 13 10:24:22 crc kubenswrapper[4632]: I0313 10:24:22.902378 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-lrjmj"] Mar 13 10:24:22 crc kubenswrapper[4632]: W0313 10:24:22.912282 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d670715_74f3_46a6_974c_b6953af9fdb7.slice/crio-4e87d116be82121d7fc1873346d45f93e13e538ff7fb3e09adc605646d99d9ef WatchSource:0}: Error finding container 4e87d116be82121d7fc1873346d45f93e13e538ff7fb3e09adc605646d99d9ef: Status 404 returned error can't find the container with id 4e87d116be82121d7fc1873346d45f93e13e538ff7fb3e09adc605646d99d9ef Mar 13 10:24:23 crc kubenswrapper[4632]: I0313 10:24:23.406745 4632 generic.go:334] "Generic (PLEG): container finished" podID="2d670715-74f3-46a6-974c-b6953af9fdb7" containerID="9bfb87771985986bb5edbb713355c76b663fe8b23df1170e73c42c65479f44df" exitCode=0 Mar 13 10:24:23 crc kubenswrapper[4632]: I0313 10:24:23.407048 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-lrjmj" event={"ID":"2d670715-74f3-46a6-974c-b6953af9fdb7","Type":"ContainerDied","Data":"9bfb87771985986bb5edbb713355c76b663fe8b23df1170e73c42c65479f44df"} Mar 13 10:24:23 crc kubenswrapper[4632]: I0313 10:24:23.407074 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-lrjmj" event={"ID":"2d670715-74f3-46a6-974c-b6953af9fdb7","Type":"ContainerStarted","Data":"4e87d116be82121d7fc1873346d45f93e13e538ff7fb3e09adc605646d99d9ef"} Mar 13 10:24:23 crc kubenswrapper[4632]: I0313 10:24:23.553014 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:23 crc kubenswrapper[4632]: E0313 10:24:23.553226 4632 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 10:24:23 crc kubenswrapper[4632]: E0313 10:24:23.554449 4632 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 10:24:23 crc kubenswrapper[4632]: E0313 10:24:23.554511 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift podName:e37b3d77-de2e-4be9-9984-550d4ba0f2f0 nodeName:}" failed. No retries permitted until 2026-03-13 10:24:39.554493624 +0000 UTC m=+1253.577023757 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift") pod "swift-storage-0" (UID: "e37b3d77-de2e-4be9-9984-550d4ba0f2f0") : configmap "swift-ring-files" not found Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.280853 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-l6hpb"] Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.281876 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l6hpb" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.285829 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.292379 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-l6hpb"] Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.295364 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qpd5p" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.367200 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjbxx\" (UniqueName: \"kubernetes.io/projected/4f1c5663-463b-45e2-b200-64e73e6d5698-kube-api-access-fjbxx\") pod \"glance-db-sync-l6hpb\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " pod="openstack/glance-db-sync-l6hpb" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.367253 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-config-data\") pod \"glance-db-sync-l6hpb\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " pod="openstack/glance-db-sync-l6hpb" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.367326 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-db-sync-config-data\") pod \"glance-db-sync-l6hpb\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " pod="openstack/glance-db-sync-l6hpb" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.367345 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-combined-ca-bundle\") pod \"glance-db-sync-l6hpb\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " pod="openstack/glance-db-sync-l6hpb" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.469295 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-db-sync-config-data\") pod \"glance-db-sync-l6hpb\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " pod="openstack/glance-db-sync-l6hpb" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.469339 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-combined-ca-bundle\") pod \"glance-db-sync-l6hpb\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " pod="openstack/glance-db-sync-l6hpb" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.469434 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjbxx\" (UniqueName: \"kubernetes.io/projected/4f1c5663-463b-45e2-b200-64e73e6d5698-kube-api-access-fjbxx\") pod \"glance-db-sync-l6hpb\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " pod="openstack/glance-db-sync-l6hpb" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.469460 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-config-data\") pod \"glance-db-sync-l6hpb\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " pod="openstack/glance-db-sync-l6hpb" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.478658 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-db-sync-config-data\") pod \"glance-db-sync-l6hpb\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " pod="openstack/glance-db-sync-l6hpb" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.487771 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-config-data\") pod \"glance-db-sync-l6hpb\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " pod="openstack/glance-db-sync-l6hpb" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.490387 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-combined-ca-bundle\") pod \"glance-db-sync-l6hpb\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " pod="openstack/glance-db-sync-l6hpb" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.493619 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjbxx\" (UniqueName: \"kubernetes.io/projected/4f1c5663-463b-45e2-b200-64e73e6d5698-kube-api-access-fjbxx\") pod \"glance-db-sync-l6hpb\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " pod="openstack/glance-db-sync-l6hpb" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.597363 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l6hpb" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.825145 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-lrjmj" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.977508 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qhtfx\" (UniqueName: \"kubernetes.io/projected/2d670715-74f3-46a6-974c-b6953af9fdb7-kube-api-access-qhtfx\") pod \"2d670715-74f3-46a6-974c-b6953af9fdb7\" (UID: \"2d670715-74f3-46a6-974c-b6953af9fdb7\") " Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.977714 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d670715-74f3-46a6-974c-b6953af9fdb7-operator-scripts\") pod \"2d670715-74f3-46a6-974c-b6953af9fdb7\" (UID: \"2d670715-74f3-46a6-974c-b6953af9fdb7\") " Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.979469 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d670715-74f3-46a6-974c-b6953af9fdb7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2d670715-74f3-46a6-974c-b6953af9fdb7" (UID: "2d670715-74f3-46a6-974c-b6953af9fdb7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:24 crc kubenswrapper[4632]: I0313 10:24:24.986742 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d670715-74f3-46a6-974c-b6953af9fdb7-kube-api-access-qhtfx" (OuterVolumeSpecName: "kube-api-access-qhtfx") pod "2d670715-74f3-46a6-974c-b6953af9fdb7" (UID: "2d670715-74f3-46a6-974c-b6953af9fdb7"). InnerVolumeSpecName "kube-api-access-qhtfx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:25 crc kubenswrapper[4632]: I0313 10:24:25.079884 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qhtfx\" (UniqueName: \"kubernetes.io/projected/2d670715-74f3-46a6-974c-b6953af9fdb7-kube-api-access-qhtfx\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:25 crc kubenswrapper[4632]: I0313 10:24:25.080233 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2d670715-74f3-46a6-974c-b6953af9fdb7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:25 crc kubenswrapper[4632]: I0313 10:24:25.266886 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-l6hpb"] Mar 13 10:24:25 crc kubenswrapper[4632]: W0313 10:24:25.285401 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f1c5663_463b_45e2_b200_64e73e6d5698.slice/crio-0a5d62eda0a21b4de62c912c034c3914a852ed117fa1d5a908a4b0e7b70dc6a3 WatchSource:0}: Error finding container 0a5d62eda0a21b4de62c912c034c3914a852ed117fa1d5a908a4b0e7b70dc6a3: Status 404 returned error can't find the container with id 0a5d62eda0a21b4de62c912c034c3914a852ed117fa1d5a908a4b0e7b70dc6a3 Mar 13 10:24:25 crc kubenswrapper[4632]: I0313 10:24:25.424196 4632 generic.go:334] "Generic (PLEG): container finished" podID="bc39c52e-008f-40c1-b93b-532707127fcd" containerID="c1060b701ea818cb4c5d1e5e94618270eed048e8bd16d50775b43c9b34c6b1b9" exitCode=0 Mar 13 10:24:25 crc kubenswrapper[4632]: I0313 10:24:25.424275 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mkdcg" event={"ID":"bc39c52e-008f-40c1-b93b-532707127fcd","Type":"ContainerDied","Data":"c1060b701ea818cb4c5d1e5e94618270eed048e8bd16d50775b43c9b34c6b1b9"} Mar 13 10:24:25 crc kubenswrapper[4632]: I0313 10:24:25.425727 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l6hpb" event={"ID":"4f1c5663-463b-45e2-b200-64e73e6d5698","Type":"ContainerStarted","Data":"0a5d62eda0a21b4de62c912c034c3914a852ed117fa1d5a908a4b0e7b70dc6a3"} Mar 13 10:24:25 crc kubenswrapper[4632]: I0313 10:24:25.427711 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-lrjmj" event={"ID":"2d670715-74f3-46a6-974c-b6953af9fdb7","Type":"ContainerDied","Data":"4e87d116be82121d7fc1873346d45f93e13e538ff7fb3e09adc605646d99d9ef"} Mar 13 10:24:25 crc kubenswrapper[4632]: I0313 10:24:25.427752 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4e87d116be82121d7fc1873346d45f93e13e538ff7fb3e09adc605646d99d9ef" Mar 13 10:24:25 crc kubenswrapper[4632]: I0313 10:24:25.427762 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-lrjmj" Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.802904 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.928397 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-swiftconf\") pod \"bc39c52e-008f-40c1-b93b-532707127fcd\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.928480 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-dispersionconf\") pod \"bc39c52e-008f-40c1-b93b-532707127fcd\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.928596 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bc39c52e-008f-40c1-b93b-532707127fcd-ring-data-devices\") pod \"bc39c52e-008f-40c1-b93b-532707127fcd\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.928634 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc39c52e-008f-40c1-b93b-532707127fcd-scripts\") pod \"bc39c52e-008f-40c1-b93b-532707127fcd\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.928695 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bc39c52e-008f-40c1-b93b-532707127fcd-etc-swift\") pod \"bc39c52e-008f-40c1-b93b-532707127fcd\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.928735 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v7f9b\" (UniqueName: \"kubernetes.io/projected/bc39c52e-008f-40c1-b93b-532707127fcd-kube-api-access-v7f9b\") pod \"bc39c52e-008f-40c1-b93b-532707127fcd\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.928786 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-combined-ca-bundle\") pod \"bc39c52e-008f-40c1-b93b-532707127fcd\" (UID: \"bc39c52e-008f-40c1-b93b-532707127fcd\") " Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.929365 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc39c52e-008f-40c1-b93b-532707127fcd-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "bc39c52e-008f-40c1-b93b-532707127fcd" (UID: "bc39c52e-008f-40c1-b93b-532707127fcd"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.929672 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc39c52e-008f-40c1-b93b-532707127fcd-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "bc39c52e-008f-40c1-b93b-532707127fcd" (UID: "bc39c52e-008f-40c1-b93b-532707127fcd"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.930278 4632 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/bc39c52e-008f-40c1-b93b-532707127fcd-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.930296 4632 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/bc39c52e-008f-40c1-b93b-532707127fcd-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.933858 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc39c52e-008f-40c1-b93b-532707127fcd-kube-api-access-v7f9b" (OuterVolumeSpecName: "kube-api-access-v7f9b") pod "bc39c52e-008f-40c1-b93b-532707127fcd" (UID: "bc39c52e-008f-40c1-b93b-532707127fcd"). InnerVolumeSpecName "kube-api-access-v7f9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.950055 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc39c52e-008f-40c1-b93b-532707127fcd-scripts" (OuterVolumeSpecName: "scripts") pod "bc39c52e-008f-40c1-b93b-532707127fcd" (UID: "bc39c52e-008f-40c1-b93b-532707127fcd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.955542 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "bc39c52e-008f-40c1-b93b-532707127fcd" (UID: "bc39c52e-008f-40c1-b93b-532707127fcd"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.970766 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "bc39c52e-008f-40c1-b93b-532707127fcd" (UID: "bc39c52e-008f-40c1-b93b-532707127fcd"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:24:26 crc kubenswrapper[4632]: I0313 10:24:26.973268 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc39c52e-008f-40c1-b93b-532707127fcd" (UID: "bc39c52e-008f-40c1-b93b-532707127fcd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:24:27 crc kubenswrapper[4632]: I0313 10:24:27.031485 4632 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:27 crc kubenswrapper[4632]: I0313 10:24:27.031527 4632 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:27 crc kubenswrapper[4632]: I0313 10:24:27.031540 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bc39c52e-008f-40c1-b93b-532707127fcd-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:27 crc kubenswrapper[4632]: I0313 10:24:27.031552 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v7f9b\" (UniqueName: \"kubernetes.io/projected/bc39c52e-008f-40c1-b93b-532707127fcd-kube-api-access-v7f9b\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:27 crc kubenswrapper[4632]: I0313 10:24:27.031563 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc39c52e-008f-40c1-b93b-532707127fcd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:27 crc kubenswrapper[4632]: I0313 10:24:27.449188 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-mkdcg" event={"ID":"bc39c52e-008f-40c1-b93b-532707127fcd","Type":"ContainerDied","Data":"b9c30e7d71115b424270718b1169d7d7c69bba98c01559c66a08b1a331e3ccdd"} Mar 13 10:24:27 crc kubenswrapper[4632]: I0313 10:24:27.449237 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b9c30e7d71115b424270718b1169d7d7c69bba98c01559c66a08b1a331e3ccdd" Mar 13 10:24:27 crc kubenswrapper[4632]: I0313 10:24:27.449300 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-mkdcg" Mar 13 10:24:28 crc kubenswrapper[4632]: I0313 10:24:28.462253 4632 generic.go:334] "Generic (PLEG): container finished" podID="159c6cee-c82b-4725-82d6-dbd27216f53c" containerID="d5bd67d741203861cfd1afa23ec3f20fd6236a99625563ac3c10816dbb2a6677" exitCode=0 Mar 13 10:24:28 crc kubenswrapper[4632]: I0313 10:24:28.462818 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"159c6cee-c82b-4725-82d6-dbd27216f53c","Type":"ContainerDied","Data":"d5bd67d741203861cfd1afa23ec3f20fd6236a99625563ac3c10816dbb2a6677"} Mar 13 10:24:28 crc kubenswrapper[4632]: I0313 10:24:28.465827 4632 generic.go:334] "Generic (PLEG): container finished" podID="211718f0-f29c-457b-bc2b-487bb76d4801" containerID="92d546a480b1e583e7b11dc48ab2d570a4a8d7af0616de2352d72ca175520f17" exitCode=0 Mar 13 10:24:28 crc kubenswrapper[4632]: I0313 10:24:28.465869 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"211718f0-f29c-457b-bc2b-487bb76d4801","Type":"ContainerDied","Data":"92d546a480b1e583e7b11dc48ab2d570a4a8d7af0616de2352d72ca175520f17"} Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.475088 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"159c6cee-c82b-4725-82d6-dbd27216f53c","Type":"ContainerStarted","Data":"d8fa91cb90a686638520d703bb5ab925cd9f40c680cdbe53067f753945b6ae3f"} Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.475384 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.478919 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"211718f0-f29c-457b-bc2b-487bb76d4801","Type":"ContainerStarted","Data":"40d92cf95f1cc26685e0359414b43dbdc31eeb90ab4b39c564b241d3fcc263fe"} Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.479213 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.516492 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.349846911 podStartE2EDuration="1m20.516476465s" podCreationTimestamp="2026-03-13 10:23:09 +0000 UTC" firstStartedPulling="2026-03-13 10:23:11.720523443 +0000 UTC m=+1165.743053576" lastFinishedPulling="2026-03-13 10:23:54.887152997 +0000 UTC m=+1208.909683130" observedRunningTime="2026-03-13 10:24:29.513441459 +0000 UTC m=+1243.535971592" watchObservedRunningTime="2026-03-13 10:24:29.516476465 +0000 UTC m=+1243.539006598" Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.531826 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-9kd7r" podUID="eab798dd-482a-4c66-983b-908966cd1f94" containerName="ovn-controller" probeResult="failure" output=< Mar 13 10:24:29 crc kubenswrapper[4632]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 13 10:24:29 crc kubenswrapper[4632]: > Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.541037 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.786566195 podStartE2EDuration="1m20.541023066s" podCreationTimestamp="2026-03-13 10:23:09 +0000 UTC" firstStartedPulling="2026-03-13 10:23:12.100430705 +0000 UTC m=+1166.122960838" lastFinishedPulling="2026-03-13 10:23:54.854887576 +0000 UTC m=+1208.877417709" observedRunningTime="2026-03-13 10:24:29.534485606 +0000 UTC m=+1243.557015739" watchObservedRunningTime="2026-03-13 10:24:29.541023066 +0000 UTC m=+1243.563553189" Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.632351 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.646500 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-c5xnp" Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.884393 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9kd7r-config-d8xq4"] Mar 13 10:24:29 crc kubenswrapper[4632]: E0313 10:24:29.884778 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc39c52e-008f-40c1-b93b-532707127fcd" containerName="swift-ring-rebalance" Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.884795 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc39c52e-008f-40c1-b93b-532707127fcd" containerName="swift-ring-rebalance" Mar 13 10:24:29 crc kubenswrapper[4632]: E0313 10:24:29.884809 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d670715-74f3-46a6-974c-b6953af9fdb7" containerName="mariadb-account-create-update" Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.884817 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d670715-74f3-46a6-974c-b6953af9fdb7" containerName="mariadb-account-create-update" Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.885061 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d670715-74f3-46a6-974c-b6953af9fdb7" containerName="mariadb-account-create-update" Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.885079 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc39c52e-008f-40c1-b93b-532707127fcd" containerName="swift-ring-rebalance" Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.885704 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.890728 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 10:24:29 crc kubenswrapper[4632]: I0313 10:24:29.908785 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9kd7r-config-d8xq4"] Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.004373 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-log-ovn\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.004518 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-run\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.004692 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8dae62c5-076b-4a06-9d12-c955d9131ef3-scripts\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.004778 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-run-ovn\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.005007 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8dae62c5-076b-4a06-9d12-c955d9131ef3-additional-scripts\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.005079 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2fkr\" (UniqueName: \"kubernetes.io/projected/8dae62c5-076b-4a06-9d12-c955d9131ef3-kube-api-access-q2fkr\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.106415 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8dae62c5-076b-4a06-9d12-c955d9131ef3-additional-scripts\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.106482 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2fkr\" (UniqueName: \"kubernetes.io/projected/8dae62c5-076b-4a06-9d12-c955d9131ef3-kube-api-access-q2fkr\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.106575 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-log-ovn\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.106628 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-run\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.106703 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8dae62c5-076b-4a06-9d12-c955d9131ef3-scripts\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.106766 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-run-ovn\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.106912 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-log-ovn\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.106987 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-run-ovn\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.107499 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-run\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.108487 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8dae62c5-076b-4a06-9d12-c955d9131ef3-additional-scripts\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.109256 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8dae62c5-076b-4a06-9d12-c955d9131ef3-scripts\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.129677 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2fkr\" (UniqueName: \"kubernetes.io/projected/8dae62c5-076b-4a06-9d12-c955d9131ef3-kube-api-access-q2fkr\") pod \"ovn-controller-9kd7r-config-d8xq4\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.209680 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:30 crc kubenswrapper[4632]: I0313 10:24:30.913692 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9kd7r-config-d8xq4"] Mar 13 10:24:31 crc kubenswrapper[4632]: I0313 10:24:31.538575 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9kd7r-config-d8xq4" event={"ID":"8dae62c5-076b-4a06-9d12-c955d9131ef3","Type":"ContainerStarted","Data":"572bb794023bd7d53a23050c721933f004db547126df9eaf9b5f8e767603f2d3"} Mar 13 10:24:31 crc kubenswrapper[4632]: I0313 10:24:31.538888 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9kd7r-config-d8xq4" event={"ID":"8dae62c5-076b-4a06-9d12-c955d9131ef3","Type":"ContainerStarted","Data":"f1388733facb7a844093a7b6aa30ee0201fcf9db5bf9311d8265080c24484714"} Mar 13 10:24:31 crc kubenswrapper[4632]: I0313 10:24:31.557970 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-9kd7r-config-d8xq4" podStartSLOduration=2.557928129 podStartE2EDuration="2.557928129s" podCreationTimestamp="2026-03-13 10:24:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:31.555880268 +0000 UTC m=+1245.578410401" watchObservedRunningTime="2026-03-13 10:24:31.557928129 +0000 UTC m=+1245.580458262" Mar 13 10:24:32 crc kubenswrapper[4632]: I0313 10:24:32.567253 4632 generic.go:334] "Generic (PLEG): container finished" podID="8dae62c5-076b-4a06-9d12-c955d9131ef3" containerID="572bb794023bd7d53a23050c721933f004db547126df9eaf9b5f8e767603f2d3" exitCode=0 Mar 13 10:24:32 crc kubenswrapper[4632]: I0313 10:24:32.567251 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9kd7r-config-d8xq4" event={"ID":"8dae62c5-076b-4a06-9d12-c955d9131ef3","Type":"ContainerDied","Data":"572bb794023bd7d53a23050c721933f004db547126df9eaf9b5f8e767603f2d3"} Mar 13 10:24:33 crc kubenswrapper[4632]: I0313 10:24:33.968447 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.098877 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-run\") pod \"8dae62c5-076b-4a06-9d12-c955d9131ef3\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.098984 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-run" (OuterVolumeSpecName: "var-run") pod "8dae62c5-076b-4a06-9d12-c955d9131ef3" (UID: "8dae62c5-076b-4a06-9d12-c955d9131ef3"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.099121 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8dae62c5-076b-4a06-9d12-c955d9131ef3-additional-scripts\") pod \"8dae62c5-076b-4a06-9d12-c955d9131ef3\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.099166 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8dae62c5-076b-4a06-9d12-c955d9131ef3-scripts\") pod \"8dae62c5-076b-4a06-9d12-c955d9131ef3\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.099188 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-log-ovn\") pod \"8dae62c5-076b-4a06-9d12-c955d9131ef3\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.099213 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2fkr\" (UniqueName: \"kubernetes.io/projected/8dae62c5-076b-4a06-9d12-c955d9131ef3-kube-api-access-q2fkr\") pod \"8dae62c5-076b-4a06-9d12-c955d9131ef3\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.099303 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "8dae62c5-076b-4a06-9d12-c955d9131ef3" (UID: "8dae62c5-076b-4a06-9d12-c955d9131ef3"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.099349 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-run-ovn\") pod \"8dae62c5-076b-4a06-9d12-c955d9131ef3\" (UID: \"8dae62c5-076b-4a06-9d12-c955d9131ef3\") " Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.099489 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "8dae62c5-076b-4a06-9d12-c955d9131ef3" (UID: "8dae62c5-076b-4a06-9d12-c955d9131ef3"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.099828 4632 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.099870 4632 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.099880 4632 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8dae62c5-076b-4a06-9d12-c955d9131ef3-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.100197 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dae62c5-076b-4a06-9d12-c955d9131ef3-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "8dae62c5-076b-4a06-9d12-c955d9131ef3" (UID: "8dae62c5-076b-4a06-9d12-c955d9131ef3"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.100521 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8dae62c5-076b-4a06-9d12-c955d9131ef3-scripts" (OuterVolumeSpecName: "scripts") pod "8dae62c5-076b-4a06-9d12-c955d9131ef3" (UID: "8dae62c5-076b-4a06-9d12-c955d9131ef3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.109233 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dae62c5-076b-4a06-9d12-c955d9131ef3-kube-api-access-q2fkr" (OuterVolumeSpecName: "kube-api-access-q2fkr") pod "8dae62c5-076b-4a06-9d12-c955d9131ef3" (UID: "8dae62c5-076b-4a06-9d12-c955d9131ef3"). InnerVolumeSpecName "kube-api-access-q2fkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.201463 4632 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8dae62c5-076b-4a06-9d12-c955d9131ef3-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.201499 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8dae62c5-076b-4a06-9d12-c955d9131ef3-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.201512 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2fkr\" (UniqueName: \"kubernetes.io/projected/8dae62c5-076b-4a06-9d12-c955d9131ef3-kube-api-access-q2fkr\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.539965 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-9kd7r" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.591488 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9kd7r-config-d8xq4" event={"ID":"8dae62c5-076b-4a06-9d12-c955d9131ef3","Type":"ContainerDied","Data":"f1388733facb7a844093a7b6aa30ee0201fcf9db5bf9311d8265080c24484714"} Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.591545 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1388733facb7a844093a7b6aa30ee0201fcf9db5bf9311d8265080c24484714" Mar 13 10:24:34 crc kubenswrapper[4632]: I0313 10:24:34.591585 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9kd7r-config-d8xq4" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.156523 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-9kd7r-config-d8xq4"] Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.178151 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-9kd7r-config-d8xq4"] Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.337895 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9kd7r-config-jtj4g"] Mar 13 10:24:35 crc kubenswrapper[4632]: E0313 10:24:35.338361 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dae62c5-076b-4a06-9d12-c955d9131ef3" containerName="ovn-config" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.338387 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dae62c5-076b-4a06-9d12-c955d9131ef3" containerName="ovn-config" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.338624 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dae62c5-076b-4a06-9d12-c955d9131ef3" containerName="ovn-config" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.339269 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.343662 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.370869 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9kd7r-config-jtj4g"] Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.431668 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-log-ovn\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.431715 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-run\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.431752 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n4j57\" (UniqueName: \"kubernetes.io/projected/67f877b5-12d3-4b48-a9eb-9ee2629e830a-kube-api-access-n4j57\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.431776 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67f877b5-12d3-4b48-a9eb-9ee2629e830a-scripts\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.431830 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-run-ovn\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.431864 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/67f877b5-12d3-4b48-a9eb-9ee2629e830a-additional-scripts\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.533668 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/67f877b5-12d3-4b48-a9eb-9ee2629e830a-additional-scripts\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.533777 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-log-ovn\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.533807 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-run\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.533848 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n4j57\" (UniqueName: \"kubernetes.io/projected/67f877b5-12d3-4b48-a9eb-9ee2629e830a-kube-api-access-n4j57\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.533880 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67f877b5-12d3-4b48-a9eb-9ee2629e830a-scripts\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.533967 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-run-ovn\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.534304 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-run-ovn\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.535428 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/67f877b5-12d3-4b48-a9eb-9ee2629e830a-additional-scripts\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.535500 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-log-ovn\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.535548 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-run\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.537786 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67f877b5-12d3-4b48-a9eb-9ee2629e830a-scripts\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.580983 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n4j57\" (UniqueName: \"kubernetes.io/projected/67f877b5-12d3-4b48-a9eb-9ee2629e830a-kube-api-access-n4j57\") pod \"ovn-controller-9kd7r-config-jtj4g\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:35 crc kubenswrapper[4632]: I0313 10:24:35.656046 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:36 crc kubenswrapper[4632]: I0313 10:24:36.056320 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8dae62c5-076b-4a06-9d12-c955d9131ef3" path="/var/lib/kubelet/pods/8dae62c5-076b-4a06-9d12-c955d9131ef3/volumes" Mar 13 10:24:39 crc kubenswrapper[4632]: I0313 10:24:39.646230 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:39 crc kubenswrapper[4632]: I0313 10:24:39.654932 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e37b3d77-de2e-4be9-9984-550d4ba0f2f0-etc-swift\") pod \"swift-storage-0\" (UID: \"e37b3d77-de2e-4be9-9984-550d4ba0f2f0\") " pod="openstack/swift-storage-0" Mar 13 10:24:39 crc kubenswrapper[4632]: I0313 10:24:39.794973 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 13 10:24:40 crc kubenswrapper[4632]: I0313 10:24:40.460502 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:24:40 crc kubenswrapper[4632]: I0313 10:24:40.460568 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:24:40 crc kubenswrapper[4632]: I0313 10:24:40.460612 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:24:40 crc kubenswrapper[4632]: I0313 10:24:40.461302 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e9a22f93dffae95945f5e47a3d15b0ebe11dc6b72712dcbe34fa0191ff687b27"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:24:40 crc kubenswrapper[4632]: I0313 10:24:40.461368 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://e9a22f93dffae95945f5e47a3d15b0ebe11dc6b72712dcbe34fa0191ff687b27" gracePeriod=600 Mar 13 10:24:40 crc kubenswrapper[4632]: I0313 10:24:40.637332 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:24:40 crc kubenswrapper[4632]: I0313 10:24:40.646780 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="e9a22f93dffae95945f5e47a3d15b0ebe11dc6b72712dcbe34fa0191ff687b27" exitCode=0 Mar 13 10:24:40 crc kubenswrapper[4632]: I0313 10:24:40.646809 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"e9a22f93dffae95945f5e47a3d15b0ebe11dc6b72712dcbe34fa0191ff687b27"} Mar 13 10:24:40 crc kubenswrapper[4632]: I0313 10:24:40.646832 4632 scope.go:117] "RemoveContainer" containerID="624a339b1e1f8b218223c2e3440b7f9925bb18567bb6def4fcf3bfc022198658" Mar 13 10:24:41 crc kubenswrapper[4632]: I0313 10:24:41.067219 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 13 10:24:42 crc kubenswrapper[4632]: I0313 10:24:42.846959 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-pnvjb"] Mar 13 10:24:42 crc kubenswrapper[4632]: I0313 10:24:42.847933 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pnvjb" Mar 13 10:24:42 crc kubenswrapper[4632]: I0313 10:24:42.870014 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-pnvjb"] Mar 13 10:24:42 crc kubenswrapper[4632]: I0313 10:24:42.921579 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb216b07-9809-4b8b-857b-ac1192747b9c-operator-scripts\") pod \"cinder-db-create-pnvjb\" (UID: \"cb216b07-9809-4b8b-857b-ac1192747b9c\") " pod="openstack/cinder-db-create-pnvjb" Mar 13 10:24:42 crc kubenswrapper[4632]: I0313 10:24:42.921670 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fgncl\" (UniqueName: \"kubernetes.io/projected/cb216b07-9809-4b8b-857b-ac1192747b9c-kube-api-access-fgncl\") pod \"cinder-db-create-pnvjb\" (UID: \"cb216b07-9809-4b8b-857b-ac1192747b9c\") " pod="openstack/cinder-db-create-pnvjb" Mar 13 10:24:42 crc kubenswrapper[4632]: I0313 10:24:42.976836 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-21a0-account-create-update-4clr7"] Mar 13 10:24:42 crc kubenswrapper[4632]: I0313 10:24:42.981029 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-21a0-account-create-update-4clr7" Mar 13 10:24:42 crc kubenswrapper[4632]: I0313 10:24:42.994503 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 13 10:24:42 crc kubenswrapper[4632]: I0313 10:24:42.995717 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-21a0-account-create-update-4clr7"] Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.024058 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fgncl\" (UniqueName: \"kubernetes.io/projected/cb216b07-9809-4b8b-857b-ac1192747b9c-kube-api-access-fgncl\") pod \"cinder-db-create-pnvjb\" (UID: \"cb216b07-9809-4b8b-857b-ac1192747b9c\") " pod="openstack/cinder-db-create-pnvjb" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.024185 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a03b92ea-cd2c-455d-a88e-1d57b958b138-operator-scripts\") pod \"cinder-21a0-account-create-update-4clr7\" (UID: \"a03b92ea-cd2c-455d-a88e-1d57b958b138\") " pod="openstack/cinder-21a0-account-create-update-4clr7" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.024334 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb216b07-9809-4b8b-857b-ac1192747b9c-operator-scripts\") pod \"cinder-db-create-pnvjb\" (UID: \"cb216b07-9809-4b8b-857b-ac1192747b9c\") " pod="openstack/cinder-db-create-pnvjb" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.024410 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skrpd\" (UniqueName: \"kubernetes.io/projected/a03b92ea-cd2c-455d-a88e-1d57b958b138-kube-api-access-skrpd\") pod \"cinder-21a0-account-create-update-4clr7\" (UID: \"a03b92ea-cd2c-455d-a88e-1d57b958b138\") " pod="openstack/cinder-21a0-account-create-update-4clr7" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.025687 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb216b07-9809-4b8b-857b-ac1192747b9c-operator-scripts\") pod \"cinder-db-create-pnvjb\" (UID: \"cb216b07-9809-4b8b-857b-ac1192747b9c\") " pod="openstack/cinder-db-create-pnvjb" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.080024 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fgncl\" (UniqueName: \"kubernetes.io/projected/cb216b07-9809-4b8b-857b-ac1192747b9c-kube-api-access-fgncl\") pod \"cinder-db-create-pnvjb\" (UID: \"cb216b07-9809-4b8b-857b-ac1192747b9c\") " pod="openstack/cinder-db-create-pnvjb" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.126213 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a03b92ea-cd2c-455d-a88e-1d57b958b138-operator-scripts\") pod \"cinder-21a0-account-create-update-4clr7\" (UID: \"a03b92ea-cd2c-455d-a88e-1d57b958b138\") " pod="openstack/cinder-21a0-account-create-update-4clr7" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.126657 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skrpd\" (UniqueName: \"kubernetes.io/projected/a03b92ea-cd2c-455d-a88e-1d57b958b138-kube-api-access-skrpd\") pod \"cinder-21a0-account-create-update-4clr7\" (UID: \"a03b92ea-cd2c-455d-a88e-1d57b958b138\") " pod="openstack/cinder-21a0-account-create-update-4clr7" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.127651 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a03b92ea-cd2c-455d-a88e-1d57b958b138-operator-scripts\") pod \"cinder-21a0-account-create-update-4clr7\" (UID: \"a03b92ea-cd2c-455d-a88e-1d57b958b138\") " pod="openstack/cinder-21a0-account-create-update-4clr7" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.159089 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skrpd\" (UniqueName: \"kubernetes.io/projected/a03b92ea-cd2c-455d-a88e-1d57b958b138-kube-api-access-skrpd\") pod \"cinder-21a0-account-create-update-4clr7\" (UID: \"a03b92ea-cd2c-455d-a88e-1d57b958b138\") " pod="openstack/cinder-21a0-account-create-update-4clr7" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.171789 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pnvjb" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.306049 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-21a0-account-create-update-4clr7" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.314238 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-kp87n"] Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.315259 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kp87n" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.322717 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-kp87n"] Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.409254 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-mq9np"] Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.410191 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mq9np" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.415910 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.416931 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.419657 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-llpcf" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.422785 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.432130 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcdcfad1-d735-4b55-ae65-0ce16bdbc79d-operator-scripts\") pod \"heat-db-create-kp87n\" (UID: \"bcdcfad1-d735-4b55-ae65-0ce16bdbc79d\") " pod="openstack/heat-db-create-kp87n" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.432414 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdvfj\" (UniqueName: \"kubernetes.io/projected/bcdcfad1-d735-4b55-ae65-0ce16bdbc79d-kube-api-access-kdvfj\") pod \"heat-db-create-kp87n\" (UID: \"bcdcfad1-d735-4b55-ae65-0ce16bdbc79d\") " pod="openstack/heat-db-create-kp87n" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.441486 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-mq9np"] Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.474151 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-b742-account-create-update-gfdkg"] Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.475089 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-b742-account-create-update-gfdkg" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.482159 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.513432 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-dwf4b"] Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.514732 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dwf4b" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.531545 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-b742-account-create-update-gfdkg"] Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.534291 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hh8q\" (UniqueName: \"kubernetes.io/projected/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-kube-api-access-5hh8q\") pod \"keystone-db-sync-mq9np\" (UID: \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\") " pod="openstack/keystone-db-sync-mq9np" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.534348 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdvfj\" (UniqueName: \"kubernetes.io/projected/bcdcfad1-d735-4b55-ae65-0ce16bdbc79d-kube-api-access-kdvfj\") pod \"heat-db-create-kp87n\" (UID: \"bcdcfad1-d735-4b55-ae65-0ce16bdbc79d\") " pod="openstack/heat-db-create-kp87n" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.534380 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-combined-ca-bundle\") pod \"keystone-db-sync-mq9np\" (UID: \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\") " pod="openstack/keystone-db-sync-mq9np" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.534400 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee94a050-f905-44f1-a5da-16536b8cdfa7-operator-scripts\") pod \"heat-b742-account-create-update-gfdkg\" (UID: \"ee94a050-f905-44f1-a5da-16536b8cdfa7\") " pod="openstack/heat-b742-account-create-update-gfdkg" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.534433 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-config-data\") pod \"keystone-db-sync-mq9np\" (UID: \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\") " pod="openstack/keystone-db-sync-mq9np" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.534460 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zg7q\" (UniqueName: \"kubernetes.io/projected/ee94a050-f905-44f1-a5da-16536b8cdfa7-kube-api-access-7zg7q\") pod \"heat-b742-account-create-update-gfdkg\" (UID: \"ee94a050-f905-44f1-a5da-16536b8cdfa7\") " pod="openstack/heat-b742-account-create-update-gfdkg" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.534501 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcdcfad1-d735-4b55-ae65-0ce16bdbc79d-operator-scripts\") pod \"heat-db-create-kp87n\" (UID: \"bcdcfad1-d735-4b55-ae65-0ce16bdbc79d\") " pod="openstack/heat-db-create-kp87n" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.535128 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcdcfad1-d735-4b55-ae65-0ce16bdbc79d-operator-scripts\") pod \"heat-db-create-kp87n\" (UID: \"bcdcfad1-d735-4b55-ae65-0ce16bdbc79d\") " pod="openstack/heat-db-create-kp87n" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.554520 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-dwf4b"] Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.595121 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdvfj\" (UniqueName: \"kubernetes.io/projected/bcdcfad1-d735-4b55-ae65-0ce16bdbc79d-kube-api-access-kdvfj\") pod \"heat-db-create-kp87n\" (UID: \"bcdcfad1-d735-4b55-ae65-0ce16bdbc79d\") " pod="openstack/heat-db-create-kp87n" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.641706 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kp87n" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.645991 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hh8q\" (UniqueName: \"kubernetes.io/projected/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-kube-api-access-5hh8q\") pod \"keystone-db-sync-mq9np\" (UID: \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\") " pod="openstack/keystone-db-sync-mq9np" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.646084 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/239c554e-360d-4f04-86f0-b2b98974bad3-operator-scripts\") pod \"neutron-db-create-dwf4b\" (UID: \"239c554e-360d-4f04-86f0-b2b98974bad3\") " pod="openstack/neutron-db-create-dwf4b" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.646128 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hprxm\" (UniqueName: \"kubernetes.io/projected/239c554e-360d-4f04-86f0-b2b98974bad3-kube-api-access-hprxm\") pod \"neutron-db-create-dwf4b\" (UID: \"239c554e-360d-4f04-86f0-b2b98974bad3\") " pod="openstack/neutron-db-create-dwf4b" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.646154 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-combined-ca-bundle\") pod \"keystone-db-sync-mq9np\" (UID: \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\") " pod="openstack/keystone-db-sync-mq9np" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.646180 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee94a050-f905-44f1-a5da-16536b8cdfa7-operator-scripts\") pod \"heat-b742-account-create-update-gfdkg\" (UID: \"ee94a050-f905-44f1-a5da-16536b8cdfa7\") " pod="openstack/heat-b742-account-create-update-gfdkg" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.646235 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-config-data\") pod \"keystone-db-sync-mq9np\" (UID: \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\") " pod="openstack/keystone-db-sync-mq9np" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.646271 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zg7q\" (UniqueName: \"kubernetes.io/projected/ee94a050-f905-44f1-a5da-16536b8cdfa7-kube-api-access-7zg7q\") pod \"heat-b742-account-create-update-gfdkg\" (UID: \"ee94a050-f905-44f1-a5da-16536b8cdfa7\") " pod="openstack/heat-b742-account-create-update-gfdkg" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.652990 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee94a050-f905-44f1-a5da-16536b8cdfa7-operator-scripts\") pod \"heat-b742-account-create-update-gfdkg\" (UID: \"ee94a050-f905-44f1-a5da-16536b8cdfa7\") " pod="openstack/heat-b742-account-create-update-gfdkg" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.674896 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-combined-ca-bundle\") pod \"keystone-db-sync-mq9np\" (UID: \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\") " pod="openstack/keystone-db-sync-mq9np" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.675632 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-config-data\") pod \"keystone-db-sync-mq9np\" (UID: \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\") " pod="openstack/keystone-db-sync-mq9np" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.707696 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hh8q\" (UniqueName: \"kubernetes.io/projected/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-kube-api-access-5hh8q\") pod \"keystone-db-sync-mq9np\" (UID: \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\") " pod="openstack/keystone-db-sync-mq9np" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.722871 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-da66-account-create-update-tk8pd"] Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.724150 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-da66-account-create-update-tk8pd" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.731662 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mq9np" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.732655 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.749015 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/239c554e-360d-4f04-86f0-b2b98974bad3-operator-scripts\") pod \"neutron-db-create-dwf4b\" (UID: \"239c554e-360d-4f04-86f0-b2b98974bad3\") " pod="openstack/neutron-db-create-dwf4b" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.749285 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hprxm\" (UniqueName: \"kubernetes.io/projected/239c554e-360d-4f04-86f0-b2b98974bad3-kube-api-access-hprxm\") pod \"neutron-db-create-dwf4b\" (UID: \"239c554e-360d-4f04-86f0-b2b98974bad3\") " pod="openstack/neutron-db-create-dwf4b" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.750155 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/239c554e-360d-4f04-86f0-b2b98974bad3-operator-scripts\") pod \"neutron-db-create-dwf4b\" (UID: \"239c554e-360d-4f04-86f0-b2b98974bad3\") " pod="openstack/neutron-db-create-dwf4b" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.766966 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-da66-account-create-update-tk8pd"] Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.790034 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-g7pfc"] Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.791003 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zg7q\" (UniqueName: \"kubernetes.io/projected/ee94a050-f905-44f1-a5da-16536b8cdfa7-kube-api-access-7zg7q\") pod \"heat-b742-account-create-update-gfdkg\" (UID: \"ee94a050-f905-44f1-a5da-16536b8cdfa7\") " pod="openstack/heat-b742-account-create-update-gfdkg" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.791317 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-g7pfc" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.793528 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-g7pfc"] Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.796316 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-b742-account-create-update-gfdkg" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.843698 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hprxm\" (UniqueName: \"kubernetes.io/projected/239c554e-360d-4f04-86f0-b2b98974bad3-kube-api-access-hprxm\") pod \"neutron-db-create-dwf4b\" (UID: \"239c554e-360d-4f04-86f0-b2b98974bad3\") " pod="openstack/neutron-db-create-dwf4b" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.850485 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsxc4\" (UniqueName: \"kubernetes.io/projected/0d045bc7-38b2-46f5-8cd8-cf634003bedf-kube-api-access-fsxc4\") pod \"neutron-da66-account-create-update-tk8pd\" (UID: \"0d045bc7-38b2-46f5-8cd8-cf634003bedf\") " pod="openstack/neutron-da66-account-create-update-tk8pd" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.850570 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km2sr\" (UniqueName: \"kubernetes.io/projected/aa0000da-8f11-4e97-8ab5-1bcfea0ac894-kube-api-access-km2sr\") pod \"barbican-db-create-g7pfc\" (UID: \"aa0000da-8f11-4e97-8ab5-1bcfea0ac894\") " pod="openstack/barbican-db-create-g7pfc" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.850601 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa0000da-8f11-4e97-8ab5-1bcfea0ac894-operator-scripts\") pod \"barbican-db-create-g7pfc\" (UID: \"aa0000da-8f11-4e97-8ab5-1bcfea0ac894\") " pod="openstack/barbican-db-create-g7pfc" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.850728 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d045bc7-38b2-46f5-8cd8-cf634003bedf-operator-scripts\") pod \"neutron-da66-account-create-update-tk8pd\" (UID: \"0d045bc7-38b2-46f5-8cd8-cf634003bedf\") " pod="openstack/neutron-da66-account-create-update-tk8pd" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.952463 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsxc4\" (UniqueName: \"kubernetes.io/projected/0d045bc7-38b2-46f5-8cd8-cf634003bedf-kube-api-access-fsxc4\") pod \"neutron-da66-account-create-update-tk8pd\" (UID: \"0d045bc7-38b2-46f5-8cd8-cf634003bedf\") " pod="openstack/neutron-da66-account-create-update-tk8pd" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.952558 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km2sr\" (UniqueName: \"kubernetes.io/projected/aa0000da-8f11-4e97-8ab5-1bcfea0ac894-kube-api-access-km2sr\") pod \"barbican-db-create-g7pfc\" (UID: \"aa0000da-8f11-4e97-8ab5-1bcfea0ac894\") " pod="openstack/barbican-db-create-g7pfc" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.952602 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa0000da-8f11-4e97-8ab5-1bcfea0ac894-operator-scripts\") pod \"barbican-db-create-g7pfc\" (UID: \"aa0000da-8f11-4e97-8ab5-1bcfea0ac894\") " pod="openstack/barbican-db-create-g7pfc" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.952660 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d045bc7-38b2-46f5-8cd8-cf634003bedf-operator-scripts\") pod \"neutron-da66-account-create-update-tk8pd\" (UID: \"0d045bc7-38b2-46f5-8cd8-cf634003bedf\") " pod="openstack/neutron-da66-account-create-update-tk8pd" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.953533 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d045bc7-38b2-46f5-8cd8-cf634003bedf-operator-scripts\") pod \"neutron-da66-account-create-update-tk8pd\" (UID: \"0d045bc7-38b2-46f5-8cd8-cf634003bedf\") " pod="openstack/neutron-da66-account-create-update-tk8pd" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.954091 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa0000da-8f11-4e97-8ab5-1bcfea0ac894-operator-scripts\") pod \"barbican-db-create-g7pfc\" (UID: \"aa0000da-8f11-4e97-8ab5-1bcfea0ac894\") " pod="openstack/barbican-db-create-g7pfc" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.985584 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km2sr\" (UniqueName: \"kubernetes.io/projected/aa0000da-8f11-4e97-8ab5-1bcfea0ac894-kube-api-access-km2sr\") pod \"barbican-db-create-g7pfc\" (UID: \"aa0000da-8f11-4e97-8ab5-1bcfea0ac894\") " pod="openstack/barbican-db-create-g7pfc" Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.994224 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-4dec-account-create-update-hfnth"] Mar 13 10:24:43 crc kubenswrapper[4632]: I0313 10:24:43.999244 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsxc4\" (UniqueName: \"kubernetes.io/projected/0d045bc7-38b2-46f5-8cd8-cf634003bedf-kube-api-access-fsxc4\") pod \"neutron-da66-account-create-update-tk8pd\" (UID: \"0d045bc7-38b2-46f5-8cd8-cf634003bedf\") " pod="openstack/neutron-da66-account-create-update-tk8pd" Mar 13 10:24:44 crc kubenswrapper[4632]: I0313 10:24:44.000732 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4dec-account-create-update-hfnth" Mar 13 10:24:44 crc kubenswrapper[4632]: I0313 10:24:44.003066 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 13 10:24:44 crc kubenswrapper[4632]: I0313 10:24:44.021139 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-4dec-account-create-update-hfnth"] Mar 13 10:24:44 crc kubenswrapper[4632]: I0313 10:24:44.055449 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64q7x\" (UniqueName: \"kubernetes.io/projected/47870992-2db9-46f4-84d9-fd50fb9851eb-kube-api-access-64q7x\") pod \"barbican-4dec-account-create-update-hfnth\" (UID: \"47870992-2db9-46f4-84d9-fd50fb9851eb\") " pod="openstack/barbican-4dec-account-create-update-hfnth" Mar 13 10:24:44 crc kubenswrapper[4632]: I0313 10:24:44.055559 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47870992-2db9-46f4-84d9-fd50fb9851eb-operator-scripts\") pod \"barbican-4dec-account-create-update-hfnth\" (UID: \"47870992-2db9-46f4-84d9-fd50fb9851eb\") " pod="openstack/barbican-4dec-account-create-update-hfnth" Mar 13 10:24:44 crc kubenswrapper[4632]: I0313 10:24:44.091843 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-da66-account-create-update-tk8pd" Mar 13 10:24:44 crc kubenswrapper[4632]: I0313 10:24:44.118584 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-g7pfc" Mar 13 10:24:44 crc kubenswrapper[4632]: I0313 10:24:44.140748 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dwf4b" Mar 13 10:24:44 crc kubenswrapper[4632]: I0313 10:24:44.156728 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47870992-2db9-46f4-84d9-fd50fb9851eb-operator-scripts\") pod \"barbican-4dec-account-create-update-hfnth\" (UID: \"47870992-2db9-46f4-84d9-fd50fb9851eb\") " pod="openstack/barbican-4dec-account-create-update-hfnth" Mar 13 10:24:44 crc kubenswrapper[4632]: I0313 10:24:44.157172 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64q7x\" (UniqueName: \"kubernetes.io/projected/47870992-2db9-46f4-84d9-fd50fb9851eb-kube-api-access-64q7x\") pod \"barbican-4dec-account-create-update-hfnth\" (UID: \"47870992-2db9-46f4-84d9-fd50fb9851eb\") " pod="openstack/barbican-4dec-account-create-update-hfnth" Mar 13 10:24:44 crc kubenswrapper[4632]: I0313 10:24:44.157670 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47870992-2db9-46f4-84d9-fd50fb9851eb-operator-scripts\") pod \"barbican-4dec-account-create-update-hfnth\" (UID: \"47870992-2db9-46f4-84d9-fd50fb9851eb\") " pod="openstack/barbican-4dec-account-create-update-hfnth" Mar 13 10:24:44 crc kubenswrapper[4632]: I0313 10:24:44.185053 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64q7x\" (UniqueName: \"kubernetes.io/projected/47870992-2db9-46f4-84d9-fd50fb9851eb-kube-api-access-64q7x\") pod \"barbican-4dec-account-create-update-hfnth\" (UID: \"47870992-2db9-46f4-84d9-fd50fb9851eb\") " pod="openstack/barbican-4dec-account-create-update-hfnth" Mar 13 10:24:44 crc kubenswrapper[4632]: I0313 10:24:44.323251 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4dec-account-create-update-hfnth" Mar 13 10:24:44 crc kubenswrapper[4632]: E0313 10:24:44.730075 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-glance-api:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:24:44 crc kubenswrapper[4632]: E0313 10:24:44.730138 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-glance-api:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:24:44 crc kubenswrapper[4632]: E0313 10:24:44.730274 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-glance-api:e43235cb19da04699a53f42b6a75afe9,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fjbxx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-l6hpb_openstack(4f1c5663-463b-45e2-b200-64e73e6d5698): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:24:44 crc kubenswrapper[4632]: E0313 10:24:44.734066 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-l6hpb" podUID="4f1c5663-463b-45e2-b200-64e73e6d5698" Mar 13 10:24:45 crc kubenswrapper[4632]: I0313 10:24:45.384891 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-21a0-account-create-update-4clr7"] Mar 13 10:24:45 crc kubenswrapper[4632]: I0313 10:24:45.635834 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-mq9np"] Mar 13 10:24:45 crc kubenswrapper[4632]: I0313 10:24:45.762924 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-21a0-account-create-update-4clr7" event={"ID":"a03b92ea-cd2c-455d-a88e-1d57b958b138","Type":"ContainerStarted","Data":"4f01ad0f17dc70f28110656c522ad63fcb71aa346a374ece30c18a84e29887e5"} Mar 13 10:24:45 crc kubenswrapper[4632]: W0313 10:24:45.766606 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode824ae7d_dbbd_496b_b8b0_8b5c59a4d419.slice/crio-1a3020d6e5b66dad152669406220e67cb7be099d82ff8fd4925d6504c1176fb1 WatchSource:0}: Error finding container 1a3020d6e5b66dad152669406220e67cb7be099d82ff8fd4925d6504c1176fb1: Status 404 returned error can't find the container with id 1a3020d6e5b66dad152669406220e67cb7be099d82ff8fd4925d6504c1176fb1 Mar 13 10:24:45 crc kubenswrapper[4632]: E0313 10:24:45.776087 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-glance-api:e43235cb19da04699a53f42b6a75afe9\\\"\"" pod="openstack/glance-db-sync-l6hpb" podUID="4f1c5663-463b-45e2-b200-64e73e6d5698" Mar 13 10:24:45 crc kubenswrapper[4632]: I0313 10:24:45.791604 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.371850 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-kp87n"] Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.383091 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9kd7r-config-jtj4g"] Mar 13 10:24:46 crc kubenswrapper[4632]: W0313 10:24:46.386635 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbcdcfad1_d735_4b55_ae65_0ce16bdbc79d.slice/crio-e1cd870de29627e6c40862edb50e2b89e40d4c4c61f895dbabfb5c6605e37291 WatchSource:0}: Error finding container e1cd870de29627e6c40862edb50e2b89e40d4c4c61f895dbabfb5c6605e37291: Status 404 returned error can't find the container with id e1cd870de29627e6c40862edb50e2b89e40d4c4c61f895dbabfb5c6605e37291 Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.438323 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-dwf4b"] Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.453012 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-da66-account-create-update-tk8pd"] Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.493744 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-g7pfc"] Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.500426 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-4dec-account-create-update-hfnth"] Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.511562 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-pnvjb"] Mar 13 10:24:46 crc kubenswrapper[4632]: W0313 10:24:46.565043 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaa0000da_8f11_4e97_8ab5_1bcfea0ac894.slice/crio-64833dd5dcca1cf0d9b0e7fcfd2870e2d6c8b40bb49e5e7b154150c2bf852051 WatchSource:0}: Error finding container 64833dd5dcca1cf0d9b0e7fcfd2870e2d6c8b40bb49e5e7b154150c2bf852051: Status 404 returned error can't find the container with id 64833dd5dcca1cf0d9b0e7fcfd2870e2d6c8b40bb49e5e7b154150c2bf852051 Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.654106 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-b742-account-create-update-gfdkg"] Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.679856 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.792830 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-da66-account-create-update-tk8pd" event={"ID":"0d045bc7-38b2-46f5-8cd8-cf634003bedf","Type":"ContainerStarted","Data":"6cabee2602a6d9e4308c8db70b1d7f8643862ae4eef1ae7803777760563d87cb"} Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.794887 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dwf4b" event={"ID":"239c554e-360d-4f04-86f0-b2b98974bad3","Type":"ContainerStarted","Data":"d57d142d20024782bc299e1b548d02139291fdbb43f3a8108c7af8762342c79e"} Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.797987 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pnvjb" event={"ID":"cb216b07-9809-4b8b-857b-ac1192747b9c","Type":"ContainerStarted","Data":"417ef1960b2d4ea70aed07efa739673778171eebc874e98d5bdf429380cac86f"} Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.814337 4632 generic.go:334] "Generic (PLEG): container finished" podID="a03b92ea-cd2c-455d-a88e-1d57b958b138" containerID="24cb5f7263654577bea6ec83ce575dcb325e9b55c8adac840790cd7a29363013" exitCode=0 Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.814424 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-21a0-account-create-update-4clr7" event={"ID":"a03b92ea-cd2c-455d-a88e-1d57b958b138","Type":"ContainerDied","Data":"24cb5f7263654577bea6ec83ce575dcb325e9b55c8adac840790cd7a29363013"} Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.845989 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9kd7r-config-jtj4g" event={"ID":"67f877b5-12d3-4b48-a9eb-9ee2629e830a","Type":"ContainerStarted","Data":"ce1ab79e8690eb9d1f16d7b7f5d9ff52729195fe7c60d80326b0842004a4a53d"} Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.852752 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"a148dfa9ef48de458189e9fda19ce88937bedd25c3ec76e22d14f43a4745805f"} Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.862265 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4dec-account-create-update-hfnth" event={"ID":"47870992-2db9-46f4-84d9-fd50fb9851eb","Type":"ContainerStarted","Data":"664b8a378c78fddcf14389393b1cc3a53fe85b08aab7f467156058884d9c4350"} Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.864194 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-b742-account-create-update-gfdkg" event={"ID":"ee94a050-f905-44f1-a5da-16536b8cdfa7","Type":"ContainerStarted","Data":"c71c108608caeb76931caff20f7b0c7e5d8d5c389c0440d31bb543598d28dfb8"} Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.865504 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"0cec076a2f38e51ffee08b1302a0f7ffd219d4ce1350c5fb23f8c33fa0bdbf2d"} Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.871708 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mq9np" event={"ID":"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419","Type":"ContainerStarted","Data":"1a3020d6e5b66dad152669406220e67cb7be099d82ff8fd4925d6504c1176fb1"} Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.879399 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-g7pfc" event={"ID":"aa0000da-8f11-4e97-8ab5-1bcfea0ac894","Type":"ContainerStarted","Data":"64833dd5dcca1cf0d9b0e7fcfd2870e2d6c8b40bb49e5e7b154150c2bf852051"} Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.882313 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-kp87n" event={"ID":"bcdcfad1-d735-4b55-ae65-0ce16bdbc79d","Type":"ContainerStarted","Data":"e1cd870de29627e6c40862edb50e2b89e40d4c4c61f895dbabfb5c6605e37291"} Mar 13 10:24:46 crc kubenswrapper[4632]: I0313 10:24:46.916553 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-kp87n" podStartSLOduration=3.916530163 podStartE2EDuration="3.916530163s" podCreationTimestamp="2026-03-13 10:24:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:46.904050406 +0000 UTC m=+1260.926580549" watchObservedRunningTime="2026-03-13 10:24:46.916530163 +0000 UTC m=+1260.939060306" Mar 13 10:24:47 crc kubenswrapper[4632]: I0313 10:24:47.894975 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-da66-account-create-update-tk8pd" event={"ID":"0d045bc7-38b2-46f5-8cd8-cf634003bedf","Type":"ContainerStarted","Data":"f13e115025698b8daa562f4881b31bb57b43cf222144f35c644ca079c94f546c"} Mar 13 10:24:47 crc kubenswrapper[4632]: I0313 10:24:47.897862 4632 generic.go:334] "Generic (PLEG): container finished" podID="239c554e-360d-4f04-86f0-b2b98974bad3" containerID="209b78ccf3afd3b3582d4d4eae9056be2d6d19f860431a427d43f1899c69be92" exitCode=0 Mar 13 10:24:47 crc kubenswrapper[4632]: I0313 10:24:47.897978 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dwf4b" event={"ID":"239c554e-360d-4f04-86f0-b2b98974bad3","Type":"ContainerDied","Data":"209b78ccf3afd3b3582d4d4eae9056be2d6d19f860431a427d43f1899c69be92"} Mar 13 10:24:47 crc kubenswrapper[4632]: I0313 10:24:47.901476 4632 generic.go:334] "Generic (PLEG): container finished" podID="bcdcfad1-d735-4b55-ae65-0ce16bdbc79d" containerID="1ead25cb79a035bd17ce1b8995cb1c20666089312b5c266ebcbccc7e66e7c0cc" exitCode=0 Mar 13 10:24:47 crc kubenswrapper[4632]: I0313 10:24:47.901575 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-kp87n" event={"ID":"bcdcfad1-d735-4b55-ae65-0ce16bdbc79d","Type":"ContainerDied","Data":"1ead25cb79a035bd17ce1b8995cb1c20666089312b5c266ebcbccc7e66e7c0cc"} Mar 13 10:24:47 crc kubenswrapper[4632]: I0313 10:24:47.904636 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4dec-account-create-update-hfnth" event={"ID":"47870992-2db9-46f4-84d9-fd50fb9851eb","Type":"ContainerStarted","Data":"dc07b5437ef3867ede6e9debff7196fad98555045e8df8dafdb4a11a7fb9808e"} Mar 13 10:24:47 crc kubenswrapper[4632]: I0313 10:24:47.934714 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-4dec-account-create-update-hfnth" podStartSLOduration=4.934692032 podStartE2EDuration="4.934692032s" podCreationTimestamp="2026-03-13 10:24:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:47.934481697 +0000 UTC m=+1261.957011840" watchObservedRunningTime="2026-03-13 10:24:47.934692032 +0000 UTC m=+1261.957222165" Mar 13 10:24:47 crc kubenswrapper[4632]: I0313 10:24:47.937136 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-da66-account-create-update-tk8pd" podStartSLOduration=4.937117261 podStartE2EDuration="4.937117261s" podCreationTimestamp="2026-03-13 10:24:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:47.920324439 +0000 UTC m=+1261.942854572" watchObservedRunningTime="2026-03-13 10:24:47.937117261 +0000 UTC m=+1261.959647394" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.351723 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-21a0-account-create-update-4clr7" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.517900 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skrpd\" (UniqueName: \"kubernetes.io/projected/a03b92ea-cd2c-455d-a88e-1d57b958b138-kube-api-access-skrpd\") pod \"a03b92ea-cd2c-455d-a88e-1d57b958b138\" (UID: \"a03b92ea-cd2c-455d-a88e-1d57b958b138\") " Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.518023 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a03b92ea-cd2c-455d-a88e-1d57b958b138-operator-scripts\") pod \"a03b92ea-cd2c-455d-a88e-1d57b958b138\" (UID: \"a03b92ea-cd2c-455d-a88e-1d57b958b138\") " Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.518983 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a03b92ea-cd2c-455d-a88e-1d57b958b138-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a03b92ea-cd2c-455d-a88e-1d57b958b138" (UID: "a03b92ea-cd2c-455d-a88e-1d57b958b138"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.534220 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a03b92ea-cd2c-455d-a88e-1d57b958b138-kube-api-access-skrpd" (OuterVolumeSpecName: "kube-api-access-skrpd") pod "a03b92ea-cd2c-455d-a88e-1d57b958b138" (UID: "a03b92ea-cd2c-455d-a88e-1d57b958b138"). InnerVolumeSpecName "kube-api-access-skrpd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.625008 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a03b92ea-cd2c-455d-a88e-1d57b958b138-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.625033 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skrpd\" (UniqueName: \"kubernetes.io/projected/a03b92ea-cd2c-455d-a88e-1d57b958b138-kube-api-access-skrpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.922242 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pnvjb" event={"ID":"cb216b07-9809-4b8b-857b-ac1192747b9c","Type":"ContainerStarted","Data":"207587c5bdcbf92f71ab5aedfecf2486734ea587705753fb95e8790e674e977d"} Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.926762 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-21a0-account-create-update-4clr7" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.934636 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-21a0-account-create-update-4clr7" event={"ID":"a03b92ea-cd2c-455d-a88e-1d57b958b138","Type":"ContainerDied","Data":"4f01ad0f17dc70f28110656c522ad63fcb71aa346a374ece30c18a84e29887e5"} Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.934785 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f01ad0f17dc70f28110656c522ad63fcb71aa346a374ece30c18a84e29887e5" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.953610 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-pnvjb" podStartSLOduration=6.953590799 podStartE2EDuration="6.953590799s" podCreationTimestamp="2026-03-13 10:24:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:48.945656064 +0000 UTC m=+1262.968186197" watchObservedRunningTime="2026-03-13 10:24:48.953590799 +0000 UTC m=+1262.976120932" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.955099 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-g7pfc" event={"ID":"aa0000da-8f11-4e97-8ab5-1bcfea0ac894","Type":"ContainerStarted","Data":"f79fdacee095a4d2c557179a3aeeb0eea1874c7280d8a656f2dd9779cf567f1e"} Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.975142 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9kd7r-config-jtj4g" event={"ID":"67f877b5-12d3-4b48-a9eb-9ee2629e830a","Type":"ContainerStarted","Data":"c6b6fdf02c5b942ff5eb86fa09449efd1927d429db47c31ad2d68c9602235d4f"} Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.991261 4632 generic.go:334] "Generic (PLEG): container finished" podID="47870992-2db9-46f4-84d9-fd50fb9851eb" containerID="dc07b5437ef3867ede6e9debff7196fad98555045e8df8dafdb4a11a7fb9808e" exitCode=0 Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.991345 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4dec-account-create-update-hfnth" event={"ID":"47870992-2db9-46f4-84d9-fd50fb9851eb","Type":"ContainerDied","Data":"dc07b5437ef3867ede6e9debff7196fad98555045e8df8dafdb4a11a7fb9808e"} Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:48.991817 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-create-g7pfc" podStartSLOduration=5.991803587 podStartE2EDuration="5.991803587s" podCreationTimestamp="2026-03-13 10:24:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:48.977089746 +0000 UTC m=+1262.999619879" watchObservedRunningTime="2026-03-13 10:24:48.991803587 +0000 UTC m=+1263.014333720" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.001301 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-b742-account-create-update-gfdkg" event={"ID":"ee94a050-f905-44f1-a5da-16536b8cdfa7","Type":"ContainerStarted","Data":"d92125a86d78e277913519dc023b0643c481c49ac75357c10f1cb11e638c36a3"} Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.004981 4632 generic.go:334] "Generic (PLEG): container finished" podID="0d045bc7-38b2-46f5-8cd8-cf634003bedf" containerID="f13e115025698b8daa562f4881b31bb57b43cf222144f35c644ca079c94f546c" exitCode=0 Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.005241 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-da66-account-create-update-tk8pd" event={"ID":"0d045bc7-38b2-46f5-8cd8-cf634003bedf","Type":"ContainerDied","Data":"f13e115025698b8daa562f4881b31bb57b43cf222144f35c644ca079c94f546c"} Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.007912 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-9kd7r-config-jtj4g" podStartSLOduration=14.007897792 podStartE2EDuration="14.007897792s" podCreationTimestamp="2026-03-13 10:24:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:49.002497439 +0000 UTC m=+1263.025027572" watchObservedRunningTime="2026-03-13 10:24:49.007897792 +0000 UTC m=+1263.030427935" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.063826 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-b742-account-create-update-gfdkg" podStartSLOduration=6.063802904 podStartE2EDuration="6.063802904s" podCreationTimestamp="2026-03-13 10:24:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:24:49.029992043 +0000 UTC m=+1263.052522196" watchObservedRunningTime="2026-03-13 10:24:49.063802904 +0000 UTC m=+1263.086333037" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.841361 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kp87n" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.850278 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dwf4b" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.892506 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hprxm\" (UniqueName: \"kubernetes.io/projected/239c554e-360d-4f04-86f0-b2b98974bad3-kube-api-access-hprxm\") pod \"239c554e-360d-4f04-86f0-b2b98974bad3\" (UID: \"239c554e-360d-4f04-86f0-b2b98974bad3\") " Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.892582 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdvfj\" (UniqueName: \"kubernetes.io/projected/bcdcfad1-d735-4b55-ae65-0ce16bdbc79d-kube-api-access-kdvfj\") pod \"bcdcfad1-d735-4b55-ae65-0ce16bdbc79d\" (UID: \"bcdcfad1-d735-4b55-ae65-0ce16bdbc79d\") " Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.892629 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcdcfad1-d735-4b55-ae65-0ce16bdbc79d-operator-scripts\") pod \"bcdcfad1-d735-4b55-ae65-0ce16bdbc79d\" (UID: \"bcdcfad1-d735-4b55-ae65-0ce16bdbc79d\") " Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.892698 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/239c554e-360d-4f04-86f0-b2b98974bad3-operator-scripts\") pod \"239c554e-360d-4f04-86f0-b2b98974bad3\" (UID: \"239c554e-360d-4f04-86f0-b2b98974bad3\") " Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.894608 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/239c554e-360d-4f04-86f0-b2b98974bad3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "239c554e-360d-4f04-86f0-b2b98974bad3" (UID: "239c554e-360d-4f04-86f0-b2b98974bad3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.895300 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bcdcfad1-d735-4b55-ae65-0ce16bdbc79d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bcdcfad1-d735-4b55-ae65-0ce16bdbc79d" (UID: "bcdcfad1-d735-4b55-ae65-0ce16bdbc79d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.899732 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcdcfad1-d735-4b55-ae65-0ce16bdbc79d-kube-api-access-kdvfj" (OuterVolumeSpecName: "kube-api-access-kdvfj") pod "bcdcfad1-d735-4b55-ae65-0ce16bdbc79d" (UID: "bcdcfad1-d735-4b55-ae65-0ce16bdbc79d"). InnerVolumeSpecName "kube-api-access-kdvfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.899877 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/239c554e-360d-4f04-86f0-b2b98974bad3-kube-api-access-hprxm" (OuterVolumeSpecName: "kube-api-access-hprxm") pod "239c554e-360d-4f04-86f0-b2b98974bad3" (UID: "239c554e-360d-4f04-86f0-b2b98974bad3"). InnerVolumeSpecName "kube-api-access-hprxm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.995050 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hprxm\" (UniqueName: \"kubernetes.io/projected/239c554e-360d-4f04-86f0-b2b98974bad3-kube-api-access-hprxm\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.995100 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdvfj\" (UniqueName: \"kubernetes.io/projected/bcdcfad1-d735-4b55-ae65-0ce16bdbc79d-kube-api-access-kdvfj\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.995115 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bcdcfad1-d735-4b55-ae65-0ce16bdbc79d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:49 crc kubenswrapper[4632]: I0313 10:24:49.995127 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/239c554e-360d-4f04-86f0-b2b98974bad3-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.019932 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-kp87n" event={"ID":"bcdcfad1-d735-4b55-ae65-0ce16bdbc79d","Type":"ContainerDied","Data":"e1cd870de29627e6c40862edb50e2b89e40d4c4c61f895dbabfb5c6605e37291"} Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.019993 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1cd870de29627e6c40862edb50e2b89e40d4c4c61f895dbabfb5c6605e37291" Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.020004 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-kp87n" Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.021787 4632 generic.go:334] "Generic (PLEG): container finished" podID="67f877b5-12d3-4b48-a9eb-9ee2629e830a" containerID="c6b6fdf02c5b942ff5eb86fa09449efd1927d429db47c31ad2d68c9602235d4f" exitCode=0 Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.021863 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9kd7r-config-jtj4g" event={"ID":"67f877b5-12d3-4b48-a9eb-9ee2629e830a","Type":"ContainerDied","Data":"c6b6fdf02c5b942ff5eb86fa09449efd1927d429db47c31ad2d68c9602235d4f"} Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.025560 4632 generic.go:334] "Generic (PLEG): container finished" podID="ee94a050-f905-44f1-a5da-16536b8cdfa7" containerID="d92125a86d78e277913519dc023b0643c481c49ac75357c10f1cb11e638c36a3" exitCode=0 Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.025674 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-b742-account-create-update-gfdkg" event={"ID":"ee94a050-f905-44f1-a5da-16536b8cdfa7","Type":"ContainerDied","Data":"d92125a86d78e277913519dc023b0643c481c49ac75357c10f1cb11e638c36a3"} Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.027143 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"00ea6bda0abc557d34319366ebb47ed2d4d334b085146aa961d521324d378058"} Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.028585 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-dwf4b" event={"ID":"239c554e-360d-4f04-86f0-b2b98974bad3","Type":"ContainerDied","Data":"d57d142d20024782bc299e1b548d02139291fdbb43f3a8108c7af8762342c79e"} Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.028628 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d57d142d20024782bc299e1b548d02139291fdbb43f3a8108c7af8762342c79e" Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.028690 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-dwf4b" Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.032988 4632 generic.go:334] "Generic (PLEG): container finished" podID="cb216b07-9809-4b8b-857b-ac1192747b9c" containerID="207587c5bdcbf92f71ab5aedfecf2486734ea587705753fb95e8790e674e977d" exitCode=0 Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.033043 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pnvjb" event={"ID":"cb216b07-9809-4b8b-857b-ac1192747b9c","Type":"ContainerDied","Data":"207587c5bdcbf92f71ab5aedfecf2486734ea587705753fb95e8790e674e977d"} Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.034547 4632 generic.go:334] "Generic (PLEG): container finished" podID="aa0000da-8f11-4e97-8ab5-1bcfea0ac894" containerID="f79fdacee095a4d2c557179a3aeeb0eea1874c7280d8a656f2dd9779cf567f1e" exitCode=0 Mar 13 10:24:50 crc kubenswrapper[4632]: I0313 10:24:50.034586 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-g7pfc" event={"ID":"aa0000da-8f11-4e97-8ab5-1bcfea0ac894","Type":"ContainerDied","Data":"f79fdacee095a4d2c557179a3aeeb0eea1874c7280d8a656f2dd9779cf567f1e"} Mar 13 10:24:51 crc kubenswrapper[4632]: I0313 10:24:51.068714 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"68efd16a592a98149fb78b8c3ca36bfb289a67545bf9bdb5079bbfc32d02d606"} Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.654280 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-g7pfc" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.678592 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pnvjb" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.686453 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4dec-account-create-update-hfnth" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.705392 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.716217 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-b742-account-create-update-gfdkg" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.724002 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-da66-account-create-update-tk8pd" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.797863 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa0000da-8f11-4e97-8ab5-1bcfea0ac894-operator-scripts\") pod \"aa0000da-8f11-4e97-8ab5-1bcfea0ac894\" (UID: \"aa0000da-8f11-4e97-8ab5-1bcfea0ac894\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.797962 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb216b07-9809-4b8b-857b-ac1192747b9c-operator-scripts\") pod \"cb216b07-9809-4b8b-857b-ac1192747b9c\" (UID: \"cb216b07-9809-4b8b-857b-ac1192747b9c\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.798026 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47870992-2db9-46f4-84d9-fd50fb9851eb-operator-scripts\") pod \"47870992-2db9-46f4-84d9-fd50fb9851eb\" (UID: \"47870992-2db9-46f4-84d9-fd50fb9851eb\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.798062 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64q7x\" (UniqueName: \"kubernetes.io/projected/47870992-2db9-46f4-84d9-fd50fb9851eb-kube-api-access-64q7x\") pod \"47870992-2db9-46f4-84d9-fd50fb9851eb\" (UID: \"47870992-2db9-46f4-84d9-fd50fb9851eb\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.798092 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-km2sr\" (UniqueName: \"kubernetes.io/projected/aa0000da-8f11-4e97-8ab5-1bcfea0ac894-kube-api-access-km2sr\") pod \"aa0000da-8f11-4e97-8ab5-1bcfea0ac894\" (UID: \"aa0000da-8f11-4e97-8ab5-1bcfea0ac894\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.798195 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fgncl\" (UniqueName: \"kubernetes.io/projected/cb216b07-9809-4b8b-857b-ac1192747b9c-kube-api-access-fgncl\") pod \"cb216b07-9809-4b8b-857b-ac1192747b9c\" (UID: \"cb216b07-9809-4b8b-857b-ac1192747b9c\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.801738 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa0000da-8f11-4e97-8ab5-1bcfea0ac894-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "aa0000da-8f11-4e97-8ab5-1bcfea0ac894" (UID: "aa0000da-8f11-4e97-8ab5-1bcfea0ac894"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.802239 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cb216b07-9809-4b8b-857b-ac1192747b9c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cb216b07-9809-4b8b-857b-ac1192747b9c" (UID: "cb216b07-9809-4b8b-857b-ac1192747b9c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.802671 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47870992-2db9-46f4-84d9-fd50fb9851eb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "47870992-2db9-46f4-84d9-fd50fb9851eb" (UID: "47870992-2db9-46f4-84d9-fd50fb9851eb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.805574 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb216b07-9809-4b8b-857b-ac1192747b9c-kube-api-access-fgncl" (OuterVolumeSpecName: "kube-api-access-fgncl") pod "cb216b07-9809-4b8b-857b-ac1192747b9c" (UID: "cb216b07-9809-4b8b-857b-ac1192747b9c"). InnerVolumeSpecName "kube-api-access-fgncl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.810438 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa0000da-8f11-4e97-8ab5-1bcfea0ac894-kube-api-access-km2sr" (OuterVolumeSpecName: "kube-api-access-km2sr") pod "aa0000da-8f11-4e97-8ab5-1bcfea0ac894" (UID: "aa0000da-8f11-4e97-8ab5-1bcfea0ac894"). InnerVolumeSpecName "kube-api-access-km2sr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.825194 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47870992-2db9-46f4-84d9-fd50fb9851eb-kube-api-access-64q7x" (OuterVolumeSpecName: "kube-api-access-64q7x") pod "47870992-2db9-46f4-84d9-fd50fb9851eb" (UID: "47870992-2db9-46f4-84d9-fd50fb9851eb"). InnerVolumeSpecName "kube-api-access-64q7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.899783 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zg7q\" (UniqueName: \"kubernetes.io/projected/ee94a050-f905-44f1-a5da-16536b8cdfa7-kube-api-access-7zg7q\") pod \"ee94a050-f905-44f1-a5da-16536b8cdfa7\" (UID: \"ee94a050-f905-44f1-a5da-16536b8cdfa7\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.899842 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-run-ovn\") pod \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.899881 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d045bc7-38b2-46f5-8cd8-cf634003bedf-operator-scripts\") pod \"0d045bc7-38b2-46f5-8cd8-cf634003bedf\" (UID: \"0d045bc7-38b2-46f5-8cd8-cf634003bedf\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.899923 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee94a050-f905-44f1-a5da-16536b8cdfa7-operator-scripts\") pod \"ee94a050-f905-44f1-a5da-16536b8cdfa7\" (UID: \"ee94a050-f905-44f1-a5da-16536b8cdfa7\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.899979 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-log-ovn\") pod \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.900037 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/67f877b5-12d3-4b48-a9eb-9ee2629e830a-additional-scripts\") pod \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.900073 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsxc4\" (UniqueName: \"kubernetes.io/projected/0d045bc7-38b2-46f5-8cd8-cf634003bedf-kube-api-access-fsxc4\") pod \"0d045bc7-38b2-46f5-8cd8-cf634003bedf\" (UID: \"0d045bc7-38b2-46f5-8cd8-cf634003bedf\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.900178 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67f877b5-12d3-4b48-a9eb-9ee2629e830a-scripts\") pod \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.900288 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-run\") pod \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.900322 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n4j57\" (UniqueName: \"kubernetes.io/projected/67f877b5-12d3-4b48-a9eb-9ee2629e830a-kube-api-access-n4j57\") pod \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\" (UID: \"67f877b5-12d3-4b48-a9eb-9ee2629e830a\") " Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.900388 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "67f877b5-12d3-4b48-a9eb-9ee2629e830a" (UID: "67f877b5-12d3-4b48-a9eb-9ee2629e830a"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.900749 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fgncl\" (UniqueName: \"kubernetes.io/projected/cb216b07-9809-4b8b-857b-ac1192747b9c-kube-api-access-fgncl\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.900771 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/aa0000da-8f11-4e97-8ab5-1bcfea0ac894-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.900784 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cb216b07-9809-4b8b-857b-ac1192747b9c-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.900797 4632 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.900809 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/47870992-2db9-46f4-84d9-fd50fb9851eb-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.900821 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64q7x\" (UniqueName: \"kubernetes.io/projected/47870992-2db9-46f4-84d9-fd50fb9851eb-kube-api-access-64q7x\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.900832 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-km2sr\" (UniqueName: \"kubernetes.io/projected/aa0000da-8f11-4e97-8ab5-1bcfea0ac894-kube-api-access-km2sr\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.900875 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee94a050-f905-44f1-a5da-16536b8cdfa7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ee94a050-f905-44f1-a5da-16536b8cdfa7" (UID: "ee94a050-f905-44f1-a5da-16536b8cdfa7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.901023 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "67f877b5-12d3-4b48-a9eb-9ee2629e830a" (UID: "67f877b5-12d3-4b48-a9eb-9ee2629e830a"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.901497 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-run" (OuterVolumeSpecName: "var-run") pod "67f877b5-12d3-4b48-a9eb-9ee2629e830a" (UID: "67f877b5-12d3-4b48-a9eb-9ee2629e830a"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.901685 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d045bc7-38b2-46f5-8cd8-cf634003bedf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d045bc7-38b2-46f5-8cd8-cf634003bedf" (UID: "0d045bc7-38b2-46f5-8cd8-cf634003bedf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.901730 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67f877b5-12d3-4b48-a9eb-9ee2629e830a-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "67f877b5-12d3-4b48-a9eb-9ee2629e830a" (UID: "67f877b5-12d3-4b48-a9eb-9ee2629e830a"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.902430 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67f877b5-12d3-4b48-a9eb-9ee2629e830a-scripts" (OuterVolumeSpecName: "scripts") pod "67f877b5-12d3-4b48-a9eb-9ee2629e830a" (UID: "67f877b5-12d3-4b48-a9eb-9ee2629e830a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.904581 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee94a050-f905-44f1-a5da-16536b8cdfa7-kube-api-access-7zg7q" (OuterVolumeSpecName: "kube-api-access-7zg7q") pod "ee94a050-f905-44f1-a5da-16536b8cdfa7" (UID: "ee94a050-f905-44f1-a5da-16536b8cdfa7"). InnerVolumeSpecName "kube-api-access-7zg7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.905425 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67f877b5-12d3-4b48-a9eb-9ee2629e830a-kube-api-access-n4j57" (OuterVolumeSpecName: "kube-api-access-n4j57") pod "67f877b5-12d3-4b48-a9eb-9ee2629e830a" (UID: "67f877b5-12d3-4b48-a9eb-9ee2629e830a"). InnerVolumeSpecName "kube-api-access-n4j57". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:54 crc kubenswrapper[4632]: I0313 10:24:54.908396 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d045bc7-38b2-46f5-8cd8-cf634003bedf-kube-api-access-fsxc4" (OuterVolumeSpecName: "kube-api-access-fsxc4") pod "0d045bc7-38b2-46f5-8cd8-cf634003bedf" (UID: "0d045bc7-38b2-46f5-8cd8-cf634003bedf"). InnerVolumeSpecName "kube-api-access-fsxc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.002972 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n4j57\" (UniqueName: \"kubernetes.io/projected/67f877b5-12d3-4b48-a9eb-9ee2629e830a-kube-api-access-n4j57\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.003333 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zg7q\" (UniqueName: \"kubernetes.io/projected/ee94a050-f905-44f1-a5da-16536b8cdfa7-kube-api-access-7zg7q\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.003353 4632 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.003368 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d045bc7-38b2-46f5-8cd8-cf634003bedf-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.003379 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ee94a050-f905-44f1-a5da-16536b8cdfa7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.003390 4632 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/67f877b5-12d3-4b48-a9eb-9ee2629e830a-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.003404 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsxc4\" (UniqueName: \"kubernetes.io/projected/0d045bc7-38b2-46f5-8cd8-cf634003bedf-kube-api-access-fsxc4\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.003417 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/67f877b5-12d3-4b48-a9eb-9ee2629e830a-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.003429 4632 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/67f877b5-12d3-4b48-a9eb-9ee2629e830a-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.127105 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-g7pfc" event={"ID":"aa0000da-8f11-4e97-8ab5-1bcfea0ac894","Type":"ContainerDied","Data":"64833dd5dcca1cf0d9b0e7fcfd2870e2d6c8b40bb49e5e7b154150c2bf852051"} Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.127146 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64833dd5dcca1cf0d9b0e7fcfd2870e2d6c8b40bb49e5e7b154150c2bf852051" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.127144 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-g7pfc" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.132191 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9kd7r-config-jtj4g" event={"ID":"67f877b5-12d3-4b48-a9eb-9ee2629e830a","Type":"ContainerDied","Data":"ce1ab79e8690eb9d1f16d7b7f5d9ff52729195fe7c60d80326b0842004a4a53d"} Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.132237 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ce1ab79e8690eb9d1f16d7b7f5d9ff52729195fe7c60d80326b0842004a4a53d" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.132318 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9kd7r-config-jtj4g" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.140243 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-4dec-account-create-update-hfnth" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.140253 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-4dec-account-create-update-hfnth" event={"ID":"47870992-2db9-46f4-84d9-fd50fb9851eb","Type":"ContainerDied","Data":"664b8a378c78fddcf14389393b1cc3a53fe85b08aab7f467156058884d9c4350"} Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.140277 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="664b8a378c78fddcf14389393b1cc3a53fe85b08aab7f467156058884d9c4350" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.142296 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-b742-account-create-update-gfdkg" event={"ID":"ee94a050-f905-44f1-a5da-16536b8cdfa7","Type":"ContainerDied","Data":"c71c108608caeb76931caff20f7b0c7e5d8d5c389c0440d31bb543598d28dfb8"} Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.142346 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c71c108608caeb76931caff20f7b0c7e5d8d5c389c0440d31bb543598d28dfb8" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.142469 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-b742-account-create-update-gfdkg" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.144917 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-da66-account-create-update-tk8pd" event={"ID":"0d045bc7-38b2-46f5-8cd8-cf634003bedf","Type":"ContainerDied","Data":"6cabee2602a6d9e4308c8db70b1d7f8643862ae4eef1ae7803777760563d87cb"} Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.144976 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6cabee2602a6d9e4308c8db70b1d7f8643862ae4eef1ae7803777760563d87cb" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.145143 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-da66-account-create-update-tk8pd" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.156014 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"5fbf766730c5e0b50f415ecb57ded1219f87c9fea289a1de721175fa49897a02"} Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.156068 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"350411d685010b4776bb5fa669ae387802d1836dc40c6f341061fd360acd8211"} Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.163060 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mq9np" event={"ID":"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419","Type":"ContainerStarted","Data":"53c212eae0f18baff6fdcd0d88db82f3271a3997b68292e7fdae508ea7808719"} Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.167994 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pnvjb" event={"ID":"cb216b07-9809-4b8b-857b-ac1192747b9c","Type":"ContainerDied","Data":"417ef1960b2d4ea70aed07efa739673778171eebc874e98d5bdf429380cac86f"} Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.168186 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="417ef1960b2d4ea70aed07efa739673778171eebc874e98d5bdf429380cac86f" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.168270 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pnvjb" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.194544 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-mq9np" podStartSLOduration=3.516220301 podStartE2EDuration="12.19441146s" podCreationTimestamp="2026-03-13 10:24:43 +0000 UTC" firstStartedPulling="2026-03-13 10:24:45.790838775 +0000 UTC m=+1259.813368908" lastFinishedPulling="2026-03-13 10:24:54.469029934 +0000 UTC m=+1268.491560067" observedRunningTime="2026-03-13 10:24:55.184359924 +0000 UTC m=+1269.206890057" watchObservedRunningTime="2026-03-13 10:24:55.19441146 +0000 UTC m=+1269.216941593" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.826091 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-9kd7r-config-jtj4g"] Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.833933 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-9kd7r-config-jtj4g"] Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.950659 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-9kd7r-config-hlhk2"] Mar 13 10:24:55 crc kubenswrapper[4632]: E0313 10:24:55.954109 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcdcfad1-d735-4b55-ae65-0ce16bdbc79d" containerName="mariadb-database-create" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954144 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcdcfad1-d735-4b55-ae65-0ce16bdbc79d" containerName="mariadb-database-create" Mar 13 10:24:55 crc kubenswrapper[4632]: E0313 10:24:55.954166 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a03b92ea-cd2c-455d-a88e-1d57b958b138" containerName="mariadb-account-create-update" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954174 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a03b92ea-cd2c-455d-a88e-1d57b958b138" containerName="mariadb-account-create-update" Mar 13 10:24:55 crc kubenswrapper[4632]: E0313 10:24:55.954192 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d045bc7-38b2-46f5-8cd8-cf634003bedf" containerName="mariadb-account-create-update" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954198 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d045bc7-38b2-46f5-8cd8-cf634003bedf" containerName="mariadb-account-create-update" Mar 13 10:24:55 crc kubenswrapper[4632]: E0313 10:24:55.954210 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67f877b5-12d3-4b48-a9eb-9ee2629e830a" containerName="ovn-config" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954216 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="67f877b5-12d3-4b48-a9eb-9ee2629e830a" containerName="ovn-config" Mar 13 10:24:55 crc kubenswrapper[4632]: E0313 10:24:55.954226 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa0000da-8f11-4e97-8ab5-1bcfea0ac894" containerName="mariadb-database-create" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954232 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa0000da-8f11-4e97-8ab5-1bcfea0ac894" containerName="mariadb-database-create" Mar 13 10:24:55 crc kubenswrapper[4632]: E0313 10:24:55.954245 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee94a050-f905-44f1-a5da-16536b8cdfa7" containerName="mariadb-account-create-update" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954253 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee94a050-f905-44f1-a5da-16536b8cdfa7" containerName="mariadb-account-create-update" Mar 13 10:24:55 crc kubenswrapper[4632]: E0313 10:24:55.954259 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb216b07-9809-4b8b-857b-ac1192747b9c" containerName="mariadb-database-create" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954265 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb216b07-9809-4b8b-857b-ac1192747b9c" containerName="mariadb-database-create" Mar 13 10:24:55 crc kubenswrapper[4632]: E0313 10:24:55.954278 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239c554e-360d-4f04-86f0-b2b98974bad3" containerName="mariadb-database-create" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954283 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="239c554e-360d-4f04-86f0-b2b98974bad3" containerName="mariadb-database-create" Mar 13 10:24:55 crc kubenswrapper[4632]: E0313 10:24:55.954295 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47870992-2db9-46f4-84d9-fd50fb9851eb" containerName="mariadb-account-create-update" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954303 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="47870992-2db9-46f4-84d9-fd50fb9851eb" containerName="mariadb-account-create-update" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954517 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa0000da-8f11-4e97-8ab5-1bcfea0ac894" containerName="mariadb-database-create" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954530 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="a03b92ea-cd2c-455d-a88e-1d57b958b138" containerName="mariadb-account-create-update" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954542 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee94a050-f905-44f1-a5da-16536b8cdfa7" containerName="mariadb-account-create-update" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954551 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb216b07-9809-4b8b-857b-ac1192747b9c" containerName="mariadb-database-create" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954562 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d045bc7-38b2-46f5-8cd8-cf634003bedf" containerName="mariadb-account-create-update" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954570 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="67f877b5-12d3-4b48-a9eb-9ee2629e830a" containerName="ovn-config" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954580 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="239c554e-360d-4f04-86f0-b2b98974bad3" containerName="mariadb-database-create" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954588 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcdcfad1-d735-4b55-ae65-0ce16bdbc79d" containerName="mariadb-database-create" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.954600 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="47870992-2db9-46f4-84d9-fd50fb9851eb" containerName="mariadb-account-create-update" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.955221 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.970531 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9kd7r-config-hlhk2"] Mar 13 10:24:55 crc kubenswrapper[4632]: I0313 10:24:55.970774 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.070138 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67f877b5-12d3-4b48-a9eb-9ee2629e830a" path="/var/lib/kubelet/pods/67f877b5-12d3-4b48-a9eb-9ee2629e830a/volumes" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.127688 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-run-ovn\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.127762 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8874c236-ccb3-45c3-9838-42542e1483fb-scripts\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.127849 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-run\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.127878 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-log-ovn\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.127901 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mklx\" (UniqueName: \"kubernetes.io/projected/8874c236-ccb3-45c3-9838-42542e1483fb-kube-api-access-2mklx\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.127982 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8874c236-ccb3-45c3-9838-42542e1483fb-additional-scripts\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.229784 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8874c236-ccb3-45c3-9838-42542e1483fb-additional-scripts\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.229864 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-run-ovn\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.229917 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8874c236-ccb3-45c3-9838-42542e1483fb-scripts\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.230012 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-run\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.230034 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-log-ovn\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.230053 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mklx\" (UniqueName: \"kubernetes.io/projected/8874c236-ccb3-45c3-9838-42542e1483fb-kube-api-access-2mklx\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.230186 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-run-ovn\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.230255 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-run\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.230587 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8874c236-ccb3-45c3-9838-42542e1483fb-additional-scripts\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.230661 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-log-ovn\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.232082 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8874c236-ccb3-45c3-9838-42542e1483fb-scripts\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.251733 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mklx\" (UniqueName: \"kubernetes.io/projected/8874c236-ccb3-45c3-9838-42542e1483fb-kube-api-access-2mklx\") pod \"ovn-controller-9kd7r-config-hlhk2\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.327858 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:56 crc kubenswrapper[4632]: I0313 10:24:56.973737 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-9kd7r-config-hlhk2"] Mar 13 10:24:57 crc kubenswrapper[4632]: I0313 10:24:57.194358 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9kd7r-config-hlhk2" event={"ID":"8874c236-ccb3-45c3-9838-42542e1483fb","Type":"ContainerStarted","Data":"de5e563237b6f5d5a900bfa4db200770d531a773a59fc045ca36a99783ac23c3"} Mar 13 10:24:57 crc kubenswrapper[4632]: I0313 10:24:57.200211 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"6410c58da9c7335dcc151a74496460faab6690d4d868622c9f53e1ac95af015d"} Mar 13 10:24:58 crc kubenswrapper[4632]: I0313 10:24:58.212891 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"2ad862ed96fbf3f56e9fff7d04adafeceb7e022528ca4c0cfd19fd10939e9b7a"} Mar 13 10:24:58 crc kubenswrapper[4632]: I0313 10:24:58.214019 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"deed0762a2707d9fc7e47fb11fd2a976ec5f0a219b2515f95c27456e30196111"} Mar 13 10:24:58 crc kubenswrapper[4632]: I0313 10:24:58.214099 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"75f72d4258d1b9a4742e65c0ab7da327b79d9d4d43591f14e17852aa410cacc0"} Mar 13 10:24:58 crc kubenswrapper[4632]: I0313 10:24:58.215386 4632 generic.go:334] "Generic (PLEG): container finished" podID="8874c236-ccb3-45c3-9838-42542e1483fb" containerID="98a44d8e524895de3db65a2da91c25a6875681d7e31dfa6eb205635df601d593" exitCode=0 Mar 13 10:24:58 crc kubenswrapper[4632]: I0313 10:24:58.215468 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9kd7r-config-hlhk2" event={"ID":"8874c236-ccb3-45c3-9838-42542e1483fb","Type":"ContainerDied","Data":"98a44d8e524895de3db65a2da91c25a6875681d7e31dfa6eb205635df601d593"} Mar 13 10:24:58 crc kubenswrapper[4632]: I0313 10:24:58.931230 4632 scope.go:117] "RemoveContainer" containerID="971cfa2ec11ce234b8c8c574daddb17b130773fddba410f62dd84c800e0f4023" Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.239358 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"c974bded2de6ee2b905b68670d69ac1e80116509a459a79b905c7f37bdacff58"} Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.626635 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.801915 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-log-ovn\") pod \"8874c236-ccb3-45c3-9838-42542e1483fb\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.802304 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-run\") pod \"8874c236-ccb3-45c3-9838-42542e1483fb\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.802354 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8874c236-ccb3-45c3-9838-42542e1483fb-additional-scripts\") pod \"8874c236-ccb3-45c3-9838-42542e1483fb\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.802373 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8874c236-ccb3-45c3-9838-42542e1483fb-scripts\") pod \"8874c236-ccb3-45c3-9838-42542e1483fb\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.802056 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "8874c236-ccb3-45c3-9838-42542e1483fb" (UID: "8874c236-ccb3-45c3-9838-42542e1483fb"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.802329 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-run" (OuterVolumeSpecName: "var-run") pod "8874c236-ccb3-45c3-9838-42542e1483fb" (UID: "8874c236-ccb3-45c3-9838-42542e1483fb"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.802441 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2mklx\" (UniqueName: \"kubernetes.io/projected/8874c236-ccb3-45c3-9838-42542e1483fb-kube-api-access-2mklx\") pod \"8874c236-ccb3-45c3-9838-42542e1483fb\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.802471 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-run-ovn\") pod \"8874c236-ccb3-45c3-9838-42542e1483fb\" (UID: \"8874c236-ccb3-45c3-9838-42542e1483fb\") " Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.802767 4632 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.802780 4632 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.802816 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "8874c236-ccb3-45c3-9838-42542e1483fb" (UID: "8874c236-ccb3-45c3-9838-42542e1483fb"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.803663 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8874c236-ccb3-45c3-9838-42542e1483fb-scripts" (OuterVolumeSpecName: "scripts") pod "8874c236-ccb3-45c3-9838-42542e1483fb" (UID: "8874c236-ccb3-45c3-9838-42542e1483fb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.803968 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8874c236-ccb3-45c3-9838-42542e1483fb-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "8874c236-ccb3-45c3-9838-42542e1483fb" (UID: "8874c236-ccb3-45c3-9838-42542e1483fb"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.811624 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8874c236-ccb3-45c3-9838-42542e1483fb-kube-api-access-2mklx" (OuterVolumeSpecName: "kube-api-access-2mklx") pod "8874c236-ccb3-45c3-9838-42542e1483fb" (UID: "8874c236-ccb3-45c3-9838-42542e1483fb"). InnerVolumeSpecName "kube-api-access-2mklx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.904251 4632 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/8874c236-ccb3-45c3-9838-42542e1483fb-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.904298 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8874c236-ccb3-45c3-9838-42542e1483fb-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.904312 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2mklx\" (UniqueName: \"kubernetes.io/projected/8874c236-ccb3-45c3-9838-42542e1483fb-kube-api-access-2mklx\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:59 crc kubenswrapper[4632]: I0313 10:24:59.904327 4632 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/8874c236-ccb3-45c3-9838-42542e1483fb-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:00 crc kubenswrapper[4632]: I0313 10:25:00.261933 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l6hpb" event={"ID":"4f1c5663-463b-45e2-b200-64e73e6d5698","Type":"ContainerStarted","Data":"bf8d93edd68f1cf79021467ff9910419baf75397a4140fb3d25bca7f97abbf70"} Mar 13 10:25:00 crc kubenswrapper[4632]: I0313 10:25:00.266797 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-9kd7r-config-hlhk2" Mar 13 10:25:00 crc kubenswrapper[4632]: I0313 10:25:00.267083 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-9kd7r-config-hlhk2" event={"ID":"8874c236-ccb3-45c3-9838-42542e1483fb","Type":"ContainerDied","Data":"de5e563237b6f5d5a900bfa4db200770d531a773a59fc045ca36a99783ac23c3"} Mar 13 10:25:00 crc kubenswrapper[4632]: I0313 10:25:00.267112 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de5e563237b6f5d5a900bfa4db200770d531a773a59fc045ca36a99783ac23c3" Mar 13 10:25:00 crc kubenswrapper[4632]: I0313 10:25:00.297767 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"992aa88321f15a34fdab3b0ce20b6c83dd648c4655d479b207bb5da20558c851"} Mar 13 10:25:00 crc kubenswrapper[4632]: I0313 10:25:00.297829 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"24f32d07c7b51462bf0a0a775221d99b9ece41557d0064e4a499eefe18ba7dfc"} Mar 13 10:25:00 crc kubenswrapper[4632]: I0313 10:25:00.297848 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"d5e01d4eefcd8e0fb049c2a6a11ed0324853c68edd619187e50b5676b13f1103"} Mar 13 10:25:00 crc kubenswrapper[4632]: I0313 10:25:00.299995 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-l6hpb" podStartSLOduration=2.853085272 podStartE2EDuration="36.299971059s" podCreationTimestamp="2026-03-13 10:24:24 +0000 UTC" firstStartedPulling="2026-03-13 10:24:25.287597404 +0000 UTC m=+1239.310127527" lastFinishedPulling="2026-03-13 10:24:58.734483181 +0000 UTC m=+1272.757013314" observedRunningTime="2026-03-13 10:25:00.279911376 +0000 UTC m=+1274.302441519" watchObservedRunningTime="2026-03-13 10:25:00.299971059 +0000 UTC m=+1274.322501192" Mar 13 10:25:00 crc kubenswrapper[4632]: I0313 10:25:00.739043 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-9kd7r-config-hlhk2"] Mar 13 10:25:00 crc kubenswrapper[4632]: I0313 10:25:00.773456 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-9kd7r-config-hlhk2"] Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.313511 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"0af8435ee83cb5cebf4c2cf61b496b0ccc8ad30e3126d780d20bab35af32a80b"} Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.313560 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"4bed88103529c366d530f8c0b3586e87d09239d52f2dc963ad771a5ba7b1873e"} Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.313572 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"e37b3d77-de2e-4be9-9984-550d4ba0f2f0","Type":"ContainerStarted","Data":"de216b68315e4d4fe5c2c01408fd20a4ef13238294cf3f3a9267a8193c8962b4"} Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.315323 4632 generic.go:334] "Generic (PLEG): container finished" podID="e824ae7d-dbbd-496b-b8b0-8b5c59a4d419" containerID="53c212eae0f18baff6fdcd0d88db82f3271a3997b68292e7fdae508ea7808719" exitCode=0 Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.315358 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mq9np" event={"ID":"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419","Type":"ContainerDied","Data":"53c212eae0f18baff6fdcd0d88db82f3271a3997b68292e7fdae508ea7808719"} Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.357566 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=43.438449855 podStartE2EDuration="55.357543615s" podCreationTimestamp="2026-03-13 10:24:06 +0000 UTC" firstStartedPulling="2026-03-13 10:24:46.813789012 +0000 UTC m=+1260.836319145" lastFinishedPulling="2026-03-13 10:24:58.732882772 +0000 UTC m=+1272.755412905" observedRunningTime="2026-03-13 10:25:01.353407143 +0000 UTC m=+1275.375937276" watchObservedRunningTime="2026-03-13 10:25:01.357543615 +0000 UTC m=+1275.380073758" Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.784666 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-756d4f5c49-ng8tb"] Mar 13 10:25:01 crc kubenswrapper[4632]: E0313 10:25:01.785625 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8874c236-ccb3-45c3-9838-42542e1483fb" containerName="ovn-config" Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.785750 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8874c236-ccb3-45c3-9838-42542e1483fb" containerName="ovn-config" Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.786061 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8874c236-ccb3-45c3-9838-42542e1483fb" containerName="ovn-config" Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.787279 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.790292 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.796398 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-756d4f5c49-ng8tb"] Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.946013 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-ovsdbserver-nb\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.946075 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-dns-svc\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.946110 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-ovsdbserver-sb\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.946351 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-config\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.946532 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5jtw\" (UniqueName: \"kubernetes.io/projected/875ab9dc-abac-45c4-86b9-b0bfccdfb240-kube-api-access-t5jtw\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:01 crc kubenswrapper[4632]: I0313 10:25:01.946836 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-dns-swift-storage-0\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.048113 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-dns-swift-storage-0\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.048415 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-ovsdbserver-nb\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.048495 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-dns-svc\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.048565 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-ovsdbserver-sb\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.048658 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-config\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.048748 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5jtw\" (UniqueName: \"kubernetes.io/projected/875ab9dc-abac-45c4-86b9-b0bfccdfb240-kube-api-access-t5jtw\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.049137 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-dns-swift-storage-0\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.049221 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-dns-svc\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.049803 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-ovsdbserver-sb\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.049817 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-config\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.050341 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-ovsdbserver-nb\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.055984 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8874c236-ccb3-45c3-9838-42542e1483fb" path="/var/lib/kubelet/pods/8874c236-ccb3-45c3-9838-42542e1483fb/volumes" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.079756 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5jtw\" (UniqueName: \"kubernetes.io/projected/875ab9dc-abac-45c4-86b9-b0bfccdfb240-kube-api-access-t5jtw\") pod \"dnsmasq-dns-756d4f5c49-ng8tb\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.103053 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.569329 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-756d4f5c49-ng8tb"] Mar 13 10:25:02 crc kubenswrapper[4632]: W0313 10:25:02.588716 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod875ab9dc_abac_45c4_86b9_b0bfccdfb240.slice/crio-0260417d40f7e52a2270d809525583f83a0891dc6b82c6c6e13d0d522dd4a9b4 WatchSource:0}: Error finding container 0260417d40f7e52a2270d809525583f83a0891dc6b82c6c6e13d0d522dd4a9b4: Status 404 returned error can't find the container with id 0260417d40f7e52a2270d809525583f83a0891dc6b82c6c6e13d0d522dd4a9b4 Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.686899 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mq9np" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.760506 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-config-data\") pod \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\" (UID: \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\") " Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.760560 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5hh8q\" (UniqueName: \"kubernetes.io/projected/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-kube-api-access-5hh8q\") pod \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\" (UID: \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\") " Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.760631 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-combined-ca-bundle\") pod \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\" (UID: \"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419\") " Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.765192 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-kube-api-access-5hh8q" (OuterVolumeSpecName: "kube-api-access-5hh8q") pod "e824ae7d-dbbd-496b-b8b0-8b5c59a4d419" (UID: "e824ae7d-dbbd-496b-b8b0-8b5c59a4d419"). InnerVolumeSpecName "kube-api-access-5hh8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.802880 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e824ae7d-dbbd-496b-b8b0-8b5c59a4d419" (UID: "e824ae7d-dbbd-496b-b8b0-8b5c59a4d419"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.829835 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-config-data" (OuterVolumeSpecName: "config-data") pod "e824ae7d-dbbd-496b-b8b0-8b5c59a4d419" (UID: "e824ae7d-dbbd-496b-b8b0-8b5c59a4d419"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.862189 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.862224 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5hh8q\" (UniqueName: \"kubernetes.io/projected/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-kube-api-access-5hh8q\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:02 crc kubenswrapper[4632]: I0313 10:25:02.862234 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.338111 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-mq9np" event={"ID":"e824ae7d-dbbd-496b-b8b0-8b5c59a4d419","Type":"ContainerDied","Data":"1a3020d6e5b66dad152669406220e67cb7be099d82ff8fd4925d6504c1176fb1"} Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.338340 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a3020d6e5b66dad152669406220e67cb7be099d82ff8fd4925d6504c1176fb1" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.339069 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-mq9np" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.339739 4632 generic.go:334] "Generic (PLEG): container finished" podID="875ab9dc-abac-45c4-86b9-b0bfccdfb240" containerID="6cc0261f52d1f9ff6e82214738332406f7144515ee6e99809f0b2f51974a5801" exitCode=0 Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.339797 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" event={"ID":"875ab9dc-abac-45c4-86b9-b0bfccdfb240","Type":"ContainerDied","Data":"6cc0261f52d1f9ff6e82214738332406f7144515ee6e99809f0b2f51974a5801"} Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.339823 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" event={"ID":"875ab9dc-abac-45c4-86b9-b0bfccdfb240","Type":"ContainerStarted","Data":"0260417d40f7e52a2270d809525583f83a0891dc6b82c6c6e13d0d522dd4a9b4"} Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.811048 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-756d4f5c49-ng8tb"] Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.821277 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-6wd56"] Mar 13 10:25:03 crc kubenswrapper[4632]: E0313 10:25:03.821746 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e824ae7d-dbbd-496b-b8b0-8b5c59a4d419" containerName="keystone-db-sync" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.821762 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e824ae7d-dbbd-496b-b8b0-8b5c59a4d419" containerName="keystone-db-sync" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.821968 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="e824ae7d-dbbd-496b-b8b0-8b5c59a4d419" containerName="keystone-db-sync" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.822553 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.828894 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6wd56"] Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.838767 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.839113 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.839302 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-llpcf" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.839464 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.851803 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.883449 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-combined-ca-bundle\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.883572 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-scripts\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.883601 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-config-data\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.883635 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-credential-keys\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.883657 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-fernet-keys\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.883677 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9qt5\" (UniqueName: \"kubernetes.io/projected/6dd01e75-b01a-439d-953a-a7b35aefaccf-kube-api-access-q9qt5\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.949871 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6844bbffb5-6qbh8"] Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.951838 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.985029 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-credential-keys\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.985089 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-fernet-keys\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.985116 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9qt5\" (UniqueName: \"kubernetes.io/projected/6dd01e75-b01a-439d-953a-a7b35aefaccf-kube-api-access-q9qt5\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.985179 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-combined-ca-bundle\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.985295 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-scripts\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:03 crc kubenswrapper[4632]: I0313 10:25:03.985335 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-config-data\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.021630 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-credential-keys\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.025026 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-scripts\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.025740 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-fernet-keys\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.026662 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-config-data\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.052234 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-combined-ca-bundle\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.083703 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6844bbffb5-6qbh8"] Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.087277 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-dns-swift-storage-0\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.087339 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-config\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.087399 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-dns-svc\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.087476 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-ovsdbserver-nb\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.087501 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slg5l\" (UniqueName: \"kubernetes.io/projected/8e8562e5-7677-460c-864c-c0f1dcd2ac41-kube-api-access-slg5l\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.087528 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-ovsdbserver-sb\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.091417 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9qt5\" (UniqueName: \"kubernetes.io/projected/6dd01e75-b01a-439d-953a-a7b35aefaccf-kube-api-access-q9qt5\") pod \"keystone-bootstrap-6wd56\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.158241 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.189111 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-dns-svc\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.189184 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-ovsdbserver-nb\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.189205 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slg5l\" (UniqueName: \"kubernetes.io/projected/8e8562e5-7677-460c-864c-c0f1dcd2ac41-kube-api-access-slg5l\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.189239 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-ovsdbserver-sb\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.189279 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-dns-swift-storage-0\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.189321 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-config\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.191749 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-dns-svc\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.192391 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-ovsdbserver-nb\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.193478 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-ovsdbserver-sb\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.194107 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-config\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.194726 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-dns-swift-storage-0\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.206924 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-7fvlk"] Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.207931 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-7fvlk" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.223665 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-vbbdq" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.223998 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.265776 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slg5l\" (UniqueName: \"kubernetes.io/projected/8e8562e5-7677-460c-864c-c0f1dcd2ac41-kube-api-access-slg5l\") pod \"dnsmasq-dns-6844bbffb5-6qbh8\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.274053 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-7fvlk"] Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.290362 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-combined-ca-bundle\") pod \"heat-db-sync-7fvlk\" (UID: \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\") " pod="openstack/heat-db-sync-7fvlk" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.290420 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-config-data\") pod \"heat-db-sync-7fvlk\" (UID: \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\") " pod="openstack/heat-db-sync-7fvlk" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.290465 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5rzh\" (UniqueName: \"kubernetes.io/projected/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-kube-api-access-n5rzh\") pod \"heat-db-sync-7fvlk\" (UID: \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\") " pod="openstack/heat-db-sync-7fvlk" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.295398 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.391503 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" event={"ID":"875ab9dc-abac-45c4-86b9-b0bfccdfb240","Type":"ContainerStarted","Data":"d5fd04632add169f284e136d8cc0cc1ed2dafea7a7d420e5a55cd81556345dd4"} Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.391637 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-combined-ca-bundle\") pod \"heat-db-sync-7fvlk\" (UID: \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\") " pod="openstack/heat-db-sync-7fvlk" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.391679 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-config-data\") pod \"heat-db-sync-7fvlk\" (UID: \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\") " pod="openstack/heat-db-sync-7fvlk" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.391688 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" podUID="875ab9dc-abac-45c4-86b9-b0bfccdfb240" containerName="dnsmasq-dns" containerID="cri-o://d5fd04632add169f284e136d8cc0cc1ed2dafea7a7d420e5a55cd81556345dd4" gracePeriod=10 Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.391724 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5rzh\" (UniqueName: \"kubernetes.io/projected/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-kube-api-access-n5rzh\") pod \"heat-db-sync-7fvlk\" (UID: \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\") " pod="openstack/heat-db-sync-7fvlk" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.392329 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.396787 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-combined-ca-bundle\") pod \"heat-db-sync-7fvlk\" (UID: \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\") " pod="openstack/heat-db-sync-7fvlk" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.405840 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-config-data\") pod \"heat-db-sync-7fvlk\" (UID: \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\") " pod="openstack/heat-db-sync-7fvlk" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.513423 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.516149 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.551402 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.551584 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.597161 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/270ebc10-986f-4473-8a5e-9094de34ae98-log-httpd\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.597242 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.597270 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-config-data\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.597335 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/270ebc10-986f-4473-8a5e-9094de34ae98-run-httpd\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.597467 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.597504 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-scripts\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.597618 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jz5nk\" (UniqueName: \"kubernetes.io/projected/270ebc10-986f-4473-8a5e-9094de34ae98-kube-api-access-jz5nk\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.615466 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-hlsnz"] Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.616840 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hlsnz" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.628633 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.628871 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-r2t7p" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.629631 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.643640 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.701078 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7221b50-7231-4ade-917e-b10f177cb539-config\") pod \"neutron-db-sync-hlsnz\" (UID: \"b7221b50-7231-4ade-917e-b10f177cb539\") " pod="openstack/neutron-db-sync-hlsnz" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.701136 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jz5nk\" (UniqueName: \"kubernetes.io/projected/270ebc10-986f-4473-8a5e-9094de34ae98-kube-api-access-jz5nk\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.701174 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7221b50-7231-4ade-917e-b10f177cb539-combined-ca-bundle\") pod \"neutron-db-sync-hlsnz\" (UID: \"b7221b50-7231-4ade-917e-b10f177cb539\") " pod="openstack/neutron-db-sync-hlsnz" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.701209 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/270ebc10-986f-4473-8a5e-9094de34ae98-log-httpd\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.701229 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.701249 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngb4t\" (UniqueName: \"kubernetes.io/projected/b7221b50-7231-4ade-917e-b10f177cb539-kube-api-access-ngb4t\") pod \"neutron-db-sync-hlsnz\" (UID: \"b7221b50-7231-4ade-917e-b10f177cb539\") " pod="openstack/neutron-db-sync-hlsnz" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.701269 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-config-data\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.701298 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/270ebc10-986f-4473-8a5e-9094de34ae98-run-httpd\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.701364 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.701391 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-scripts\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.702776 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/270ebc10-986f-4473-8a5e-9094de34ae98-run-httpd\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.703302 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/270ebc10-986f-4473-8a5e-9094de34ae98-log-httpd\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.706115 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" podStartSLOduration=3.706095517 podStartE2EDuration="3.706095517s" podCreationTimestamp="2026-03-13 10:25:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:25:04.696166944 +0000 UTC m=+1278.718697087" watchObservedRunningTime="2026-03-13 10:25:04.706095517 +0000 UTC m=+1278.728625660" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.728273 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-config-data\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.728910 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.729646 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-scripts\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.754040 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.809821 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-hlsnz"] Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.810742 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7221b50-7231-4ade-917e-b10f177cb539-combined-ca-bundle\") pod \"neutron-db-sync-hlsnz\" (UID: \"b7221b50-7231-4ade-917e-b10f177cb539\") " pod="openstack/neutron-db-sync-hlsnz" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.810781 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngb4t\" (UniqueName: \"kubernetes.io/projected/b7221b50-7231-4ade-917e-b10f177cb539-kube-api-access-ngb4t\") pod \"neutron-db-sync-hlsnz\" (UID: \"b7221b50-7231-4ade-917e-b10f177cb539\") " pod="openstack/neutron-db-sync-hlsnz" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.810903 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7221b50-7231-4ade-917e-b10f177cb539-config\") pod \"neutron-db-sync-hlsnz\" (UID: \"b7221b50-7231-4ade-917e-b10f177cb539\") " pod="openstack/neutron-db-sync-hlsnz" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.815265 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jz5nk\" (UniqueName: \"kubernetes.io/projected/270ebc10-986f-4473-8a5e-9094de34ae98-kube-api-access-jz5nk\") pod \"ceilometer-0\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.849881 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="1761ca69-46fd-4375-af60-22b3e77c19a2" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.867965 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7221b50-7231-4ade-917e-b10f177cb539-config\") pod \"neutron-db-sync-hlsnz\" (UID: \"b7221b50-7231-4ade-917e-b10f177cb539\") " pod="openstack/neutron-db-sync-hlsnz" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.868465 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7221b50-7231-4ade-917e-b10f177cb539-combined-ca-bundle\") pod \"neutron-db-sync-hlsnz\" (UID: \"b7221b50-7231-4ade-917e-b10f177cb539\") " pod="openstack/neutron-db-sync-hlsnz" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.869235 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.901194 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6fb489b64f-prckv"] Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.913017 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.921164 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngb4t\" (UniqueName: \"kubernetes.io/projected/b7221b50-7231-4ade-917e-b10f177cb539-kube-api-access-ngb4t\") pod \"neutron-db-sync-hlsnz\" (UID: \"b7221b50-7231-4ade-917e-b10f177cb539\") " pod="openstack/neutron-db-sync-hlsnz" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.928303 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.944901 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.945118 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-59mgb" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.945220 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.948341 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6fb489b64f-prckv"] Mar 13 10:25:04 crc kubenswrapper[4632]: I0313 10:25:04.969474 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hlsnz" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.013442 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00c4bdc6-a22c-4ab6-b898-cf591b92756b-scripts\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.013726 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00c4bdc6-a22c-4ab6-b898-cf591b92756b-config-data\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.013837 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00c4bdc6-a22c-4ab6-b898-cf591b92756b-horizon-secret-key\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.013922 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00c4bdc6-a22c-4ab6-b898-cf591b92756b-logs\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.014553 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gwx9r\" (UniqueName: \"kubernetes.io/projected/00c4bdc6-a22c-4ab6-b898-cf591b92756b-kube-api-access-gwx9r\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.025862 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-kq8lc"] Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.027033 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.046325 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-j7c52" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.046559 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.046699 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.119509 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-combined-ca-bundle\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.119574 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00c4bdc6-a22c-4ab6-b898-cf591b92756b-config-data\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.119599 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-scripts\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.119640 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-db-sync-config-data\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.119676 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00c4bdc6-a22c-4ab6-b898-cf591b92756b-horizon-secret-key\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.119702 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00c4bdc6-a22c-4ab6-b898-cf591b92756b-logs\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.119743 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m5bn\" (UniqueName: \"kubernetes.io/projected/8f916c05-f172-42b6-9b13-0c8d2058bfb1-kube-api-access-5m5bn\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.119785 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gwx9r\" (UniqueName: \"kubernetes.io/projected/00c4bdc6-a22c-4ab6-b898-cf591b92756b-kube-api-access-gwx9r\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.119829 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-config-data\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.119895 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f916c05-f172-42b6-9b13-0c8d2058bfb1-etc-machine-id\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.119999 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00c4bdc6-a22c-4ab6-b898-cf591b92756b-scripts\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.120783 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00c4bdc6-a22c-4ab6-b898-cf591b92756b-scripts\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.129865 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00c4bdc6-a22c-4ab6-b898-cf591b92756b-config-data\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.130192 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00c4bdc6-a22c-4ab6-b898-cf591b92756b-logs\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.211022 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-kq8lc"] Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.223041 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f916c05-f172-42b6-9b13-0c8d2058bfb1-etc-machine-id\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.223175 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-combined-ca-bundle\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.223217 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-scripts\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.223266 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-db-sync-config-data\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.223338 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m5bn\" (UniqueName: \"kubernetes.io/projected/8f916c05-f172-42b6-9b13-0c8d2058bfb1-kube-api-access-5m5bn\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.223425 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-config-data\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.233794 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-scripts\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.233886 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f916c05-f172-42b6-9b13-0c8d2058bfb1-etc-machine-id\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.239709 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-combined-ca-bundle\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.245443 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gwx9r\" (UniqueName: \"kubernetes.io/projected/00c4bdc6-a22c-4ab6-b898-cf591b92756b-kube-api-access-gwx9r\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.246175 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-db-sync-config-data\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.246199 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-config-data\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.258413 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00c4bdc6-a22c-4ab6-b898-cf591b92756b-horizon-secret-key\") pod \"horizon-6fb489b64f-prckv\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.296748 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.389722 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m5bn\" (UniqueName: \"kubernetes.io/projected/8f916c05-f172-42b6-9b13-0c8d2058bfb1-kube-api-access-5m5bn\") pod \"cinder-db-sync-kq8lc\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.432346 4632 generic.go:334] "Generic (PLEG): container finished" podID="875ab9dc-abac-45c4-86b9-b0bfccdfb240" containerID="d5fd04632add169f284e136d8cc0cc1ed2dafea7a7d420e5a55cd81556345dd4" exitCode=0 Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.432395 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" event={"ID":"875ab9dc-abac-45c4-86b9-b0bfccdfb240","Type":"ContainerDied","Data":"d5fd04632add169f284e136d8cc0cc1ed2dafea7a7d420e5a55cd81556345dd4"} Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.452212 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-zdgpw"] Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.453978 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zdgpw" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.467726 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.467917 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-m45mn" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.483099 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-zdgpw"] Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.542124 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-htnd9"] Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.543431 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.545171 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/418cb883-abd1-46b4-957f-0a40f3e62297-db-sync-config-data\") pod \"barbican-db-sync-zdgpw\" (UID: \"418cb883-abd1-46b4-957f-0a40f3e62297\") " pod="openstack/barbican-db-sync-zdgpw" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.545298 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgn44\" (UniqueName: \"kubernetes.io/projected/418cb883-abd1-46b4-957f-0a40f3e62297-kube-api-access-zgn44\") pod \"barbican-db-sync-zdgpw\" (UID: \"418cb883-abd1-46b4-957f-0a40f3e62297\") " pod="openstack/barbican-db-sync-zdgpw" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.545334 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418cb883-abd1-46b4-957f-0a40f3e62297-combined-ca-bundle\") pod \"barbican-db-sync-zdgpw\" (UID: \"418cb883-abd1-46b4-957f-0a40f3e62297\") " pod="openstack/barbican-db-sync-zdgpw" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.570615 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.574181 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-6tvl4" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.587501 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.620753 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5rzh\" (UniqueName: \"kubernetes.io/projected/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-kube-api-access-n5rzh\") pod \"heat-db-sync-7fvlk\" (UID: \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\") " pod="openstack/heat-db-sync-7fvlk" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.631072 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6844bbffb5-6qbh8"] Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.647239 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-scripts\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.647331 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-combined-ca-bundle\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.647384 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgn44\" (UniqueName: \"kubernetes.io/projected/418cb883-abd1-46b4-957f-0a40f3e62297-kube-api-access-zgn44\") pod \"barbican-db-sync-zdgpw\" (UID: \"418cb883-abd1-46b4-957f-0a40f3e62297\") " pod="openstack/barbican-db-sync-zdgpw" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.647410 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e92afa62-9c75-4e0e-92f4-76e57328d7a0-logs\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.647458 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418cb883-abd1-46b4-957f-0a40f3e62297-combined-ca-bundle\") pod \"barbican-db-sync-zdgpw\" (UID: \"418cb883-abd1-46b4-957f-0a40f3e62297\") " pod="openstack/barbican-db-sync-zdgpw" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.647521 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/418cb883-abd1-46b4-957f-0a40f3e62297-db-sync-config-data\") pod \"barbican-db-sync-zdgpw\" (UID: \"418cb883-abd1-46b4-957f-0a40f3e62297\") " pod="openstack/barbican-db-sync-zdgpw" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.647562 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mqp5h\" (UniqueName: \"kubernetes.io/projected/e92afa62-9c75-4e0e-92f4-76e57328d7a0-kube-api-access-mqp5h\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.647611 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-config-data\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.652071 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-htnd9"] Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.680069 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418cb883-abd1-46b4-957f-0a40f3e62297-combined-ca-bundle\") pod \"barbican-db-sync-zdgpw\" (UID: \"418cb883-abd1-46b4-957f-0a40f3e62297\") " pod="openstack/barbican-db-sync-zdgpw" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.680477 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.688919 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgn44\" (UniqueName: \"kubernetes.io/projected/418cb883-abd1-46b4-957f-0a40f3e62297-kube-api-access-zgn44\") pod \"barbican-db-sync-zdgpw\" (UID: \"418cb883-abd1-46b4-957f-0a40f3e62297\") " pod="openstack/barbican-db-sync-zdgpw" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.698295 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/418cb883-abd1-46b4-957f-0a40f3e62297-db-sync-config-data\") pod \"barbican-db-sync-zdgpw\" (UID: \"418cb883-abd1-46b4-957f-0a40f3e62297\") " pod="openstack/barbican-db-sync-zdgpw" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.699067 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d8f9dd5cc-6nktg"] Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.711069 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.748801 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e92afa62-9c75-4e0e-92f4-76e57328d7a0-logs\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.749535 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e92afa62-9c75-4e0e-92f4-76e57328d7a0-logs\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.749689 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mqp5h\" (UniqueName: \"kubernetes.io/projected/e92afa62-9c75-4e0e-92f4-76e57328d7a0-kube-api-access-mqp5h\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.749733 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-config-data\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.750086 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-scripts\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.766610 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-config-data\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.766776 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-combined-ca-bundle\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.768836 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d8f9dd5cc-6nktg"] Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.796246 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-scripts\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.809186 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-combined-ca-bundle\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.820219 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zdgpw" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.830768 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mqp5h\" (UniqueName: \"kubernetes.io/projected/e92afa62-9c75-4e0e-92f4-76e57328d7a0-kube-api-access-mqp5h\") pod \"placement-db-sync-htnd9\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.872495 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-ovsdbserver-nb\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.872623 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-dns-swift-storage-0\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.872793 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-config\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.872845 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29zt6\" (UniqueName: \"kubernetes.io/projected/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-kube-api-access-29zt6\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.872993 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-ovsdbserver-sb\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.873099 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-dns-svc\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.889885 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-htnd9" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.893630 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-7fvlk" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.937398 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-67d6b4b8f7-nrxn8"] Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.939211 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.956537 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-6wd56"] Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.980172 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-ovsdbserver-sb\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.980237 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-dns-svc\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.980361 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-ovsdbserver-nb\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.980384 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-dns-swift-storage-0\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.980457 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-config\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.980477 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29zt6\" (UniqueName: \"kubernetes.io/projected/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-kube-api-access-29zt6\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.981853 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-ovsdbserver-nb\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.982425 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-ovsdbserver-sb\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.982735 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-dns-swift-storage-0\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.983055 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-dns-svc\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.983307 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-config\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:05 crc kubenswrapper[4632]: I0313 10:25:05.975925 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67d6b4b8f7-nrxn8"] Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.039400 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6844bbffb5-6qbh8"] Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.091324 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxkl9\" (UniqueName: \"kubernetes.io/projected/95fe9a38-2b32-411e-9121-ad4cc32f159e-kube-api-access-zxkl9\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.091423 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95fe9a38-2b32-411e-9121-ad4cc32f159e-scripts\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.091455 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/95fe9a38-2b32-411e-9121-ad4cc32f159e-config-data\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.091487 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95fe9a38-2b32-411e-9121-ad4cc32f159e-logs\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.091534 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/95fe9a38-2b32-411e-9121-ad4cc32f159e-horizon-secret-key\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.148519 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29zt6\" (UniqueName: \"kubernetes.io/projected/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-kube-api-access-29zt6\") pod \"dnsmasq-dns-7d8f9dd5cc-6nktg\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.203628 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95fe9a38-2b32-411e-9121-ad4cc32f159e-scripts\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.203704 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/95fe9a38-2b32-411e-9121-ad4cc32f159e-config-data\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.204659 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95fe9a38-2b32-411e-9121-ad4cc32f159e-scripts\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.207275 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/95fe9a38-2b32-411e-9121-ad4cc32f159e-config-data\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.215224 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95fe9a38-2b32-411e-9121-ad4cc32f159e-logs\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.215367 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/95fe9a38-2b32-411e-9121-ad4cc32f159e-horizon-secret-key\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.215513 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxkl9\" (UniqueName: \"kubernetes.io/projected/95fe9a38-2b32-411e-9121-ad4cc32f159e-kube-api-access-zxkl9\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.224795 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95fe9a38-2b32-411e-9121-ad4cc32f159e-logs\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.232411 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/95fe9a38-2b32-411e-9121-ad4cc32f159e-horizon-secret-key\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: W0313 10:25:06.255776 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6dd01e75_b01a_439d_953a_a7b35aefaccf.slice/crio-e5726b2fe96111fb5381a471a1eb71f1473219059f290427afb5975ccc268d97 WatchSource:0}: Error finding container e5726b2fe96111fb5381a471a1eb71f1473219059f290427afb5975ccc268d97: Status 404 returned error can't find the container with id e5726b2fe96111fb5381a471a1eb71f1473219059f290427afb5975ccc268d97 Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.318601 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-hlsnz"] Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.345969 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxkl9\" (UniqueName: \"kubernetes.io/projected/95fe9a38-2b32-411e-9121-ad4cc32f159e-kube-api-access-zxkl9\") pod \"horizon-67d6b4b8f7-nrxn8\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.384548 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.454826 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.466164 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" event={"ID":"8e8562e5-7677-460c-864c-c0f1dcd2ac41","Type":"ContainerStarted","Data":"b4b78a7ce795d6bbb416eb8a947d7920ad2bf057be657ab85e7df8a0742013f4"} Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.470415 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6wd56" event={"ID":"6dd01e75-b01a-439d-953a-a7b35aefaccf","Type":"ContainerStarted","Data":"e5726b2fe96111fb5381a471a1eb71f1473219059f290427afb5975ccc268d97"} Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.531236 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.531079 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-756d4f5c49-ng8tb" event={"ID":"875ab9dc-abac-45c4-86b9-b0bfccdfb240","Type":"ContainerDied","Data":"0260417d40f7e52a2270d809525583f83a0891dc6b82c6c6e13d0d522dd4a9b4"} Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.531540 4632 scope.go:117] "RemoveContainer" containerID="d5fd04632add169f284e136d8cc0cc1ed2dafea7a7d420e5a55cd81556345dd4" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.544584 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hlsnz" event={"ID":"b7221b50-7231-4ade-917e-b10f177cb539","Type":"ContainerStarted","Data":"20f645b899e167ff59a24d843990ef38d86d73ef7009bca8f9190936862bedaf"} Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.616557 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.625737 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-ovsdbserver-sb\") pod \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.625799 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5jtw\" (UniqueName: \"kubernetes.io/projected/875ab9dc-abac-45c4-86b9-b0bfccdfb240-kube-api-access-t5jtw\") pod \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.625857 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-config\") pod \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.625920 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-dns-svc\") pod \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.626006 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-ovsdbserver-nb\") pod \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.626228 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-dns-swift-storage-0\") pod \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\" (UID: \"875ab9dc-abac-45c4-86b9-b0bfccdfb240\") " Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.637492 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/875ab9dc-abac-45c4-86b9-b0bfccdfb240-kube-api-access-t5jtw" (OuterVolumeSpecName: "kube-api-access-t5jtw") pod "875ab9dc-abac-45c4-86b9-b0bfccdfb240" (UID: "875ab9dc-abac-45c4-86b9-b0bfccdfb240"). InnerVolumeSpecName "kube-api-access-t5jtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.666921 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.808119 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5jtw\" (UniqueName: \"kubernetes.io/projected/875ab9dc-abac-45c4-86b9-b0bfccdfb240-kube-api-access-t5jtw\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:06 crc kubenswrapper[4632]: I0313 10:25:06.829373 4632 scope.go:117] "RemoveContainer" containerID="6cc0261f52d1f9ff6e82214738332406f7144515ee6e99809f0b2f51974a5801" Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.044190 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6fb489b64f-prckv"] Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.066386 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "875ab9dc-abac-45c4-86b9-b0bfccdfb240" (UID: "875ab9dc-abac-45c4-86b9-b0bfccdfb240"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.151150 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.158297 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-zdgpw"] Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.172851 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "875ab9dc-abac-45c4-86b9-b0bfccdfb240" (UID: "875ab9dc-abac-45c4-86b9-b0bfccdfb240"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.173931 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-config" (OuterVolumeSpecName: "config") pod "875ab9dc-abac-45c4-86b9-b0bfccdfb240" (UID: "875ab9dc-abac-45c4-86b9-b0bfccdfb240"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.197990 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-kq8lc"] Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.209279 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "875ab9dc-abac-45c4-86b9-b0bfccdfb240" (UID: "875ab9dc-abac-45c4-86b9-b0bfccdfb240"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:07 crc kubenswrapper[4632]: W0313 10:25:07.249713 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f916c05_f172_42b6_9b13_0c8d2058bfb1.slice/crio-8ac8055b0e5fc8cb1135e4ae559dd9794358a9f9dfb68fd20402b62c57115f00 WatchSource:0}: Error finding container 8ac8055b0e5fc8cb1135e4ae559dd9794358a9f9dfb68fd20402b62c57115f00: Status 404 returned error can't find the container with id 8ac8055b0e5fc8cb1135e4ae559dd9794358a9f9dfb68fd20402b62c57115f00 Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.254403 4632 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.254557 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.254614 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.259357 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "875ab9dc-abac-45c4-86b9-b0bfccdfb240" (UID: "875ab9dc-abac-45c4-86b9-b0bfccdfb240"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.365572 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/875ab9dc-abac-45c4-86b9-b0bfccdfb240-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.375391 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-htnd9"] Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.464978 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-7fvlk"] Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.552210 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d8f9dd5cc-6nktg"] Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.612955 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"270ebc10-986f-4473-8a5e-9094de34ae98","Type":"ContainerStarted","Data":"c4fcb786f7a33daa32bea87a76b7b56e9f86402051990ca301fe80823cca805f"} Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.643640 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-756d4f5c49-ng8tb"] Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.649973 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-756d4f5c49-ng8tb"] Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.650386 4632 generic.go:334] "Generic (PLEG): container finished" podID="8e8562e5-7677-460c-864c-c0f1dcd2ac41" containerID="33cfae1bedfa94ab3959b48b4e6591f364a299ec2695a359dd153fc66fde7615" exitCode=0 Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.650662 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" event={"ID":"8e8562e5-7677-460c-864c-c0f1dcd2ac41","Type":"ContainerDied","Data":"33cfae1bedfa94ab3959b48b4e6591f364a299ec2695a359dd153fc66fde7615"} Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.682513 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kq8lc" event={"ID":"8f916c05-f172-42b6-9b13-0c8d2058bfb1","Type":"ContainerStarted","Data":"8ac8055b0e5fc8cb1135e4ae559dd9794358a9f9dfb68fd20402b62c57115f00"} Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.696200 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-67d6b4b8f7-nrxn8"] Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.696356 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-htnd9" event={"ID":"e92afa62-9c75-4e0e-92f4-76e57328d7a0","Type":"ContainerStarted","Data":"fa8253910988ff0dbee81a3230f0ff84637c4204c805ed0e40f0cc26f23d5381"} Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.696465 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zdgpw" event={"ID":"418cb883-abd1-46b4-957f-0a40f3e62297","Type":"ContainerStarted","Data":"47e1c2b826ae3f1aaa52b7a4210b405df85537a8c7de35fb1657923a6d754982"} Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.702026 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-7fvlk" event={"ID":"d722ddd7-e65d-44f7-a02d-18ddf126ccf5","Type":"ContainerStarted","Data":"76d57552a9eced6e283cb6dee93cf8db23032b8fbb20e4a910d615de236f52d7"} Mar 13 10:25:07 crc kubenswrapper[4632]: I0313 10:25:07.722918 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fb489b64f-prckv" event={"ID":"00c4bdc6-a22c-4ab6-b898-cf591b92756b","Type":"ContainerStarted","Data":"1e0d86fdbf39635fdea4aed078faa89b9573bea1f02b182e9ea0c1a965b0c550"} Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.062636 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="875ab9dc-abac-45c4-86b9-b0bfccdfb240" path="/var/lib/kubelet/pods/875ab9dc-abac-45c4-86b9-b0bfccdfb240/volumes" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.397542 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.455070 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-ovsdbserver-nb\") pod \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.455125 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-dns-svc\") pod \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.455176 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slg5l\" (UniqueName: \"kubernetes.io/projected/8e8562e5-7677-460c-864c-c0f1dcd2ac41-kube-api-access-slg5l\") pod \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.455220 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-config\") pod \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.455287 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-dns-swift-storage-0\") pod \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.455315 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-ovsdbserver-sb\") pod \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.501060 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8e8562e5-7677-460c-864c-c0f1dcd2ac41" (UID: "8e8562e5-7677-460c-864c-c0f1dcd2ac41"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.518156 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e8562e5-7677-460c-864c-c0f1dcd2ac41-kube-api-access-slg5l" (OuterVolumeSpecName: "kube-api-access-slg5l") pod "8e8562e5-7677-460c-864c-c0f1dcd2ac41" (UID: "8e8562e5-7677-460c-864c-c0f1dcd2ac41"). InnerVolumeSpecName "kube-api-access-slg5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.537339 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8e8562e5-7677-460c-864c-c0f1dcd2ac41" (UID: "8e8562e5-7677-460c-864c-c0f1dcd2ac41"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.558470 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8e8562e5-7677-460c-864c-c0f1dcd2ac41" (UID: "8e8562e5-7677-460c-864c-c0f1dcd2ac41"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.559398 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-ovsdbserver-sb\") pod \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\" (UID: \"8e8562e5-7677-460c-864c-c0f1dcd2ac41\") " Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.559881 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.559896 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slg5l\" (UniqueName: \"kubernetes.io/projected/8e8562e5-7677-460c-864c-c0f1dcd2ac41-kube-api-access-slg5l\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.559907 4632 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:08 crc kubenswrapper[4632]: W0313 10:25:08.560027 4632 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/8e8562e5-7677-460c-864c-c0f1dcd2ac41/volumes/kubernetes.io~configmap/ovsdbserver-sb Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.560042 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8e8562e5-7677-460c-864c-c0f1dcd2ac41" (UID: "8e8562e5-7677-460c-864c-c0f1dcd2ac41"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.596567 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-config" (OuterVolumeSpecName: "config") pod "8e8562e5-7677-460c-864c-c0f1dcd2ac41" (UID: "8e8562e5-7677-460c-864c-c0f1dcd2ac41"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.629314 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8e8562e5-7677-460c-864c-c0f1dcd2ac41" (UID: "8e8562e5-7677-460c-864c-c0f1dcd2ac41"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.662963 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.662992 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.663002 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8e8562e5-7677-460c-864c-c0f1dcd2ac41-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.768418 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hlsnz" event={"ID":"b7221b50-7231-4ade-917e-b10f177cb539","Type":"ContainerStarted","Data":"a4f9bd4f877455829b998ee69c6d5f9dd7fb999a6d06fe2960e4af1bfddc1eb0"} Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.776469 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67d6b4b8f7-nrxn8" event={"ID":"95fe9a38-2b32-411e-9121-ad4cc32f159e","Type":"ContainerStarted","Data":"23d0d6f6bc6174b2a86ec905a9477b2974881387bec66374cfa55dca37114aec"} Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.784582 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" event={"ID":"8e8562e5-7677-460c-864c-c0f1dcd2ac41","Type":"ContainerDied","Data":"b4b78a7ce795d6bbb416eb8a947d7920ad2bf057be657ab85e7df8a0742013f4"} Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.784899 4632 scope.go:117] "RemoveContainer" containerID="33cfae1bedfa94ab3959b48b4e6591f364a299ec2695a359dd153fc66fde7615" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.785022 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6844bbffb5-6qbh8" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.920167 4632 generic.go:334] "Generic (PLEG): container finished" podID="78e29b83-b50e-46db-a8d6-bba0ecfb5c08" containerID="80dc69ee9ec968911aeb73d01429c087d3144584fba19b07c0a8b37e75187f19" exitCode=0 Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.920638 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" event={"ID":"78e29b83-b50e-46db-a8d6-bba0ecfb5c08","Type":"ContainerDied","Data":"80dc69ee9ec968911aeb73d01429c087d3144584fba19b07c0a8b37e75187f19"} Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.920748 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" event={"ID":"78e29b83-b50e-46db-a8d6-bba0ecfb5c08","Type":"ContainerStarted","Data":"5187e6e9a0835d8922aa8452723fd7620bf5222c8a96f16a5be9778d8386494d"} Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.928358 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-hlsnz" podStartSLOduration=4.928337615 podStartE2EDuration="4.928337615s" podCreationTimestamp="2026-03-13 10:25:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:25:08.838313727 +0000 UTC m=+1282.860843870" watchObservedRunningTime="2026-03-13 10:25:08.928337615 +0000 UTC m=+1282.950867748" Mar 13 10:25:08 crc kubenswrapper[4632]: I0313 10:25:08.991247 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6wd56" event={"ID":"6dd01e75-b01a-439d-953a-a7b35aefaccf","Type":"ContainerStarted","Data":"19adb417107921a77df964ab1bd8c8cf0029e40afcac705a66952307655b68b9"} Mar 13 10:25:09 crc kubenswrapper[4632]: I0313 10:25:09.057250 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6844bbffb5-6qbh8"] Mar 13 10:25:09 crc kubenswrapper[4632]: I0313 10:25:09.094549 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6844bbffb5-6qbh8"] Mar 13 10:25:09 crc kubenswrapper[4632]: I0313 10:25:09.164046 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-6wd56" podStartSLOduration=6.164022857 podStartE2EDuration="6.164022857s" podCreationTimestamp="2026-03-13 10:25:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:25:09.125710938 +0000 UTC m=+1283.148241071" watchObservedRunningTime="2026-03-13 10:25:09.164022857 +0000 UTC m=+1283.186552990" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.027778 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" event={"ID":"78e29b83-b50e-46db-a8d6-bba0ecfb5c08","Type":"ContainerStarted","Data":"0f3b47ae46ac068badd9fe9f0befa9613632a7901ef641ad38d1419cf04cc4af"} Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.027829 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.086213 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" podStartSLOduration=5.086145511 podStartE2EDuration="5.086145511s" podCreationTimestamp="2026-03-13 10:25:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:25:10.084129672 +0000 UTC m=+1284.106659815" watchObservedRunningTime="2026-03-13 10:25:10.086145511 +0000 UTC m=+1284.108675654" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.089894 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e8562e5-7677-460c-864c-c0f1dcd2ac41" path="/var/lib/kubelet/pods/8e8562e5-7677-460c-864c-c0f1dcd2ac41/volumes" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.091240 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6fb489b64f-prckv"] Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.139152 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.170459 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7ff9d5cddf-cz85p"] Mar 13 10:25:10 crc kubenswrapper[4632]: E0313 10:25:10.170882 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875ab9dc-abac-45c4-86b9-b0bfccdfb240" containerName="init" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.170899 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="875ab9dc-abac-45c4-86b9-b0bfccdfb240" containerName="init" Mar 13 10:25:10 crc kubenswrapper[4632]: E0313 10:25:10.170922 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="875ab9dc-abac-45c4-86b9-b0bfccdfb240" containerName="dnsmasq-dns" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.170930 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="875ab9dc-abac-45c4-86b9-b0bfccdfb240" containerName="dnsmasq-dns" Mar 13 10:25:10 crc kubenswrapper[4632]: E0313 10:25:10.170970 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8562e5-7677-460c-864c-c0f1dcd2ac41" containerName="init" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.170986 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8562e5-7677-460c-864c-c0f1dcd2ac41" containerName="init" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.171206 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="875ab9dc-abac-45c4-86b9-b0bfccdfb240" containerName="dnsmasq-dns" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.171224 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e8562e5-7677-460c-864c-c0f1dcd2ac41" containerName="init" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.172309 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.199133 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7ff9d5cddf-cz85p"] Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.329912 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgt2s\" (UniqueName: \"kubernetes.io/projected/930f1246-53c8-4970-af1f-a7ef0ae42648-kube-api-access-rgt2s\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.329997 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/930f1246-53c8-4970-af1f-a7ef0ae42648-scripts\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.330061 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/930f1246-53c8-4970-af1f-a7ef0ae42648-config-data\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.330121 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/930f1246-53c8-4970-af1f-a7ef0ae42648-logs\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.330140 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/930f1246-53c8-4970-af1f-a7ef0ae42648-horizon-secret-key\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.431166 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/930f1246-53c8-4970-af1f-a7ef0ae42648-logs\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.431205 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/930f1246-53c8-4970-af1f-a7ef0ae42648-horizon-secret-key\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.431256 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgt2s\" (UniqueName: \"kubernetes.io/projected/930f1246-53c8-4970-af1f-a7ef0ae42648-kube-api-access-rgt2s\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.431286 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/930f1246-53c8-4970-af1f-a7ef0ae42648-scripts\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.431337 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/930f1246-53c8-4970-af1f-a7ef0ae42648-config-data\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.431656 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/930f1246-53c8-4970-af1f-a7ef0ae42648-logs\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.432464 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/930f1246-53c8-4970-af1f-a7ef0ae42648-scripts\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.433753 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/930f1246-53c8-4970-af1f-a7ef0ae42648-config-data\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.445354 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/930f1246-53c8-4970-af1f-a7ef0ae42648-horizon-secret-key\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.465558 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgt2s\" (UniqueName: \"kubernetes.io/projected/930f1246-53c8-4970-af1f-a7ef0ae42648-kube-api-access-rgt2s\") pod \"horizon-7ff9d5cddf-cz85p\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:10 crc kubenswrapper[4632]: I0313 10:25:10.517783 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:11 crc kubenswrapper[4632]: I0313 10:25:11.280006 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7ff9d5cddf-cz85p"] Mar 13 10:25:12 crc kubenswrapper[4632]: I0313 10:25:12.096052 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7ff9d5cddf-cz85p" event={"ID":"930f1246-53c8-4970-af1f-a7ef0ae42648","Type":"ContainerStarted","Data":"adf64e4e5f85756a4d7fe85854309c14415a387724ea67892ccde00c8e6a4b0e"} Mar 13 10:25:14 crc kubenswrapper[4632]: I0313 10:25:14.141065 4632 generic.go:334] "Generic (PLEG): container finished" podID="6dd01e75-b01a-439d-953a-a7b35aefaccf" containerID="19adb417107921a77df964ab1bd8c8cf0029e40afcac705a66952307655b68b9" exitCode=0 Mar 13 10:25:14 crc kubenswrapper[4632]: I0313 10:25:14.141295 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6wd56" event={"ID":"6dd01e75-b01a-439d-953a-a7b35aefaccf","Type":"ContainerDied","Data":"19adb417107921a77df964ab1bd8c8cf0029e40afcac705a66952307655b68b9"} Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.009445 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67d6b4b8f7-nrxn8"] Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.029723 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7bdb5f7878-ng2k2"] Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.031235 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.035621 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.078018 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bdb5f7878-ng2k2"] Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.138221 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7ff9d5cddf-cz85p"] Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.149998 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e4afb91-ce26-4325-89c9-2542da2ec48a-scripts\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.150085 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-horizon-secret-key\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.150121 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e4afb91-ce26-4325-89c9-2542da2ec48a-logs\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.150234 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntccx\" (UniqueName: \"kubernetes.io/projected/3e4afb91-ce26-4325-89c9-2542da2ec48a-kube-api-access-ntccx\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.150274 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-combined-ca-bundle\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.150300 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e4afb91-ce26-4325-89c9-2542da2ec48a-config-data\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.150434 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-horizon-tls-certs\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.213461 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-689764498d-rg7vt"] Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.217425 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.254209 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-combined-ca-bundle\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.254267 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e4afb91-ce26-4325-89c9-2542da2ec48a-config-data\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.254358 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-horizon-tls-certs\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.254413 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e4afb91-ce26-4325-89c9-2542da2ec48a-scripts\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.254446 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-horizon-secret-key\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.254466 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e4afb91-ce26-4325-89c9-2542da2ec48a-logs\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.254515 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntccx\" (UniqueName: \"kubernetes.io/projected/3e4afb91-ce26-4325-89c9-2542da2ec48a-kube-api-access-ntccx\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.256012 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e4afb91-ce26-4325-89c9-2542da2ec48a-scripts\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.257224 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e4afb91-ce26-4325-89c9-2542da2ec48a-config-data\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.258754 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-689764498d-rg7vt"] Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.259737 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e4afb91-ce26-4325-89c9-2542da2ec48a-logs\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.280246 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-horizon-tls-certs\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.301928 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-horizon-secret-key\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.308802 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-combined-ca-bundle\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.327755 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntccx\" (UniqueName: \"kubernetes.io/projected/3e4afb91-ce26-4325-89c9-2542da2ec48a-kube-api-access-ntccx\") pod \"horizon-7bdb5f7878-ng2k2\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.356974 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-combined-ca-bundle\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.357044 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-scripts\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.357094 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-config-data\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.357136 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-logs\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.357154 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-horizon-tls-certs\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.357175 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbnrh\" (UniqueName: \"kubernetes.io/projected/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-kube-api-access-xbnrh\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.357216 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-horizon-secret-key\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.392968 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.468438 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-combined-ca-bundle\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.468508 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-scripts\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.468567 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-config-data\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.468612 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-logs\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.468639 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-horizon-tls-certs\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.468663 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbnrh\" (UniqueName: \"kubernetes.io/projected/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-kube-api-access-xbnrh\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.468685 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-horizon-secret-key\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.473039 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-config-data\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.481261 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-scripts\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.489481 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-logs\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.564901 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-combined-ca-bundle\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.566956 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-horizon-secret-key\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.572983 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbnrh\" (UniqueName: \"kubernetes.io/projected/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-kube-api-access-xbnrh\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.618689 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c-horizon-tls-certs\") pod \"horizon-689764498d-rg7vt\" (UID: \"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c\") " pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:15 crc kubenswrapper[4632]: I0313 10:25:15.855787 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:25:16 crc kubenswrapper[4632]: I0313 10:25:16.387072 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:25:16 crc kubenswrapper[4632]: I0313 10:25:16.475957 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b59dbc87f-7zwrj"] Mar 13 10:25:16 crc kubenswrapper[4632]: I0313 10:25:16.476372 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerName="dnsmasq-dns" containerID="cri-o://f1255f2b0d97d7bcc13a7045fc5d8e4778eece89f9f6f1d468ae8c05e428c6f7" gracePeriod=10 Mar 13 10:25:16 crc kubenswrapper[4632]: I0313 10:25:16.719333 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Mar 13 10:25:17 crc kubenswrapper[4632]: I0313 10:25:17.187239 4632 generic.go:334] "Generic (PLEG): container finished" podID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerID="f1255f2b0d97d7bcc13a7045fc5d8e4778eece89f9f6f1d468ae8c05e428c6f7" exitCode=0 Mar 13 10:25:17 crc kubenswrapper[4632]: I0313 10:25:17.187279 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" event={"ID":"7203640d-964c-4c28-8cc2-6a7ae27cdab3","Type":"ContainerDied","Data":"f1255f2b0d97d7bcc13a7045fc5d8e4778eece89f9f6f1d468ae8c05e428c6f7"} Mar 13 10:25:21 crc kubenswrapper[4632]: I0313 10:25:21.224986 4632 generic.go:334] "Generic (PLEG): container finished" podID="4f1c5663-463b-45e2-b200-64e73e6d5698" containerID="bf8d93edd68f1cf79021467ff9910419baf75397a4140fb3d25bca7f97abbf70" exitCode=0 Mar 13 10:25:21 crc kubenswrapper[4632]: I0313 10:25:21.225711 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l6hpb" event={"ID":"4f1c5663-463b-45e2-b200-64e73e6d5698","Type":"ContainerDied","Data":"bf8d93edd68f1cf79021467ff9910419baf75397a4140fb3d25bca7f97abbf70"} Mar 13 10:25:21 crc kubenswrapper[4632]: I0313 10:25:21.719349 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Mar 13 10:25:26 crc kubenswrapper[4632]: I0313 10:25:26.719552 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Mar 13 10:25:26 crc kubenswrapper[4632]: I0313 10:25:26.720800 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:25:31 crc kubenswrapper[4632]: I0313 10:25:31.718774 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: connect: connection refused" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.325899 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.379660 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-6wd56" event={"ID":"6dd01e75-b01a-439d-953a-a7b35aefaccf","Type":"ContainerDied","Data":"e5726b2fe96111fb5381a471a1eb71f1473219059f290427afb5975ccc268d97"} Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.379730 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5726b2fe96111fb5381a471a1eb71f1473219059f290427afb5975ccc268d97" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.379745 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-6wd56" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.460358 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q9qt5\" (UniqueName: \"kubernetes.io/projected/6dd01e75-b01a-439d-953a-a7b35aefaccf-kube-api-access-q9qt5\") pod \"6dd01e75-b01a-439d-953a-a7b35aefaccf\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.460840 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-scripts\") pod \"6dd01e75-b01a-439d-953a-a7b35aefaccf\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.460953 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-fernet-keys\") pod \"6dd01e75-b01a-439d-953a-a7b35aefaccf\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.460987 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-credential-keys\") pod \"6dd01e75-b01a-439d-953a-a7b35aefaccf\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.461032 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-config-data\") pod \"6dd01e75-b01a-439d-953a-a7b35aefaccf\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.461081 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-combined-ca-bundle\") pod \"6dd01e75-b01a-439d-953a-a7b35aefaccf\" (UID: \"6dd01e75-b01a-439d-953a-a7b35aefaccf\") " Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.470086 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6dd01e75-b01a-439d-953a-a7b35aefaccf" (UID: "6dd01e75-b01a-439d-953a-a7b35aefaccf"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.470129 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6dd01e75-b01a-439d-953a-a7b35aefaccf" (UID: "6dd01e75-b01a-439d-953a-a7b35aefaccf"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.470177 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-scripts" (OuterVolumeSpecName: "scripts") pod "6dd01e75-b01a-439d-953a-a7b35aefaccf" (UID: "6dd01e75-b01a-439d-953a-a7b35aefaccf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.473656 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6dd01e75-b01a-439d-953a-a7b35aefaccf-kube-api-access-q9qt5" (OuterVolumeSpecName: "kube-api-access-q9qt5") pod "6dd01e75-b01a-439d-953a-a7b35aefaccf" (UID: "6dd01e75-b01a-439d-953a-a7b35aefaccf"). InnerVolumeSpecName "kube-api-access-q9qt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.503339 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-config-data" (OuterVolumeSpecName: "config-data") pod "6dd01e75-b01a-439d-953a-a7b35aefaccf" (UID: "6dd01e75-b01a-439d-953a-a7b35aefaccf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.503551 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6dd01e75-b01a-439d-953a-a7b35aefaccf" (UID: "6dd01e75-b01a-439d-953a-a7b35aefaccf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.563650 4632 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.563694 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.563706 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.563718 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q9qt5\" (UniqueName: \"kubernetes.io/projected/6dd01e75-b01a-439d-953a-a7b35aefaccf-kube-api-access-q9qt5\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.563732 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:33 crc kubenswrapper[4632]: I0313 10:25:33.563743 4632 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6dd01e75-b01a-439d-953a-a7b35aefaccf-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:33 crc kubenswrapper[4632]: E0313 10:25:33.720270 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-heat-engine:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:25:33 crc kubenswrapper[4632]: E0313 10:25:33.720631 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-heat-engine:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:25:33 crc kubenswrapper[4632]: E0313 10:25:33.720778 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-heat-engine:e43235cb19da04699a53f42b6a75afe9,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-n5rzh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-7fvlk_openstack(d722ddd7-e65d-44f7-a02d-18ddf126ccf5): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:25:33 crc kubenswrapper[4632]: E0313 10:25:33.722599 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-7fvlk" podUID="d722ddd7-e65d-44f7-a02d-18ddf126ccf5" Mar 13 10:25:33 crc kubenswrapper[4632]: E0313 10:25:33.729979 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:25:33 crc kubenswrapper[4632]: E0313 10:25:33.730034 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:25:33 crc kubenswrapper[4632]: E0313 10:25:33.730150 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:e43235cb19da04699a53f42b6a75afe9,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n65dh55dh9fh567h567h7ch66ch6ch5d6hfbh658hfchbch5dch59h55fh94hbch5d5hbch685h649h87h5h56bh5bh56fh588h9h5c6h54dh696q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gwx9r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6fb489b64f-prckv_openstack(00c4bdc6-a22c-4ab6-b898-cf591b92756b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:25:33 crc kubenswrapper[4632]: E0313 10:25:33.732311 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:e43235cb19da04699a53f42b6a75afe9\\\"\"]" pod="openstack/horizon-6fb489b64f-prckv" podUID="00c4bdc6-a22c-4ab6-b898-cf591b92756b" Mar 13 10:25:34 crc kubenswrapper[4632]: E0313 10:25:34.402967 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-heat-engine:e43235cb19da04699a53f42b6a75afe9\\\"\"" pod="openstack/heat-db-sync-7fvlk" podUID="d722ddd7-e65d-44f7-a02d-18ddf126ccf5" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.471144 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-6wd56"] Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.481109 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-6wd56"] Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.537880 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-x8tq8"] Mar 13 10:25:34 crc kubenswrapper[4632]: E0313 10:25:34.539541 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6dd01e75-b01a-439d-953a-a7b35aefaccf" containerName="keystone-bootstrap" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.539653 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6dd01e75-b01a-439d-953a-a7b35aefaccf" containerName="keystone-bootstrap" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.539987 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6dd01e75-b01a-439d-953a-a7b35aefaccf" containerName="keystone-bootstrap" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.541323 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.543848 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.544701 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.545496 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-llpcf" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.546104 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.546888 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.552204 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-x8tq8"] Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.694293 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-scripts\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.694368 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln474\" (UniqueName: \"kubernetes.io/projected/d8d0f662-d180-4137-8107-e465c5fb0621-kube-api-access-ln474\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.694517 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-combined-ca-bundle\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.694551 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-fernet-keys\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.694620 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-config-data\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.695431 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-credential-keys\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.798668 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-credential-keys\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.798774 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-scripts\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.798801 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln474\" (UniqueName: \"kubernetes.io/projected/d8d0f662-d180-4137-8107-e465c5fb0621-kube-api-access-ln474\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.798839 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-combined-ca-bundle\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.798863 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-fernet-keys\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.798880 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-config-data\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.813138 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-scripts\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.814302 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-fernet-keys\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.814378 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-credential-keys\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.821496 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-combined-ca-bundle\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.828847 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-config-data\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.832330 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln474\" (UniqueName: \"kubernetes.io/projected/d8d0f662-d180-4137-8107-e465c5fb0621-kube-api-access-ln474\") pod \"keystone-bootstrap-x8tq8\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:34 crc kubenswrapper[4632]: I0313 10:25:34.936654 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.057346 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6dd01e75-b01a-439d-953a-a7b35aefaccf" path="/var/lib/kubelet/pods/6dd01e75-b01a-439d-953a-a7b35aefaccf/volumes" Mar 13 10:25:36 crc kubenswrapper[4632]: E0313 10:25:36.211020 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-placement-api:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:25:36 crc kubenswrapper[4632]: E0313 10:25:36.211082 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-placement-api:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:25:36 crc kubenswrapper[4632]: E0313 10:25:36.211206 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-placement-api:e43235cb19da04699a53f42b6a75afe9,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mqp5h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-htnd9_openstack(e92afa62-9c75-4e0e-92f4-76e57328d7a0): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:25:36 crc kubenswrapper[4632]: E0313 10:25:36.212399 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-htnd9" podUID="e92afa62-9c75-4e0e-92f4-76e57328d7a0" Mar 13 10:25:36 crc kubenswrapper[4632]: E0313 10:25:36.223051 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:25:36 crc kubenswrapper[4632]: E0313 10:25:36.223113 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:25:36 crc kubenswrapper[4632]: E0313 10:25:36.223245 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:e43235cb19da04699a53f42b6a75afe9,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n56dhbh68dhfbh569h676h64fh69h669h65fh64fh85h5c7h588h5bdh76h65chd7h5b8h5ch677hc7h64ch8bh5fdh75h7dh5cbhf9h76h688h5d4q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rgt2s,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7ff9d5cddf-cz85p_openstack(930f1246-53c8-4970-af1f-a7ef0ae42648): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:25:36 crc kubenswrapper[4632]: E0313 10:25:36.226586 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-horizon:e43235cb19da04699a53f42b6a75afe9\\\"\"]" pod="openstack/horizon-7ff9d5cddf-cz85p" podUID="930f1246-53c8-4970-af1f-a7ef0ae42648" Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.307482 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l6hpb" Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.429876 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-combined-ca-bundle\") pod \"4f1c5663-463b-45e2-b200-64e73e6d5698\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.429927 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-config-data\") pod \"4f1c5663-463b-45e2-b200-64e73e6d5698\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.430085 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-db-sync-config-data\") pod \"4f1c5663-463b-45e2-b200-64e73e6d5698\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.430147 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjbxx\" (UniqueName: \"kubernetes.io/projected/4f1c5663-463b-45e2-b200-64e73e6d5698-kube-api-access-fjbxx\") pod \"4f1c5663-463b-45e2-b200-64e73e6d5698\" (UID: \"4f1c5663-463b-45e2-b200-64e73e6d5698\") " Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.457052 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f1c5663-463b-45e2-b200-64e73e6d5698-kube-api-access-fjbxx" (OuterVolumeSpecName: "kube-api-access-fjbxx") pod "4f1c5663-463b-45e2-b200-64e73e6d5698" (UID: "4f1c5663-463b-45e2-b200-64e73e6d5698"). InnerVolumeSpecName "kube-api-access-fjbxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.464232 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4f1c5663-463b-45e2-b200-64e73e6d5698" (UID: "4f1c5663-463b-45e2-b200-64e73e6d5698"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.514455 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-l6hpb" Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.514933 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-l6hpb" event={"ID":"4f1c5663-463b-45e2-b200-64e73e6d5698","Type":"ContainerDied","Data":"0a5d62eda0a21b4de62c912c034c3914a852ed117fa1d5a908a4b0e7b70dc6a3"} Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.514990 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a5d62eda0a21b4de62c912c034c3914a852ed117fa1d5a908a4b0e7b70dc6a3" Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.521067 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f1c5663-463b-45e2-b200-64e73e6d5698" (UID: "4f1c5663-463b-45e2-b200-64e73e6d5698"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:25:36 crc kubenswrapper[4632]: E0313 10:25:36.521240 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-placement-api:e43235cb19da04699a53f42b6a75afe9\\\"\"" pod="openstack/placement-db-sync-htnd9" podUID="e92afa62-9c75-4e0e-92f4-76e57328d7a0" Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.532440 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.532480 4632 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.532494 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjbxx\" (UniqueName: \"kubernetes.io/projected/4f1c5663-463b-45e2-b200-64e73e6d5698-kube-api-access-fjbxx\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.609130 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-config-data" (OuterVolumeSpecName: "config-data") pod "4f1c5663-463b-45e2-b200-64e73e6d5698" (UID: "4f1c5663-463b-45e2-b200-64e73e6d5698"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:25:36 crc kubenswrapper[4632]: I0313 10:25:36.636227 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f1c5663-463b-45e2-b200-64e73e6d5698-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.168286 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5db978f585-jtbcw"] Mar 13 10:25:38 crc kubenswrapper[4632]: E0313 10:25:38.169472 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f1c5663-463b-45e2-b200-64e73e6d5698" containerName="glance-db-sync" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.169495 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f1c5663-463b-45e2-b200-64e73e6d5698" containerName="glance-db-sync" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.169931 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f1c5663-463b-45e2-b200-64e73e6d5698" containerName="glance-db-sync" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.171992 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.175869 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5db978f585-jtbcw"] Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.238632 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-ovsdbserver-sb\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.238768 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsqxq\" (UniqueName: \"kubernetes.io/projected/4cf1d659-89cc-471b-8089-bc85f7ab3578-kube-api-access-gsqxq\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.238845 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-dns-swift-storage-0\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.238880 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-ovsdbserver-nb\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.238910 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-config\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.239081 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-dns-svc\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.340880 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-dns-swift-storage-0\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.342133 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-ovsdbserver-nb\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.342265 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-dns-swift-storage-0\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.342571 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-config\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.342652 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-ovsdbserver-nb\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.342808 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-dns-svc\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.342905 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-ovsdbserver-sb\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.343166 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsqxq\" (UniqueName: \"kubernetes.io/projected/4cf1d659-89cc-471b-8089-bc85f7ab3578-kube-api-access-gsqxq\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.343784 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-ovsdbserver-sb\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.344019 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-dns-svc\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.344849 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-config\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.366833 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsqxq\" (UniqueName: \"kubernetes.io/projected/4cf1d659-89cc-471b-8089-bc85f7ab3578-kube-api-access-gsqxq\") pod \"dnsmasq-dns-5db978f585-jtbcw\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.550799 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.777250 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.790815 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.793629 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.793897 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qpd5p" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.795265 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.799504 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.855222 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37dc6e5d-eb14-4cef-9451-7c567c6c9068-logs\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.855417 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-config-data\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.855529 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37dc6e5d-eb14-4cef-9451-7c567c6c9068-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.855593 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-scripts\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.856138 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.856364 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.856398 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rqnd\" (UniqueName: \"kubernetes.io/projected/37dc6e5d-eb14-4cef-9451-7c567c6c9068-kube-api-access-5rqnd\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.958037 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37dc6e5d-eb14-4cef-9451-7c567c6c9068-logs\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.958102 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-config-data\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.958147 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37dc6e5d-eb14-4cef-9451-7c567c6c9068-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.958167 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-scripts\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.958196 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.958289 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.958761 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.959116 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rqnd\" (UniqueName: \"kubernetes.io/projected/37dc6e5d-eb14-4cef-9451-7c567c6c9068-kube-api-access-5rqnd\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.959332 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37dc6e5d-eb14-4cef-9451-7c567c6c9068-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.959412 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37dc6e5d-eb14-4cef-9451-7c567c6c9068-logs\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.963921 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-scripts\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.964759 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:38 crc kubenswrapper[4632]: I0313 10:25:38.974388 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-config-data\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.023196 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.023847 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rqnd\" (UniqueName: \"kubernetes.io/projected/37dc6e5d-eb14-4cef-9451-7c567c6c9068-kube-api-access-5rqnd\") pod \"glance-default-external-api-0\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " pod="openstack/glance-default-external-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: E0313 10:25:39.113144 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-barbican-api:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:25:39 crc kubenswrapper[4632]: E0313 10:25:39.113203 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-barbican-api:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:25:39 crc kubenswrapper[4632]: E0313 10:25:39.113380 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-barbican-api:e43235cb19da04699a53f42b6a75afe9,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zgn44,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-zdgpw_openstack(418cb883-abd1-46b4-957f-0a40f3e62297): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:25:39 crc kubenswrapper[4632]: E0313 10:25:39.114758 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-zdgpw" podUID="418cb883-abd1-46b4-957f-0a40f3e62297" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.121763 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.299433 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.301264 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.303675 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.323033 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.366024 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1756bbdc-3e6c-4815-96a7-0620f7400cb7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.366093 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.366168 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.366264 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.366349 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.366382 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1756bbdc-3e6c-4815-96a7-0620f7400cb7-logs\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.366419 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jn72\" (UniqueName: \"kubernetes.io/projected/1756bbdc-3e6c-4815-96a7-0620f7400cb7-kube-api-access-7jn72\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.467805 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.467882 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.467908 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1756bbdc-3e6c-4815-96a7-0620f7400cb7-logs\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.467935 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jn72\" (UniqueName: \"kubernetes.io/projected/1756bbdc-3e6c-4815-96a7-0620f7400cb7-kube-api-access-7jn72\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.468031 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1756bbdc-3e6c-4815-96a7-0620f7400cb7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.468062 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.468107 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.468403 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.469299 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1756bbdc-3e6c-4815-96a7-0620f7400cb7-logs\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.469439 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1756bbdc-3e6c-4815-96a7-0620f7400cb7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.474819 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.474985 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.475889 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.490038 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jn72\" (UniqueName: \"kubernetes.io/projected/1756bbdc-3e6c-4815-96a7-0620f7400cb7-kube-api-access-7jn72\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.503236 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:25:39 crc kubenswrapper[4632]: E0313 10:25:39.548254 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-barbican-api:e43235cb19da04699a53f42b6a75afe9\\\"\"" pod="openstack/barbican-db-sync-zdgpw" podUID="418cb883-abd1-46b4-957f-0a40f3e62297" Mar 13 10:25:39 crc kubenswrapper[4632]: I0313 10:25:39.628775 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 10:25:41 crc kubenswrapper[4632]: I0313 10:25:41.104578 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:25:41 crc kubenswrapper[4632]: I0313 10:25:41.220578 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:25:41 crc kubenswrapper[4632]: I0313 10:25:41.566357 4632 generic.go:334] "Generic (PLEG): container finished" podID="b7221b50-7231-4ade-917e-b10f177cb539" containerID="a4f9bd4f877455829b998ee69c6d5f9dd7fb999a6d06fe2960e4af1bfddc1eb0" exitCode=0 Mar 13 10:25:41 crc kubenswrapper[4632]: I0313 10:25:41.566415 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hlsnz" event={"ID":"b7221b50-7231-4ade-917e-b10f177cb539","Type":"ContainerDied","Data":"a4f9bd4f877455829b998ee69c6d5f9dd7fb999a6d06fe2960e4af1bfddc1eb0"} Mar 13 10:25:41 crc kubenswrapper[4632]: I0313 10:25:41.719073 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: i/o timeout" Mar 13 10:25:46 crc kubenswrapper[4632]: I0313 10:25:46.719411 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: i/o timeout" Mar 13 10:25:51 crc kubenswrapper[4632]: I0313 10:25:51.720179 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: i/o timeout" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.430668 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.435395 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.445869 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.461763 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hlsnz" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.542261 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/930f1246-53c8-4970-af1f-a7ef0ae42648-logs\") pod \"930f1246-53c8-4970-af1f-a7ef0ae42648\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.542362 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gwx9r\" (UniqueName: \"kubernetes.io/projected/00c4bdc6-a22c-4ab6-b898-cf591b92756b-kube-api-access-gwx9r\") pod \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.542411 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r95hn\" (UniqueName: \"kubernetes.io/projected/7203640d-964c-4c28-8cc2-6a7ae27cdab3-kube-api-access-r95hn\") pod \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.542444 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-ovsdbserver-sb\") pod \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.542476 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00c4bdc6-a22c-4ab6-b898-cf591b92756b-config-data\") pod \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.542566 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00c4bdc6-a22c-4ab6-b898-cf591b92756b-logs\") pod \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.542598 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgt2s\" (UniqueName: \"kubernetes.io/projected/930f1246-53c8-4970-af1f-a7ef0ae42648-kube-api-access-rgt2s\") pod \"930f1246-53c8-4970-af1f-a7ef0ae42648\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.542635 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/930f1246-53c8-4970-af1f-a7ef0ae42648-scripts\") pod \"930f1246-53c8-4970-af1f-a7ef0ae42648\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.542662 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-dns-svc\") pod \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.542699 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-config\") pod \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.542743 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00c4bdc6-a22c-4ab6-b898-cf591b92756b-horizon-secret-key\") pod \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.543457 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/930f1246-53c8-4970-af1f-a7ef0ae42648-config-data\") pod \"930f1246-53c8-4970-af1f-a7ef0ae42648\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.543554 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00c4bdc6-a22c-4ab6-b898-cf591b92756b-scripts\") pod \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\" (UID: \"00c4bdc6-a22c-4ab6-b898-cf591b92756b\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.543625 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/930f1246-53c8-4970-af1f-a7ef0ae42648-horizon-secret-key\") pod \"930f1246-53c8-4970-af1f-a7ef0ae42648\" (UID: \"930f1246-53c8-4970-af1f-a7ef0ae42648\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.543674 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-ovsdbserver-nb\") pod \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\" (UID: \"7203640d-964c-4c28-8cc2-6a7ae27cdab3\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.543454 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/930f1246-53c8-4970-af1f-a7ef0ae42648-logs" (OuterVolumeSpecName: "logs") pod "930f1246-53c8-4970-af1f-a7ef0ae42648" (UID: "930f1246-53c8-4970-af1f-a7ef0ae42648"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.545732 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00c4bdc6-a22c-4ab6-b898-cf591b92756b-logs" (OuterVolumeSpecName: "logs") pod "00c4bdc6-a22c-4ab6-b898-cf591b92756b" (UID: "00c4bdc6-a22c-4ab6-b898-cf591b92756b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.547072 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00c4bdc6-a22c-4ab6-b898-cf591b92756b-config-data" (OuterVolumeSpecName: "config-data") pod "00c4bdc6-a22c-4ab6-b898-cf591b92756b" (UID: "00c4bdc6-a22c-4ab6-b898-cf591b92756b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.580844 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/930f1246-53c8-4970-af1f-a7ef0ae42648-config-data" (OuterVolumeSpecName: "config-data") pod "930f1246-53c8-4970-af1f-a7ef0ae42648" (UID: "930f1246-53c8-4970-af1f-a7ef0ae42648"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.581754 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00c4bdc6-a22c-4ab6-b898-cf591b92756b-scripts" (OuterVolumeSpecName: "scripts") pod "00c4bdc6-a22c-4ab6-b898-cf591b92756b" (UID: "00c4bdc6-a22c-4ab6-b898-cf591b92756b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.582193 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/930f1246-53c8-4970-af1f-a7ef0ae42648-kube-api-access-rgt2s" (OuterVolumeSpecName: "kube-api-access-rgt2s") pod "930f1246-53c8-4970-af1f-a7ef0ae42648" (UID: "930f1246-53c8-4970-af1f-a7ef0ae42648"). InnerVolumeSpecName "kube-api-access-rgt2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.582743 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/930f1246-53c8-4970-af1f-a7ef0ae42648-scripts" (OuterVolumeSpecName: "scripts") pod "930f1246-53c8-4970-af1f-a7ef0ae42648" (UID: "930f1246-53c8-4970-af1f-a7ef0ae42648"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.608409 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7203640d-964c-4c28-8cc2-6a7ae27cdab3-kube-api-access-r95hn" (OuterVolumeSpecName: "kube-api-access-r95hn") pod "7203640d-964c-4c28-8cc2-6a7ae27cdab3" (UID: "7203640d-964c-4c28-8cc2-6a7ae27cdab3"). InnerVolumeSpecName "kube-api-access-r95hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.608583 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00c4bdc6-a22c-4ab6-b898-cf591b92756b-kube-api-access-gwx9r" (OuterVolumeSpecName: "kube-api-access-gwx9r") pod "00c4bdc6-a22c-4ab6-b898-cf591b92756b" (UID: "00c4bdc6-a22c-4ab6-b898-cf591b92756b"). InnerVolumeSpecName "kube-api-access-gwx9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.608680 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/930f1246-53c8-4970-af1f-a7ef0ae42648-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "930f1246-53c8-4970-af1f-a7ef0ae42648" (UID: "930f1246-53c8-4970-af1f-a7ef0ae42648"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.614969 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00c4bdc6-a22c-4ab6-b898-cf591b92756b-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "00c4bdc6-a22c-4ab6-b898-cf591b92756b" (UID: "00c4bdc6-a22c-4ab6-b898-cf591b92756b"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.634154 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7203640d-964c-4c28-8cc2-6a7ae27cdab3" (UID: "7203640d-964c-4c28-8cc2-6a7ae27cdab3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.652609 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngb4t\" (UniqueName: \"kubernetes.io/projected/b7221b50-7231-4ade-917e-b10f177cb539-kube-api-access-ngb4t\") pod \"b7221b50-7231-4ade-917e-b10f177cb539\" (UID: \"b7221b50-7231-4ade-917e-b10f177cb539\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.652741 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7221b50-7231-4ade-917e-b10f177cb539-combined-ca-bundle\") pod \"b7221b50-7231-4ade-917e-b10f177cb539\" (UID: \"b7221b50-7231-4ade-917e-b10f177cb539\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.652879 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7221b50-7231-4ade-917e-b10f177cb539-config\") pod \"b7221b50-7231-4ade-917e-b10f177cb539\" (UID: \"b7221b50-7231-4ade-917e-b10f177cb539\") " Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.654909 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgt2s\" (UniqueName: \"kubernetes.io/projected/930f1246-53c8-4970-af1f-a7ef0ae42648-kube-api-access-rgt2s\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.654930 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/930f1246-53c8-4970-af1f-a7ef0ae42648-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.654981 4632 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00c4bdc6-a22c-4ab6-b898-cf591b92756b-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.654991 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/930f1246-53c8-4970-af1f-a7ef0ae42648-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.654999 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00c4bdc6-a22c-4ab6-b898-cf591b92756b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.655008 4632 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/930f1246-53c8-4970-af1f-a7ef0ae42648-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.655016 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.655024 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/930f1246-53c8-4970-af1f-a7ef0ae42648-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.655054 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gwx9r\" (UniqueName: \"kubernetes.io/projected/00c4bdc6-a22c-4ab6-b898-cf591b92756b-kube-api-access-gwx9r\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.655063 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r95hn\" (UniqueName: \"kubernetes.io/projected/7203640d-964c-4c28-8cc2-6a7ae27cdab3-kube-api-access-r95hn\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.655072 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00c4bdc6-a22c-4ab6-b898-cf591b92756b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.655081 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00c4bdc6-a22c-4ab6-b898-cf591b92756b-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.656761 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7221b50-7231-4ade-917e-b10f177cb539-kube-api-access-ngb4t" (OuterVolumeSpecName: "kube-api-access-ngb4t") pod "b7221b50-7231-4ade-917e-b10f177cb539" (UID: "b7221b50-7231-4ade-917e-b10f177cb539"). InnerVolumeSpecName "kube-api-access-ngb4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.683153 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7203640d-964c-4c28-8cc2-6a7ae27cdab3" (UID: "7203640d-964c-4c28-8cc2-6a7ae27cdab3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.688786 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.688797 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" event={"ID":"7203640d-964c-4c28-8cc2-6a7ae27cdab3","Type":"ContainerDied","Data":"501cdcd9d1f38a4b8b82ad7d76e2b6765f391cfadd65ee750e8254d78d76de84"} Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.689099 4632 scope.go:117] "RemoveContainer" containerID="f1255f2b0d97d7bcc13a7045fc5d8e4778eece89f9f6f1d468ae8c05e428c6f7" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.692025 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-hlsnz" event={"ID":"b7221b50-7231-4ade-917e-b10f177cb539","Type":"ContainerDied","Data":"20f645b899e167ff59a24d843990ef38d86d73ef7009bca8f9190936862bedaf"} Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.692194 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="20f645b899e167ff59a24d843990ef38d86d73ef7009bca8f9190936862bedaf" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.692442 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-hlsnz" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.694176 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-config" (OuterVolumeSpecName: "config") pod "7203640d-964c-4c28-8cc2-6a7ae27cdab3" (UID: "7203640d-964c-4c28-8cc2-6a7ae27cdab3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.695250 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6fb489b64f-prckv" event={"ID":"00c4bdc6-a22c-4ab6-b898-cf591b92756b","Type":"ContainerDied","Data":"1e0d86fdbf39635fdea4aed078faa89b9573bea1f02b182e9ea0c1a965b0c550"} Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.695414 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6fb489b64f-prckv" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.697739 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7ff9d5cddf-cz85p" event={"ID":"930f1246-53c8-4970-af1f-a7ef0ae42648","Type":"ContainerDied","Data":"adf64e4e5f85756a4d7fe85854309c14415a387724ea67892ccde00c8e6a4b0e"} Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.697910 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7ff9d5cddf-cz85p" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.711904 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7221b50-7231-4ade-917e-b10f177cb539-config" (OuterVolumeSpecName: "config") pod "b7221b50-7231-4ade-917e-b10f177cb539" (UID: "b7221b50-7231-4ade-917e-b10f177cb539"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.714903 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7203640d-964c-4c28-8cc2-6a7ae27cdab3" (UID: "7203640d-964c-4c28-8cc2-6a7ae27cdab3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.715091 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b7221b50-7231-4ade-917e-b10f177cb539-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b7221b50-7231-4ade-917e-b10f177cb539" (UID: "b7221b50-7231-4ade-917e-b10f177cb539"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.756645 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7221b50-7231-4ade-917e-b10f177cb539-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.756687 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.756701 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngb4t\" (UniqueName: \"kubernetes.io/projected/b7221b50-7231-4ade-917e-b10f177cb539-kube-api-access-ngb4t\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.756712 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.756722 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7203640d-964c-4c28-8cc2-6a7ae27cdab3-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.756732 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7221b50-7231-4ade-917e-b10f177cb539-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.773930 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7ff9d5cddf-cz85p"] Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.797384 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7ff9d5cddf-cz85p"] Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.817813 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6fb489b64f-prckv"] Mar 13 10:25:52 crc kubenswrapper[4632]: I0313 10:25:52.829487 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6fb489b64f-prckv"] Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.030075 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b59dbc87f-7zwrj"] Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.037418 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b59dbc87f-7zwrj"] Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.904511 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5db978f585-jtbcw"] Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.940100 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5c86b4b888-l9574"] Mar 13 10:25:53 crc kubenswrapper[4632]: E0313 10:25:53.941300 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerName="dnsmasq-dns" Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.941321 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerName="dnsmasq-dns" Mar 13 10:25:53 crc kubenswrapper[4632]: E0313 10:25:53.941332 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7221b50-7231-4ade-917e-b10f177cb539" containerName="neutron-db-sync" Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.941338 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7221b50-7231-4ade-917e-b10f177cb539" containerName="neutron-db-sync" Mar 13 10:25:53 crc kubenswrapper[4632]: E0313 10:25:53.941361 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerName="init" Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.941369 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerName="init" Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.941528 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerName="dnsmasq-dns" Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.941550 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7221b50-7231-4ade-917e-b10f177cb539" containerName="neutron-db-sync" Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.942402 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.952825 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-r2t7p" Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.953053 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.953198 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.953328 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 13 10:25:53 crc kubenswrapper[4632]: I0313 10:25:53.961910 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c86b4b888-l9574"] Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.066470 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00c4bdc6-a22c-4ab6-b898-cf591b92756b" path="/var/lib/kubelet/pods/00c4bdc6-a22c-4ab6-b898-cf591b92756b/volumes" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.066900 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" path="/var/lib/kubelet/pods/7203640d-964c-4c28-8cc2-6a7ae27cdab3/volumes" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.067775 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="930f1246-53c8-4970-af1f-a7ef0ae42648" path="/var/lib/kubelet/pods/930f1246-53c8-4970-af1f-a7ef0ae42648/volumes" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.087896 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5776d95bfc-hl9dv"] Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.092813 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lsp7\" (UniqueName: \"kubernetes.io/projected/6d73a499-d334-4a7a-9783-640b98760672-kube-api-access-6lsp7\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.092886 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-httpd-config\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.093029 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-ovndb-tls-certs\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.093070 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-combined-ca-bundle\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.093102 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-config\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.093267 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.102950 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5776d95bfc-hl9dv"] Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.199052 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-ovsdbserver-nb\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.199495 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-dns-swift-storage-0\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.199555 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-ovsdbserver-sb\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.199611 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-ovndb-tls-certs\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.199645 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-dns-svc\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.199671 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-combined-ca-bundle\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.199702 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-config\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.199749 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4rzv\" (UniqueName: \"kubernetes.io/projected/ff547198-2736-4059-8e66-e63ea9ce7345-kube-api-access-v4rzv\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.199796 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-config\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.199825 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lsp7\" (UniqueName: \"kubernetes.io/projected/6d73a499-d334-4a7a-9783-640b98760672-kube-api-access-6lsp7\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.199871 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-httpd-config\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.208290 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-config\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.208752 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-httpd-config\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.212733 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-combined-ca-bundle\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.222486 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-ovndb-tls-certs\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.223837 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lsp7\" (UniqueName: \"kubernetes.io/projected/6d73a499-d334-4a7a-9783-640b98760672-kube-api-access-6lsp7\") pod \"neutron-5c86b4b888-l9574\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.276268 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.301337 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-config\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.301671 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-ovsdbserver-nb\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.301815 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-dns-swift-storage-0\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.301925 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-ovsdbserver-sb\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.302101 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-dns-svc\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.302314 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-config\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.302466 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4rzv\" (UniqueName: \"kubernetes.io/projected/ff547198-2736-4059-8e66-e63ea9ce7345-kube-api-access-v4rzv\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.302876 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-dns-swift-storage-0\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.303489 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-ovsdbserver-nb\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.304299 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-dns-svc\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.304789 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-ovsdbserver-sb\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.344656 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4rzv\" (UniqueName: \"kubernetes.io/projected/ff547198-2736-4059-8e66-e63ea9ce7345-kube-api-access-v4rzv\") pod \"dnsmasq-dns-5776d95bfc-hl9dv\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:54 crc kubenswrapper[4632]: I0313 10:25:54.446605 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:55 crc kubenswrapper[4632]: E0313 10:25:55.073319 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:25:55 crc kubenswrapper[4632]: E0313 10:25:55.073382 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:25:55 crc kubenswrapper[4632]: E0313 10:25:55.073507 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:e43235cb19da04699a53f42b6a75afe9,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5m5bn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-kq8lc_openstack(8f916c05-f172-42b6-9b13-0c8d2058bfb1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:25:55 crc kubenswrapper[4632]: E0313 10:25:55.074634 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-kq8lc" podUID="8f916c05-f172-42b6-9b13-0c8d2058bfb1" Mar 13 10:25:55 crc kubenswrapper[4632]: I0313 10:25:55.348024 4632 scope.go:117] "RemoveContainer" containerID="f74cf11731f4fec2422112ef6bdd1e43cc133692a8363ef95d5bb5847ffb0fd1" Mar 13 10:25:55 crc kubenswrapper[4632]: I0313 10:25:55.548445 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7bdb5f7878-ng2k2"] Mar 13 10:25:55 crc kubenswrapper[4632]: I0313 10:25:55.700925 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-689764498d-rg7vt"] Mar 13 10:25:55 crc kubenswrapper[4632]: I0313 10:25:55.876605 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb5f7878-ng2k2" event={"ID":"3e4afb91-ce26-4325-89c9-2542da2ec48a","Type":"ContainerStarted","Data":"aaad122938f426786c8baabdc4555594b0ba0e55f0c39302b9bf84230f06cfd1"} Mar 13 10:25:55 crc kubenswrapper[4632]: E0313 10:25:55.944420 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-cinder-api:e43235cb19da04699a53f42b6a75afe9\\\"\"" pod="openstack/cinder-db-sync-kq8lc" podUID="8f916c05-f172-42b6-9b13-0c8d2058bfb1" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.091086 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-x8tq8"] Mar 13 10:25:56 crc kubenswrapper[4632]: W0313 10:25:56.095746 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8d0f662_d180_4137_8107_e465c5fb0621.slice/crio-58b9dfc8050bba09291b639a8d3d5cc84a9643af0afb63feda9e26973d06a678 WatchSource:0}: Error finding container 58b9dfc8050bba09291b639a8d3d5cc84a9643af0afb63feda9e26973d06a678: Status 404 returned error can't find the container with id 58b9dfc8050bba09291b639a8d3d5cc84a9643af0afb63feda9e26973d06a678 Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.205441 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.343808 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5db978f585-jtbcw"] Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.471166 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-64bdffbb5c-mpfvf"] Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.473028 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.492416 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.492701 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.505803 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-64bdffbb5c-mpfvf"] Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.631081 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-ovndb-tls-certs\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.636895 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-public-tls-certs\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.636965 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-httpd-config\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.637645 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-config\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.637684 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-internal-tls-certs\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.637745 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-combined-ca-bundle\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.637781 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbbzk\" (UniqueName: \"kubernetes.io/projected/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-kube-api-access-rbbzk\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.723147 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5b59dbc87f-7zwrj" podUID="7203640d-964c-4c28-8cc2-6a7ae27cdab3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.114:5353: i/o timeout" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.753360 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-config\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.766324 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-config\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.766736 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-internal-tls-certs\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.766899 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-combined-ca-bundle\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.766986 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbbzk\" (UniqueName: \"kubernetes.io/projected/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-kube-api-access-rbbzk\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.767116 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-ovndb-tls-certs\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.767337 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-public-tls-certs\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.767441 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-httpd-config\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.789968 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-public-tls-certs\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.795028 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5c86b4b888-l9574"] Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.803136 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-ovndb-tls-certs\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.820568 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5776d95bfc-hl9dv"] Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.833677 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-httpd-config\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.845767 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-combined-ca-bundle\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.846831 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-internal-tls-certs\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.927956 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbbzk\" (UniqueName: \"kubernetes.io/projected/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-kube-api-access-rbbzk\") pod \"neutron-64bdffbb5c-mpfvf\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.956212 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67d6b4b8f7-nrxn8" event={"ID":"95fe9a38-2b32-411e-9121-ad4cc32f159e","Type":"ContainerStarted","Data":"d9db78843b825b24c0eab6345b91a7657d2b3f0bb64d65b5dcc125b1edeeb022"} Mar 13 10:25:56 crc kubenswrapper[4632]: I0313 10:25:56.990256 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x8tq8" event={"ID":"d8d0f662-d180-4137-8107-e465c5fb0621","Type":"ContainerStarted","Data":"58b9dfc8050bba09291b639a8d3d5cc84a9643af0afb63feda9e26973d06a678"} Mar 13 10:25:57 crc kubenswrapper[4632]: I0313 10:25:57.029397 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-htnd9" event={"ID":"e92afa62-9c75-4e0e-92f4-76e57328d7a0","Type":"ContainerStarted","Data":"68a82ec143a93c9f66b6d5e73e70ead182bba11acadf06a0bc0700ee8971357d"} Mar 13 10:25:57 crc kubenswrapper[4632]: I0313 10:25:57.052793 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5db978f585-jtbcw" event={"ID":"4cf1d659-89cc-471b-8089-bc85f7ab3578","Type":"ContainerStarted","Data":"4e0560a7572c040fe38f67491b2340496562a7d42aebf98395ce318dd739fcfb"} Mar 13 10:25:57 crc kubenswrapper[4632]: I0313 10:25:57.065390 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-htnd9" podStartSLOduration=4.126363025 podStartE2EDuration="52.065376503s" podCreationTimestamp="2026-03-13 10:25:05 +0000 UTC" firstStartedPulling="2026-03-13 10:25:07.408847427 +0000 UTC m=+1281.431377560" lastFinishedPulling="2026-03-13 10:25:55.347860905 +0000 UTC m=+1329.370391038" observedRunningTime="2026-03-13 10:25:57.060789674 +0000 UTC m=+1331.083319807" watchObservedRunningTime="2026-03-13 10:25:57.065376503 +0000 UTC m=+1331.087906636" Mar 13 10:25:57 crc kubenswrapper[4632]: I0313 10:25:57.078299 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689764498d-rg7vt" event={"ID":"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c","Type":"ContainerStarted","Data":"3e7612a846cadd8420f2057181569ce83941e65dffcec61def9bcce804b35eef"} Mar 13 10:25:57 crc kubenswrapper[4632]: I0313 10:25:57.102804 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-7fvlk" event={"ID":"d722ddd7-e65d-44f7-a02d-18ddf126ccf5","Type":"ContainerStarted","Data":"3ef3ce34ce4d2a0d8d000d31874aca20b10c953ddde87f68a0b04979e69b8bae"} Mar 13 10:25:57 crc kubenswrapper[4632]: I0313 10:25:57.133494 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"270ebc10-986f-4473-8a5e-9094de34ae98","Type":"ContainerStarted","Data":"27c121915dbbdfc336d1bc55bed50eb5edaf76e1bc92f4f6b5e249f4ffe5098a"} Mar 13 10:25:57 crc kubenswrapper[4632]: I0313 10:25:57.151581 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-7fvlk" podStartSLOduration=5.195070397 podStartE2EDuration="53.15155934s" podCreationTimestamp="2026-03-13 10:25:04 +0000 UTC" firstStartedPulling="2026-03-13 10:25:07.529247871 +0000 UTC m=+1281.551778004" lastFinishedPulling="2026-03-13 10:25:55.485736814 +0000 UTC m=+1329.508266947" observedRunningTime="2026-03-13 10:25:57.13380559 +0000 UTC m=+1331.156335723" watchObservedRunningTime="2026-03-13 10:25:57.15155934 +0000 UTC m=+1331.174089473" Mar 13 10:25:57 crc kubenswrapper[4632]: I0313 10:25:57.303715 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:25:57 crc kubenswrapper[4632]: I0313 10:25:57.335986 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.137100 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.223520 4632 generic.go:334] "Generic (PLEG): container finished" podID="ff547198-2736-4059-8e66-e63ea9ce7345" containerID="dec2a00b325c16f4a1d001f23d5e8b1ffdb30f4c935f90c479b4c2928a1f9cbd" exitCode=0 Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.224485 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" event={"ID":"ff547198-2736-4059-8e66-e63ea9ce7345","Type":"ContainerDied","Data":"dec2a00b325c16f4a1d001f23d5e8b1ffdb30f4c935f90c479b4c2928a1f9cbd"} Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.224520 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" event={"ID":"ff547198-2736-4059-8e66-e63ea9ce7345","Type":"ContainerStarted","Data":"8db6fac31f3928e6490a77faa8cf72ab51791153ec4fce9dafd1cd9fb950c31f"} Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.244987 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-64bdffbb5c-mpfvf"] Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.296015 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689764498d-rg7vt" event={"ID":"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c","Type":"ContainerStarted","Data":"b6cdbfe3937cc8607d510b86100785b83eab81056229525992fbe23bfebc3c39"} Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.334021 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c86b4b888-l9574" event={"ID":"6d73a499-d334-4a7a-9783-640b98760672","Type":"ContainerStarted","Data":"8c839401b1db62da93454588496b8ab534c9e6313aa3bcb0003cb9137b63b2ca"} Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.334070 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c86b4b888-l9574" event={"ID":"6d73a499-d334-4a7a-9783-640b98760672","Type":"ContainerStarted","Data":"c201c6ed0f734df3747387db31697b083007f33831a2be5b5b4d93d97a61d2c9"} Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.336686 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"37dc6e5d-eb14-4cef-9451-7c567c6c9068","Type":"ContainerStarted","Data":"064fad76db398b97a6a04386f16fbe17c9bebd9b23d2f3264f42bd5bbfc7916f"} Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.348311 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb5f7878-ng2k2" event={"ID":"3e4afb91-ce26-4325-89c9-2542da2ec48a","Type":"ContainerStarted","Data":"0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589"} Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.400406 4632 generic.go:334] "Generic (PLEG): container finished" podID="4cf1d659-89cc-471b-8089-bc85f7ab3578" containerID="259722717f7860244e919e91ec8af7531cfacc106ccbd95a7ce0bb8509700a95" exitCode=0 Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.400504 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5db978f585-jtbcw" event={"ID":"4cf1d659-89cc-471b-8089-bc85f7ab3578","Type":"ContainerDied","Data":"259722717f7860244e919e91ec8af7531cfacc106ccbd95a7ce0bb8509700a95"} Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.435585 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zdgpw" event={"ID":"418cb883-abd1-46b4-957f-0a40f3e62297","Type":"ContainerStarted","Data":"3672f721f5cc963fe48f19a0fe26275ae0f1cbd82fd44ed2d6b14dcbb240be1d"} Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.474425 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-zdgpw" podStartSLOduration=4.941126353 podStartE2EDuration="53.47440765s" podCreationTimestamp="2026-03-13 10:25:05 +0000 UTC" firstStartedPulling="2026-03-13 10:25:07.184145893 +0000 UTC m=+1281.206676026" lastFinishedPulling="2026-03-13 10:25:55.71742719 +0000 UTC m=+1329.739957323" observedRunningTime="2026-03-13 10:25:58.470693111 +0000 UTC m=+1332.493223255" watchObservedRunningTime="2026-03-13 10:25:58.47440765 +0000 UTC m=+1332.496937783" Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.477234 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67d6b4b8f7-nrxn8" event={"ID":"95fe9a38-2b32-411e-9121-ad4cc32f159e","Type":"ContainerStarted","Data":"f8238ac2122bdce07c274a4f41c5a0d859a4162d57594e52444f5d2a425d1e7b"} Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.477379 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67d6b4b8f7-nrxn8" podUID="95fe9a38-2b32-411e-9121-ad4cc32f159e" containerName="horizon-log" containerID="cri-o://d9db78843b825b24c0eab6345b91a7657d2b3f0bb64d65b5dcc125b1edeeb022" gracePeriod=30 Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.477857 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-67d6b4b8f7-nrxn8" podUID="95fe9a38-2b32-411e-9121-ad4cc32f159e" containerName="horizon" containerID="cri-o://f8238ac2122bdce07c274a4f41c5a0d859a4162d57594e52444f5d2a425d1e7b" gracePeriod=30 Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.501602 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x8tq8" event={"ID":"d8d0f662-d180-4137-8107-e465c5fb0621","Type":"ContainerStarted","Data":"3b5385b113397b9418c59a941d2a27f232c7b0df4b245db65886e55380c57297"} Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.553058 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-67d6b4b8f7-nrxn8" podStartSLOduration=8.912867412 podStartE2EDuration="53.553034588s" podCreationTimestamp="2026-03-13 10:25:05 +0000 UTC" firstStartedPulling="2026-03-13 10:25:07.693195043 +0000 UTC m=+1281.715725176" lastFinishedPulling="2026-03-13 10:25:52.333362219 +0000 UTC m=+1326.355892352" observedRunningTime="2026-03-13 10:25:58.531367535 +0000 UTC m=+1332.553897668" watchObservedRunningTime="2026-03-13 10:25:58.553034588 +0000 UTC m=+1332.575564721" Mar 13 10:25:58 crc kubenswrapper[4632]: I0313 10:25:58.647963 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-x8tq8" podStartSLOduration=24.64792298 podStartE2EDuration="24.64792298s" podCreationTimestamp="2026-03-13 10:25:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:25:58.567603582 +0000 UTC m=+1332.590133715" watchObservedRunningTime="2026-03-13 10:25:58.64792298 +0000 UTC m=+1332.670453113" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.093020 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.204365 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-config\") pod \"4cf1d659-89cc-471b-8089-bc85f7ab3578\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.204527 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-ovsdbserver-nb\") pod \"4cf1d659-89cc-471b-8089-bc85f7ab3578\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.204733 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsqxq\" (UniqueName: \"kubernetes.io/projected/4cf1d659-89cc-471b-8089-bc85f7ab3578-kube-api-access-gsqxq\") pod \"4cf1d659-89cc-471b-8089-bc85f7ab3578\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.204793 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-dns-swift-storage-0\") pod \"4cf1d659-89cc-471b-8089-bc85f7ab3578\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.204843 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-ovsdbserver-sb\") pod \"4cf1d659-89cc-471b-8089-bc85f7ab3578\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.204924 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-dns-svc\") pod \"4cf1d659-89cc-471b-8089-bc85f7ab3578\" (UID: \"4cf1d659-89cc-471b-8089-bc85f7ab3578\") " Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.240553 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cf1d659-89cc-471b-8089-bc85f7ab3578-kube-api-access-gsqxq" (OuterVolumeSpecName: "kube-api-access-gsqxq") pod "4cf1d659-89cc-471b-8089-bc85f7ab3578" (UID: "4cf1d659-89cc-471b-8089-bc85f7ab3578"). InnerVolumeSpecName "kube-api-access-gsqxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.307329 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsqxq\" (UniqueName: \"kubernetes.io/projected/4cf1d659-89cc-471b-8089-bc85f7ab3578-kube-api-access-gsqxq\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.320249 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-config" (OuterVolumeSpecName: "config") pod "4cf1d659-89cc-471b-8089-bc85f7ab3578" (UID: "4cf1d659-89cc-471b-8089-bc85f7ab3578"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.320592 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "4cf1d659-89cc-471b-8089-bc85f7ab3578" (UID: "4cf1d659-89cc-471b-8089-bc85f7ab3578"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.322321 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "4cf1d659-89cc-471b-8089-bc85f7ab3578" (UID: "4cf1d659-89cc-471b-8089-bc85f7ab3578"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.336351 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "4cf1d659-89cc-471b-8089-bc85f7ab3578" (UID: "4cf1d659-89cc-471b-8089-bc85f7ab3578"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.372570 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "4cf1d659-89cc-471b-8089-bc85f7ab3578" (UID: "4cf1d659-89cc-471b-8089-bc85f7ab3578"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.409672 4632 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.409711 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.409726 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.409739 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.409750 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/4cf1d659-89cc-471b-8089-bc85f7ab3578-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.557734 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"37dc6e5d-eb14-4cef-9451-7c567c6c9068","Type":"ContainerStarted","Data":"4cf0cebf490653caa207d6e711c23d84bd0c1109c2c15d6d1a3ec573b2a4d48f"} Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.583819 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb5f7878-ng2k2" event={"ID":"3e4afb91-ce26-4325-89c9-2542da2ec48a","Type":"ContainerStarted","Data":"dc4a058f6feb7822333693352f32f5677ff03988b7b5b71005c85c4bf733b402"} Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.603336 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c86b4b888-l9574" event={"ID":"6d73a499-d334-4a7a-9783-640b98760672","Type":"ContainerStarted","Data":"e005b4f09b297f1fe00efd39c9534b7382173cd69b88dca5466ba89c0f3c0de7"} Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.604565 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.623164 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" event={"ID":"ff547198-2736-4059-8e66-e63ea9ce7345","Type":"ContainerStarted","Data":"1076485d4d02b6cacd1f94b4c459b88d5309d73c47777ad04b4bed1ee81eb7ff"} Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.624001 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.636931 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7bdb5f7878-ng2k2" podStartSLOduration=45.636912938 podStartE2EDuration="45.636912938s" podCreationTimestamp="2026-03-13 10:25:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:25:59.633762624 +0000 UTC m=+1333.656292767" watchObservedRunningTime="2026-03-13 10:25:59.636912938 +0000 UTC m=+1333.659443071" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.657790 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" podStartSLOduration=5.657773102 podStartE2EDuration="5.657773102s" podCreationTimestamp="2026-03-13 10:25:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:25:59.656146283 +0000 UTC m=+1333.678676426" watchObservedRunningTime="2026-03-13 10:25:59.657773102 +0000 UTC m=+1333.680303235" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.662270 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64bdffbb5c-mpfvf" event={"ID":"6c867fc1-05ed-46c3-99dc-71ef8a09dad3","Type":"ContainerStarted","Data":"027b2c4436a3d137f7ef6a7921904bf128e17aa7812143af60d4d11a546759da"} Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.662322 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64bdffbb5c-mpfvf" event={"ID":"6c867fc1-05ed-46c3-99dc-71ef8a09dad3","Type":"ContainerStarted","Data":"bb71081b64258f79a4055c8e129128f47654fe94235aa2a730194da521f70fe1"} Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.682729 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1756bbdc-3e6c-4815-96a7-0620f7400cb7","Type":"ContainerStarted","Data":"4958903559f9d5de9098d0d27c704deb245f10fc57158f3d446a7bff788fb121"} Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.698975 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5db978f585-jtbcw" event={"ID":"4cf1d659-89cc-471b-8089-bc85f7ab3578","Type":"ContainerDied","Data":"4e0560a7572c040fe38f67491b2340496562a7d42aebf98395ce318dd739fcfb"} Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.699040 4632 scope.go:117] "RemoveContainer" containerID="259722717f7860244e919e91ec8af7531cfacc106ccbd95a7ce0bb8509700a95" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.699202 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5db978f585-jtbcw" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.703290 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5c86b4b888-l9574" podStartSLOduration=6.703268546 podStartE2EDuration="6.703268546s" podCreationTimestamp="2026-03-13 10:25:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:25:59.682533256 +0000 UTC m=+1333.705063409" watchObservedRunningTime="2026-03-13 10:25:59.703268546 +0000 UTC m=+1333.725798679" Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.733070 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689764498d-rg7vt" event={"ID":"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c","Type":"ContainerStarted","Data":"8ce0185281fb59d0c6bda2b2c484ad3711b4bd3b729b4b8677e75ca6b8e1f739"} Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.810980 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5db978f585-jtbcw"] Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.860692 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5db978f585-jtbcw"] Mar 13 10:25:59 crc kubenswrapper[4632]: I0313 10:25:59.869646 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-689764498d-rg7vt" podStartSLOduration=44.869622229 podStartE2EDuration="44.869622229s" podCreationTimestamp="2026-03-13 10:25:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:25:59.805676557 +0000 UTC m=+1333.828206690" watchObservedRunningTime="2026-03-13 10:25:59.869622229 +0000 UTC m=+1333.892152372" Mar 13 10:26:00 crc kubenswrapper[4632]: I0313 10:26:00.058293 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cf1d659-89cc-471b-8089-bc85f7ab3578" path="/var/lib/kubelet/pods/4cf1d659-89cc-471b-8089-bc85f7ab3578/volumes" Mar 13 10:26:00 crc kubenswrapper[4632]: I0313 10:26:00.225045 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556626-z45rd"] Mar 13 10:26:00 crc kubenswrapper[4632]: E0313 10:26:00.225629 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cf1d659-89cc-471b-8089-bc85f7ab3578" containerName="init" Mar 13 10:26:00 crc kubenswrapper[4632]: I0313 10:26:00.225649 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cf1d659-89cc-471b-8089-bc85f7ab3578" containerName="init" Mar 13 10:26:00 crc kubenswrapper[4632]: I0313 10:26:00.225964 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cf1d659-89cc-471b-8089-bc85f7ab3578" containerName="init" Mar 13 10:26:00 crc kubenswrapper[4632]: I0313 10:26:00.226734 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556626-z45rd" Mar 13 10:26:00 crc kubenswrapper[4632]: I0313 10:26:00.229323 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:26:00 crc kubenswrapper[4632]: I0313 10:26:00.231570 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:26:00 crc kubenswrapper[4632]: I0313 10:26:00.232570 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:26:00 crc kubenswrapper[4632]: I0313 10:26:00.249057 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556626-z45rd"] Mar 13 10:26:00 crc kubenswrapper[4632]: I0313 10:26:00.363500 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9ghq\" (UniqueName: \"kubernetes.io/projected/8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27-kube-api-access-r9ghq\") pod \"auto-csr-approver-29556626-z45rd\" (UID: \"8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27\") " pod="openshift-infra/auto-csr-approver-29556626-z45rd" Mar 13 10:26:00 crc kubenswrapper[4632]: I0313 10:26:00.465225 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9ghq\" (UniqueName: \"kubernetes.io/projected/8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27-kube-api-access-r9ghq\") pod \"auto-csr-approver-29556626-z45rd\" (UID: \"8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27\") " pod="openshift-infra/auto-csr-approver-29556626-z45rd" Mar 13 10:26:00 crc kubenswrapper[4632]: I0313 10:26:00.512043 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9ghq\" (UniqueName: \"kubernetes.io/projected/8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27-kube-api-access-r9ghq\") pod \"auto-csr-approver-29556626-z45rd\" (UID: \"8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27\") " pod="openshift-infra/auto-csr-approver-29556626-z45rd" Mar 13 10:26:00 crc kubenswrapper[4632]: I0313 10:26:00.562262 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556626-z45rd" Mar 13 10:26:00 crc kubenswrapper[4632]: I0313 10:26:00.746754 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1756bbdc-3e6c-4815-96a7-0620f7400cb7","Type":"ContainerStarted","Data":"7d8fa15092ec71e6c12fe0e4bfd626668295f5b687014027b1d5515acb53e02d"} Mar 13 10:26:04 crc kubenswrapper[4632]: I0313 10:26:04.264418 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556626-z45rd"] Mar 13 10:26:04 crc kubenswrapper[4632]: I0313 10:26:04.450108 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:26:04 crc kubenswrapper[4632]: I0313 10:26:04.584780 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d8f9dd5cc-6nktg"] Mar 13 10:26:04 crc kubenswrapper[4632]: I0313 10:26:04.595470 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" podUID="78e29b83-b50e-46db-a8d6-bba0ecfb5c08" containerName="dnsmasq-dns" containerID="cri-o://0f3b47ae46ac068badd9fe9f0befa9613632a7901ef641ad38d1419cf04cc4af" gracePeriod=10 Mar 13 10:26:04 crc kubenswrapper[4632]: I0313 10:26:04.879527 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64bdffbb5c-mpfvf" event={"ID":"6c867fc1-05ed-46c3-99dc-71ef8a09dad3","Type":"ContainerStarted","Data":"bbc256375bc79a61ff656574ec8a596aed3314e7ad4cd2f7fcf6a7462aee3274"} Mar 13 10:26:04 crc kubenswrapper[4632]: I0313 10:26:04.880263 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:26:04 crc kubenswrapper[4632]: I0313 10:26:04.889079 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556626-z45rd" event={"ID":"8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27","Type":"ContainerStarted","Data":"6d61d4f7c0c3be011ec1f0f84978bb250c298a5c766226d624254ef183165b94"} Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.394483 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.395429 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.523806 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.546522 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-64bdffbb5c-mpfvf" podStartSLOduration=9.546502741 podStartE2EDuration="9.546502741s" podCreationTimestamp="2026-03-13 10:25:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:04.920661708 +0000 UTC m=+1338.943191841" watchObservedRunningTime="2026-03-13 10:26:05.546502741 +0000 UTC m=+1339.569032874" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.694669 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29zt6\" (UniqueName: \"kubernetes.io/projected/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-kube-api-access-29zt6\") pod \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.694751 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-config\") pod \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.694794 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-dns-swift-storage-0\") pod \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.694830 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-ovsdbserver-nb\") pod \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.695021 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-dns-svc\") pod \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.695082 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-ovsdbserver-sb\") pod \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\" (UID: \"78e29b83-b50e-46db-a8d6-bba0ecfb5c08\") " Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.704299 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-kube-api-access-29zt6" (OuterVolumeSpecName: "kube-api-access-29zt6") pod "78e29b83-b50e-46db-a8d6-bba0ecfb5c08" (UID: "78e29b83-b50e-46db-a8d6-bba0ecfb5c08"). InnerVolumeSpecName "kube-api-access-29zt6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.797530 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29zt6\" (UniqueName: \"kubernetes.io/projected/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-kube-api-access-29zt6\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.812003 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "78e29b83-b50e-46db-a8d6-bba0ecfb5c08" (UID: "78e29b83-b50e-46db-a8d6-bba0ecfb5c08"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.812347 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "78e29b83-b50e-46db-a8d6-bba0ecfb5c08" (UID: "78e29b83-b50e-46db-a8d6-bba0ecfb5c08"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.856696 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.858334 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.862698 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "78e29b83-b50e-46db-a8d6-bba0ecfb5c08" (UID: "78e29b83-b50e-46db-a8d6-bba0ecfb5c08"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.881926 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "78e29b83-b50e-46db-a8d6-bba0ecfb5c08" (UID: "78e29b83-b50e-46db-a8d6-bba0ecfb5c08"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.885639 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-config" (OuterVolumeSpecName: "config") pod "78e29b83-b50e-46db-a8d6-bba0ecfb5c08" (UID: "78e29b83-b50e-46db-a8d6-bba0ecfb5c08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.900685 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.900719 4632 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.900735 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.900749 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.900759 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/78e29b83-b50e-46db-a8d6-bba0ecfb5c08-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.917549 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"37dc6e5d-eb14-4cef-9451-7c567c6c9068","Type":"ContainerStarted","Data":"82cadbe6c9dcc57c6c514bcdc80e74daabbed007f9160e45e58a214195d92a1e"} Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.917718 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="37dc6e5d-eb14-4cef-9451-7c567c6c9068" containerName="glance-log" containerID="cri-o://4cf0cebf490653caa207d6e711c23d84bd0c1109c2c15d6d1a3ec573b2a4d48f" gracePeriod=30 Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.918153 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="37dc6e5d-eb14-4cef-9451-7c567c6c9068" containerName="glance-httpd" containerID="cri-o://82cadbe6c9dcc57c6c514bcdc80e74daabbed007f9160e45e58a214195d92a1e" gracePeriod=30 Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.937688 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"270ebc10-986f-4473-8a5e-9094de34ae98","Type":"ContainerStarted","Data":"b63cc4f80efbb7b17b044808a5b6c8d5aa98b9e2ae8e38ab95a55c4e3ba911d1"} Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.948211 4632 generic.go:334] "Generic (PLEG): container finished" podID="78e29b83-b50e-46db-a8d6-bba0ecfb5c08" containerID="0f3b47ae46ac068badd9fe9f0befa9613632a7901ef641ad38d1419cf04cc4af" exitCode=0 Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.948279 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" event={"ID":"78e29b83-b50e-46db-a8d6-bba0ecfb5c08","Type":"ContainerDied","Data":"0f3b47ae46ac068badd9fe9f0befa9613632a7901ef641ad38d1419cf04cc4af"} Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.948308 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" event={"ID":"78e29b83-b50e-46db-a8d6-bba0ecfb5c08","Type":"ContainerDied","Data":"5187e6e9a0835d8922aa8452723fd7620bf5222c8a96f16a5be9778d8386494d"} Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.948325 4632 scope.go:117] "RemoveContainer" containerID="0f3b47ae46ac068badd9fe9f0befa9613632a7901ef641ad38d1419cf04cc4af" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.948474 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d8f9dd5cc-6nktg" Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.980078 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1756bbdc-3e6c-4815-96a7-0620f7400cb7" containerName="glance-log" containerID="cri-o://7d8fa15092ec71e6c12fe0e4bfd626668295f5b687014027b1d5515acb53e02d" gracePeriod=30 Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.980656 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1756bbdc-3e6c-4815-96a7-0620f7400cb7","Type":"ContainerStarted","Data":"859ff52dd31106c27267e2e88fdd9f3088d59cb66aea2491829bb3b779d0c030"} Mar 13 10:26:05 crc kubenswrapper[4632]: I0313 10:26:05.983252 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="1756bbdc-3e6c-4815-96a7-0620f7400cb7" containerName="glance-httpd" containerID="cri-o://859ff52dd31106c27267e2e88fdd9f3088d59cb66aea2491829bb3b779d0c030" gracePeriod=30 Mar 13 10:26:06 crc kubenswrapper[4632]: I0313 10:26:06.014798 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=29.01478288 podStartE2EDuration="29.01478288s" podCreationTimestamp="2026-03-13 10:25:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:05.960259601 +0000 UTC m=+1339.982789754" watchObservedRunningTime="2026-03-13 10:26:06.01478288 +0000 UTC m=+1340.037313013" Mar 13 10:26:06 crc kubenswrapper[4632]: I0313 10:26:06.019459 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=28.01944629 podStartE2EDuration="28.01944629s" podCreationTimestamp="2026-03-13 10:25:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:06.010012007 +0000 UTC m=+1340.032542160" watchObservedRunningTime="2026-03-13 10:26:06.01944629 +0000 UTC m=+1340.041976423" Mar 13 10:26:06 crc kubenswrapper[4632]: I0313 10:26:06.086697 4632 scope.go:117] "RemoveContainer" containerID="80dc69ee9ec968911aeb73d01429c087d3144584fba19b07c0a8b37e75187f19" Mar 13 10:26:06 crc kubenswrapper[4632]: I0313 10:26:06.115493 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d8f9dd5cc-6nktg"] Mar 13 10:26:06 crc kubenswrapper[4632]: I0313 10:26:06.127303 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d8f9dd5cc-6nktg"] Mar 13 10:26:06 crc kubenswrapper[4632]: I0313 10:26:06.406587 4632 scope.go:117] "RemoveContainer" containerID="0f3b47ae46ac068badd9fe9f0befa9613632a7901ef641ad38d1419cf04cc4af" Mar 13 10:26:06 crc kubenswrapper[4632]: E0313 10:26:06.409756 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f3b47ae46ac068badd9fe9f0befa9613632a7901ef641ad38d1419cf04cc4af\": container with ID starting with 0f3b47ae46ac068badd9fe9f0befa9613632a7901ef641ad38d1419cf04cc4af not found: ID does not exist" containerID="0f3b47ae46ac068badd9fe9f0befa9613632a7901ef641ad38d1419cf04cc4af" Mar 13 10:26:06 crc kubenswrapper[4632]: I0313 10:26:06.409830 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f3b47ae46ac068badd9fe9f0befa9613632a7901ef641ad38d1419cf04cc4af"} err="failed to get container status \"0f3b47ae46ac068badd9fe9f0befa9613632a7901ef641ad38d1419cf04cc4af\": rpc error: code = NotFound desc = could not find container \"0f3b47ae46ac068badd9fe9f0befa9613632a7901ef641ad38d1419cf04cc4af\": container with ID starting with 0f3b47ae46ac068badd9fe9f0befa9613632a7901ef641ad38d1419cf04cc4af not found: ID does not exist" Mar 13 10:26:06 crc kubenswrapper[4632]: I0313 10:26:06.409876 4632 scope.go:117] "RemoveContainer" containerID="80dc69ee9ec968911aeb73d01429c087d3144584fba19b07c0a8b37e75187f19" Mar 13 10:26:06 crc kubenswrapper[4632]: E0313 10:26:06.410640 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80dc69ee9ec968911aeb73d01429c087d3144584fba19b07c0a8b37e75187f19\": container with ID starting with 80dc69ee9ec968911aeb73d01429c087d3144584fba19b07c0a8b37e75187f19 not found: ID does not exist" containerID="80dc69ee9ec968911aeb73d01429c087d3144584fba19b07c0a8b37e75187f19" Mar 13 10:26:06 crc kubenswrapper[4632]: I0313 10:26:06.410671 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80dc69ee9ec968911aeb73d01429c087d3144584fba19b07c0a8b37e75187f19"} err="failed to get container status \"80dc69ee9ec968911aeb73d01429c087d3144584fba19b07c0a8b37e75187f19\": rpc error: code = NotFound desc = could not find container \"80dc69ee9ec968911aeb73d01429c087d3144584fba19b07c0a8b37e75187f19\": container with ID starting with 80dc69ee9ec968911aeb73d01429c087d3144584fba19b07c0a8b37e75187f19 not found: ID does not exist" Mar 13 10:26:06 crc kubenswrapper[4632]: I0313 10:26:06.618142 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.086109 4632 generic.go:334] "Generic (PLEG): container finished" podID="1756bbdc-3e6c-4815-96a7-0620f7400cb7" containerID="859ff52dd31106c27267e2e88fdd9f3088d59cb66aea2491829bb3b779d0c030" exitCode=143 Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.086562 4632 generic.go:334] "Generic (PLEG): container finished" podID="1756bbdc-3e6c-4815-96a7-0620f7400cb7" containerID="7d8fa15092ec71e6c12fe0e4bfd626668295f5b687014027b1d5515acb53e02d" exitCode=143 Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.086636 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1756bbdc-3e6c-4815-96a7-0620f7400cb7","Type":"ContainerDied","Data":"859ff52dd31106c27267e2e88fdd9f3088d59cb66aea2491829bb3b779d0c030"} Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.086664 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1756bbdc-3e6c-4815-96a7-0620f7400cb7","Type":"ContainerDied","Data":"7d8fa15092ec71e6c12fe0e4bfd626668295f5b687014027b1d5515acb53e02d"} Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.099036 4632 generic.go:334] "Generic (PLEG): container finished" podID="37dc6e5d-eb14-4cef-9451-7c567c6c9068" containerID="82cadbe6c9dcc57c6c514bcdc80e74daabbed007f9160e45e58a214195d92a1e" exitCode=143 Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.099072 4632 generic.go:334] "Generic (PLEG): container finished" podID="37dc6e5d-eb14-4cef-9451-7c567c6c9068" containerID="4cf0cebf490653caa207d6e711c23d84bd0c1109c2c15d6d1a3ec573b2a4d48f" exitCode=143 Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.099118 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"37dc6e5d-eb14-4cef-9451-7c567c6c9068","Type":"ContainerDied","Data":"82cadbe6c9dcc57c6c514bcdc80e74daabbed007f9160e45e58a214195d92a1e"} Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.099152 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"37dc6e5d-eb14-4cef-9451-7c567c6c9068","Type":"ContainerDied","Data":"4cf0cebf490653caa207d6e711c23d84bd0c1109c2c15d6d1a3ec573b2a4d48f"} Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.101994 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556626-z45rd" event={"ID":"8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27","Type":"ContainerStarted","Data":"e8fc7f9526396e3f4333f93ccef86f72aee3214939c63a5e8145c990bbf9d938"} Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.137216 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556626-z45rd" podStartSLOduration=6.105890334 podStartE2EDuration="7.137198552s" podCreationTimestamp="2026-03-13 10:26:00 +0000 UTC" firstStartedPulling="2026-03-13 10:26:04.256514908 +0000 UTC m=+1338.279045041" lastFinishedPulling="2026-03-13 10:26:05.287823126 +0000 UTC m=+1339.310353259" observedRunningTime="2026-03-13 10:26:07.132212913 +0000 UTC m=+1341.154743046" watchObservedRunningTime="2026-03-13 10:26:07.137198552 +0000 UTC m=+1341.159728685" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.323778 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.345966 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.452138 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.452225 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-combined-ca-bundle\") pod \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.452282 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37dc6e5d-eb14-4cef-9451-7c567c6c9068-httpd-run\") pod \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.452311 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-scripts\") pod \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.452340 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jn72\" (UniqueName: \"kubernetes.io/projected/1756bbdc-3e6c-4815-96a7-0620f7400cb7-kube-api-access-7jn72\") pod \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.452367 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rqnd\" (UniqueName: \"kubernetes.io/projected/37dc6e5d-eb14-4cef-9451-7c567c6c9068-kube-api-access-5rqnd\") pod \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.452391 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1756bbdc-3e6c-4815-96a7-0620f7400cb7-logs\") pod \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.452431 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-combined-ca-bundle\") pod \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.452481 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-scripts\") pod \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.452542 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37dc6e5d-eb14-4cef-9451-7c567c6c9068-logs\") pod \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.452581 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-config-data\") pod \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\" (UID: \"37dc6e5d-eb14-4cef-9451-7c567c6c9068\") " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.452601 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.452646 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1756bbdc-3e6c-4815-96a7-0620f7400cb7-httpd-run\") pod \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.452663 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-config-data\") pod \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\" (UID: \"1756bbdc-3e6c-4815-96a7-0620f7400cb7\") " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.453541 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1756bbdc-3e6c-4815-96a7-0620f7400cb7-logs" (OuterVolumeSpecName: "logs") pod "1756bbdc-3e6c-4815-96a7-0620f7400cb7" (UID: "1756bbdc-3e6c-4815-96a7-0620f7400cb7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.454105 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37dc6e5d-eb14-4cef-9451-7c567c6c9068-logs" (OuterVolumeSpecName: "logs") pod "37dc6e5d-eb14-4cef-9451-7c567c6c9068" (UID: "37dc6e5d-eb14-4cef-9451-7c567c6c9068"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.465227 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1756bbdc-3e6c-4815-96a7-0620f7400cb7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1756bbdc-3e6c-4815-96a7-0620f7400cb7" (UID: "1756bbdc-3e6c-4815-96a7-0620f7400cb7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.495539 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/37dc6e5d-eb14-4cef-9451-7c567c6c9068-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "37dc6e5d-eb14-4cef-9451-7c567c6c9068" (UID: "37dc6e5d-eb14-4cef-9451-7c567c6c9068"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.510133 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "1756bbdc-3e6c-4815-96a7-0620f7400cb7" (UID: "1756bbdc-3e6c-4815-96a7-0620f7400cb7"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.510270 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "37dc6e5d-eb14-4cef-9451-7c567c6c9068" (UID: "37dc6e5d-eb14-4cef-9451-7c567c6c9068"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.535310 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1756bbdc-3e6c-4815-96a7-0620f7400cb7-kube-api-access-7jn72" (OuterVolumeSpecName: "kube-api-access-7jn72") pod "1756bbdc-3e6c-4815-96a7-0620f7400cb7" (UID: "1756bbdc-3e6c-4815-96a7-0620f7400cb7"). InnerVolumeSpecName "kube-api-access-7jn72". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.535336 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-scripts" (OuterVolumeSpecName: "scripts") pod "37dc6e5d-eb14-4cef-9451-7c567c6c9068" (UID: "37dc6e5d-eb14-4cef-9451-7c567c6c9068"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.537105 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-scripts" (OuterVolumeSpecName: "scripts") pod "1756bbdc-3e6c-4815-96a7-0620f7400cb7" (UID: "1756bbdc-3e6c-4815-96a7-0620f7400cb7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.549535 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/37dc6e5d-eb14-4cef-9451-7c567c6c9068-kube-api-access-5rqnd" (OuterVolumeSpecName: "kube-api-access-5rqnd") pod "37dc6e5d-eb14-4cef-9451-7c567c6c9068" (UID: "37dc6e5d-eb14-4cef-9451-7c567c6c9068"). InnerVolumeSpecName "kube-api-access-5rqnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.554711 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/37dc6e5d-eb14-4cef-9451-7c567c6c9068-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.554780 4632 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.554807 4632 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1756bbdc-3e6c-4815-96a7-0620f7400cb7-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.554820 4632 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.554830 4632 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/37dc6e5d-eb14-4cef-9451-7c567c6c9068-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.554840 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.554849 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jn72\" (UniqueName: \"kubernetes.io/projected/1756bbdc-3e6c-4815-96a7-0620f7400cb7-kube-api-access-7jn72\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.554860 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rqnd\" (UniqueName: \"kubernetes.io/projected/37dc6e5d-eb14-4cef-9451-7c567c6c9068-kube-api-access-5rqnd\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.554868 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1756bbdc-3e6c-4815-96a7-0620f7400cb7-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.554890 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.588531 4632 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.601293 4632 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.802577 4632 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.802611 4632 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.878133 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-config-data" (OuterVolumeSpecName: "config-data") pod "37dc6e5d-eb14-4cef-9451-7c567c6c9068" (UID: "37dc6e5d-eb14-4cef-9451-7c567c6c9068"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.894358 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "37dc6e5d-eb14-4cef-9451-7c567c6c9068" (UID: "37dc6e5d-eb14-4cef-9451-7c567c6c9068"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.915198 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.915264 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/37dc6e5d-eb14-4cef-9451-7c567c6c9068-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.936791 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1756bbdc-3e6c-4815-96a7-0620f7400cb7" (UID: "1756bbdc-3e6c-4815-96a7-0620f7400cb7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:07 crc kubenswrapper[4632]: I0313 10:26:07.943099 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-config-data" (OuterVolumeSpecName: "config-data") pod "1756bbdc-3e6c-4815-96a7-0620f7400cb7" (UID: "1756bbdc-3e6c-4815-96a7-0620f7400cb7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.017441 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.017482 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1756bbdc-3e6c-4815-96a7-0620f7400cb7-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.077170 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78e29b83-b50e-46db-a8d6-bba0ecfb5c08" path="/var/lib/kubelet/pods/78e29b83-b50e-46db-a8d6-bba0ecfb5c08/volumes" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.186577 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"37dc6e5d-eb14-4cef-9451-7c567c6c9068","Type":"ContainerDied","Data":"064fad76db398b97a6a04386f16fbe17c9bebd9b23d2f3264f42bd5bbfc7916f"} Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.186655 4632 scope.go:117] "RemoveContainer" containerID="82cadbe6c9dcc57c6c514bcdc80e74daabbed007f9160e45e58a214195d92a1e" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.186925 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.256738 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.260589 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"1756bbdc-3e6c-4815-96a7-0620f7400cb7","Type":"ContainerDied","Data":"4958903559f9d5de9098d0d27c704deb245f10fc57158f3d446a7bff788fb121"} Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.300720 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.337191 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.381014 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.465419 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.497912 4632 scope.go:117] "RemoveContainer" containerID="4cf0cebf490653caa207d6e711c23d84bd0c1109c2c15d6d1a3ec573b2a4d48f" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.511852 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:26:08 crc kubenswrapper[4632]: E0313 10:26:08.529790 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78e29b83-b50e-46db-a8d6-bba0ecfb5c08" containerName="dnsmasq-dns" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.529843 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="78e29b83-b50e-46db-a8d6-bba0ecfb5c08" containerName="dnsmasq-dns" Mar 13 10:26:08 crc kubenswrapper[4632]: E0313 10:26:08.529862 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78e29b83-b50e-46db-a8d6-bba0ecfb5c08" containerName="init" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.529871 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="78e29b83-b50e-46db-a8d6-bba0ecfb5c08" containerName="init" Mar 13 10:26:08 crc kubenswrapper[4632]: E0313 10:26:08.529890 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37dc6e5d-eb14-4cef-9451-7c567c6c9068" containerName="glance-httpd" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.529897 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="37dc6e5d-eb14-4cef-9451-7c567c6c9068" containerName="glance-httpd" Mar 13 10:26:08 crc kubenswrapper[4632]: E0313 10:26:08.529913 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="37dc6e5d-eb14-4cef-9451-7c567c6c9068" containerName="glance-log" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.529920 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="37dc6e5d-eb14-4cef-9451-7c567c6c9068" containerName="glance-log" Mar 13 10:26:08 crc kubenswrapper[4632]: E0313 10:26:08.529955 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1756bbdc-3e6c-4815-96a7-0620f7400cb7" containerName="glance-log" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.529962 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="1756bbdc-3e6c-4815-96a7-0620f7400cb7" containerName="glance-log" Mar 13 10:26:08 crc kubenswrapper[4632]: E0313 10:26:08.529979 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1756bbdc-3e6c-4815-96a7-0620f7400cb7" containerName="glance-httpd" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.529987 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="1756bbdc-3e6c-4815-96a7-0620f7400cb7" containerName="glance-httpd" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.530235 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="1756bbdc-3e6c-4815-96a7-0620f7400cb7" containerName="glance-log" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.530267 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="1756bbdc-3e6c-4815-96a7-0620f7400cb7" containerName="glance-httpd" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.530285 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="37dc6e5d-eb14-4cef-9451-7c567c6c9068" containerName="glance-log" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.530298 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="37dc6e5d-eb14-4cef-9451-7c567c6c9068" containerName="glance-httpd" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.530315 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="78e29b83-b50e-46db-a8d6-bba0ecfb5c08" containerName="dnsmasq-dns" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.531740 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.549536 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.582305 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.582567 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.582664 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.582971 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-qpd5p" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.604502 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.621901 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.629355 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.629665 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.658222 4632 scope.go:117] "RemoveContainer" containerID="859ff52dd31106c27267e2e88fdd9f3088d59cb66aea2491829bb3b779d0c030" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.668558 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.689673 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.689749 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbnq2\" (UniqueName: \"kubernetes.io/projected/62c1f3f8-e898-4481-88e0-49f0c20228a4-kube-api-access-gbnq2\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.689814 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-config-data\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.689843 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62c1f3f8-e898-4481-88e0-49f0c20228a4-logs\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.689897 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.689990 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.690009 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-scripts\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.690070 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/62c1f3f8-e898-4481-88e0-49f0c20228a4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.792648 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/050df504-63b9-4453-be2b-f3b0315fb801-logs\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.792763 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/62c1f3f8-e898-4481-88e0-49f0c20228a4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.792815 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.792839 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/050df504-63b9-4453-be2b-f3b0315fb801-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.792870 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.792908 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5bl8\" (UniqueName: \"kubernetes.io/projected/050df504-63b9-4453-be2b-f3b0315fb801-kube-api-access-z5bl8\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.792961 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.793049 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbnq2\" (UniqueName: \"kubernetes.io/projected/62c1f3f8-e898-4481-88e0-49f0c20228a4-kube-api-access-gbnq2\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.793077 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-config-data\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.793104 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62c1f3f8-e898-4481-88e0-49f0c20228a4-logs\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.793159 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-scripts\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.793195 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.793275 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.793329 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-config-data\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.793360 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.793390 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-scripts\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.795072 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/62c1f3f8-e898-4481-88e0-49f0c20228a4-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.795808 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.803417 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62c1f3f8-e898-4481-88e0-49f0c20228a4-logs\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.818398 4632 scope.go:117] "RemoveContainer" containerID="7d8fa15092ec71e6c12fe0e4bfd626668295f5b687014027b1d5515acb53e02d" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.843919 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.845622 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.858302 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbnq2\" (UniqueName: \"kubernetes.io/projected/62c1f3f8-e898-4481-88e0-49f0c20228a4-kube-api-access-gbnq2\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.886513 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-scripts\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.888228 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-config-data\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.892856 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " pod="openstack/glance-default-external-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.894589 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/050df504-63b9-4453-be2b-f3b0315fb801-logs\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.894658 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.894676 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/050df504-63b9-4453-be2b-f3b0315fb801-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.894696 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.894718 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5bl8\" (UniqueName: \"kubernetes.io/projected/050df504-63b9-4453-be2b-f3b0315fb801-kube-api-access-z5bl8\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.894793 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-scripts\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.894849 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.894879 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-config-data\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.900791 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/050df504-63b9-4453-be2b-f3b0315fb801-logs\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.901255 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.908957 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/050df504-63b9-4453-be2b-f3b0315fb801-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.925196 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.926179 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-config-data\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.966906 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.967721 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-scripts\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.986500 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5bl8\" (UniqueName: \"kubernetes.io/projected/050df504-63b9-4453-be2b-f3b0315fb801-kube-api-access-z5bl8\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:08 crc kubenswrapper[4632]: I0313 10:26:08.990296 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:26:09 crc kubenswrapper[4632]: I0313 10:26:09.183037 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 10:26:09 crc kubenswrapper[4632]: I0313 10:26:09.284035 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 10:26:10 crc kubenswrapper[4632]: I0313 10:26:10.103506 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1756bbdc-3e6c-4815-96a7-0620f7400cb7" path="/var/lib/kubelet/pods/1756bbdc-3e6c-4815-96a7-0620f7400cb7/volumes" Mar 13 10:26:10 crc kubenswrapper[4632]: I0313 10:26:10.106002 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="37dc6e5d-eb14-4cef-9451-7c567c6c9068" path="/var/lib/kubelet/pods/37dc6e5d-eb14-4cef-9451-7c567c6c9068/volumes" Mar 13 10:26:10 crc kubenswrapper[4632]: I0313 10:26:10.427830 4632 generic.go:334] "Generic (PLEG): container finished" podID="8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27" containerID="e8fc7f9526396e3f4333f93ccef86f72aee3214939c63a5e8145c990bbf9d938" exitCode=0 Mar 13 10:26:10 crc kubenswrapper[4632]: I0313 10:26:10.427904 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556626-z45rd" event={"ID":"8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27","Type":"ContainerDied","Data":"e8fc7f9526396e3f4333f93ccef86f72aee3214939c63a5e8145c990bbf9d938"} Mar 13 10:26:10 crc kubenswrapper[4632]: I0313 10:26:10.647440 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:26:10 crc kubenswrapper[4632]: W0313 10:26:10.661084 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62c1f3f8_e898_4481_88e0_49f0c20228a4.slice/crio-da67a58c5a020c95fb415df6f51542675c8d6697cd1fafcacdcc7d6081f0a9ff WatchSource:0}: Error finding container da67a58c5a020c95fb415df6f51542675c8d6697cd1fafcacdcc7d6081f0a9ff: Status 404 returned error can't find the container with id da67a58c5a020c95fb415df6f51542675c8d6697cd1fafcacdcc7d6081f0a9ff Mar 13 10:26:10 crc kubenswrapper[4632]: I0313 10:26:10.860403 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:26:11 crc kubenswrapper[4632]: I0313 10:26:11.480752 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kq8lc" event={"ID":"8f916c05-f172-42b6-9b13-0c8d2058bfb1","Type":"ContainerStarted","Data":"6d5ac5d7a6aab5517e4300c2e14808710d4f8cfa4977c9841f6552b262144012"} Mar 13 10:26:11 crc kubenswrapper[4632]: I0313 10:26:11.493927 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"62c1f3f8-e898-4481-88e0-49f0c20228a4","Type":"ContainerStarted","Data":"da67a58c5a020c95fb415df6f51542675c8d6697cd1fafcacdcc7d6081f0a9ff"} Mar 13 10:26:11 crc kubenswrapper[4632]: I0313 10:26:11.497790 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"050df504-63b9-4453-be2b-f3b0315fb801","Type":"ContainerStarted","Data":"639dbfcf9c85b2d6df276ce37ddb572204028d4a54aa36f7c4d3026c9ff6abfc"} Mar 13 10:26:11 crc kubenswrapper[4632]: I0313 10:26:11.521994 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-kq8lc" podStartSLOduration=7.33701618 podStartE2EDuration="1m7.521970178s" podCreationTimestamp="2026-03-13 10:25:04 +0000 UTC" firstStartedPulling="2026-03-13 10:25:07.27532772 +0000 UTC m=+1281.297857853" lastFinishedPulling="2026-03-13 10:26:07.460281718 +0000 UTC m=+1341.482811851" observedRunningTime="2026-03-13 10:26:11.515435574 +0000 UTC m=+1345.537965717" watchObservedRunningTime="2026-03-13 10:26:11.521970178 +0000 UTC m=+1345.544500331" Mar 13 10:26:12 crc kubenswrapper[4632]: I0313 10:26:12.003170 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556626-z45rd" Mar 13 10:26:12 crc kubenswrapper[4632]: I0313 10:26:12.164590 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r9ghq\" (UniqueName: \"kubernetes.io/projected/8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27-kube-api-access-r9ghq\") pod \"8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27\" (UID: \"8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27\") " Mar 13 10:26:12 crc kubenswrapper[4632]: I0313 10:26:12.198223 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27-kube-api-access-r9ghq" (OuterVolumeSpecName: "kube-api-access-r9ghq") pod "8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27" (UID: "8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27"). InnerVolumeSpecName "kube-api-access-r9ghq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:12 crc kubenswrapper[4632]: I0313 10:26:12.283223 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r9ghq\" (UniqueName: \"kubernetes.io/projected/8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27-kube-api-access-r9ghq\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:12 crc kubenswrapper[4632]: I0313 10:26:12.533623 4632 generic.go:334] "Generic (PLEG): container finished" podID="e92afa62-9c75-4e0e-92f4-76e57328d7a0" containerID="68a82ec143a93c9f66b6d5e73e70ead182bba11acadf06a0bc0700ee8971357d" exitCode=0 Mar 13 10:26:12 crc kubenswrapper[4632]: I0313 10:26:12.533743 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-htnd9" event={"ID":"e92afa62-9c75-4e0e-92f4-76e57328d7a0","Type":"ContainerDied","Data":"68a82ec143a93c9f66b6d5e73e70ead182bba11acadf06a0bc0700ee8971357d"} Mar 13 10:26:12 crc kubenswrapper[4632]: I0313 10:26:12.557645 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"62c1f3f8-e898-4481-88e0-49f0c20228a4","Type":"ContainerStarted","Data":"75e2995816c15269a0e0bb8513c4f7b9cace1b33dd417df2fc8f694c18b89fa0"} Mar 13 10:26:12 crc kubenswrapper[4632]: I0313 10:26:12.586227 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556626-z45rd" Mar 13 10:26:12 crc kubenswrapper[4632]: I0313 10:26:12.586327 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556626-z45rd" event={"ID":"8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27","Type":"ContainerDied","Data":"6d61d4f7c0c3be011ec1f0f84978bb250c298a5c766226d624254ef183165b94"} Mar 13 10:26:12 crc kubenswrapper[4632]: I0313 10:26:12.586377 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d61d4f7c0c3be011ec1f0f84978bb250c298a5c766226d624254ef183165b94" Mar 13 10:26:12 crc kubenswrapper[4632]: I0313 10:26:12.589251 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"050df504-63b9-4453-be2b-f3b0315fb801","Type":"ContainerStarted","Data":"4286cd55d064d024725ded90d153143e568de28aeedc6a6060f69501102dd4cb"} Mar 13 10:26:12 crc kubenswrapper[4632]: I0313 10:26:12.625324 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556620-42vs6"] Mar 13 10:26:12 crc kubenswrapper[4632]: I0313 10:26:12.650531 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556620-42vs6"] Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.064745 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28dbef1d-ca7f-4387-80af-8dffbfe92895" path="/var/lib/kubelet/pods/28dbef1d-ca7f-4387-80af-8dffbfe92895/volumes" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.401645 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-htnd9" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.461267 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e92afa62-9c75-4e0e-92f4-76e57328d7a0-logs\") pod \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.461337 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mqp5h\" (UniqueName: \"kubernetes.io/projected/e92afa62-9c75-4e0e-92f4-76e57328d7a0-kube-api-access-mqp5h\") pod \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.461388 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-config-data\") pod \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.461558 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-combined-ca-bundle\") pod \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.461699 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-scripts\") pod \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\" (UID: \"e92afa62-9c75-4e0e-92f4-76e57328d7a0\") " Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.470373 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e92afa62-9c75-4e0e-92f4-76e57328d7a0-logs" (OuterVolumeSpecName: "logs") pod "e92afa62-9c75-4e0e-92f4-76e57328d7a0" (UID: "e92afa62-9c75-4e0e-92f4-76e57328d7a0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.518360 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e92afa62-9c75-4e0e-92f4-76e57328d7a0-kube-api-access-mqp5h" (OuterVolumeSpecName: "kube-api-access-mqp5h") pod "e92afa62-9c75-4e0e-92f4-76e57328d7a0" (UID: "e92afa62-9c75-4e0e-92f4-76e57328d7a0"). InnerVolumeSpecName "kube-api-access-mqp5h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.518509 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-scripts" (OuterVolumeSpecName: "scripts") pod "e92afa62-9c75-4e0e-92f4-76e57328d7a0" (UID: "e92afa62-9c75-4e0e-92f4-76e57328d7a0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.561649 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-config-data" (OuterVolumeSpecName: "config-data") pod "e92afa62-9c75-4e0e-92f4-76e57328d7a0" (UID: "e92afa62-9c75-4e0e-92f4-76e57328d7a0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.564211 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.564486 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e92afa62-9c75-4e0e-92f4-76e57328d7a0-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.564585 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mqp5h\" (UniqueName: \"kubernetes.io/projected/e92afa62-9c75-4e0e-92f4-76e57328d7a0-kube-api-access-mqp5h\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.564832 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.568563 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e92afa62-9c75-4e0e-92f4-76e57328d7a0" (UID: "e92afa62-9c75-4e0e-92f4-76e57328d7a0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.666096 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e92afa62-9c75-4e0e-92f4-76e57328d7a0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.687502 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-htnd9" event={"ID":"e92afa62-9c75-4e0e-92f4-76e57328d7a0","Type":"ContainerDied","Data":"fa8253910988ff0dbee81a3230f0ff84637c4204c805ed0e40f0cc26f23d5381"} Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.688050 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa8253910988ff0dbee81a3230f0ff84637c4204c805ed0e40f0cc26f23d5381" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.688205 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-htnd9" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.698313 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"62c1f3f8-e898-4481-88e0-49f0c20228a4","Type":"ContainerStarted","Data":"60e19c69317a817c5bf104bc8691bdf46121d52039ad19099e25f869718b8e19"} Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.719152 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"050df504-63b9-4453-be2b-f3b0315fb801","Type":"ContainerStarted","Data":"a06e9823c7700968605c221a9839cf4f237fe6a7eee8836d69bade62686f4372"} Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.732021 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-7dd5c7bdcd-4969b"] Mar 13 10:26:14 crc kubenswrapper[4632]: E0313 10:26:14.732606 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e92afa62-9c75-4e0e-92f4-76e57328d7a0" containerName="placement-db-sync" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.732626 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e92afa62-9c75-4e0e-92f4-76e57328d7a0" containerName="placement-db-sync" Mar 13 10:26:14 crc kubenswrapper[4632]: E0313 10:26:14.732641 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27" containerName="oc" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.732648 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27" containerName="oc" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.732852 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27" containerName="oc" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.732873 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="e92afa62-9c75-4e0e-92f4-76e57328d7a0" containerName="placement-db-sync" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.734682 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.770476 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7dd5c7bdcd-4969b"] Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.776913 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.776889377 podStartE2EDuration="6.776889377s" podCreationTimestamp="2026-03-13 10:26:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:14.730526841 +0000 UTC m=+1348.753056984" watchObservedRunningTime="2026-03-13 10:26:14.776889377 +0000 UTC m=+1348.799419510" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.777905 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.778396 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-6tvl4" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.778647 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.778983 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.779180 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.814692 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.81467343 podStartE2EDuration="6.81467343s" podCreationTimestamp="2026-03-13 10:26:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:14.813784109 +0000 UTC m=+1348.836314242" watchObservedRunningTime="2026-03-13 10:26:14.81467343 +0000 UTC m=+1348.837203573" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.871038 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-internal-tls-certs\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.871416 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-config-data\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.871599 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-combined-ca-bundle\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.871753 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xk8t2\" (UniqueName: \"kubernetes.io/projected/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-kube-api-access-xk8t2\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.871902 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-public-tls-certs\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.873542 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-logs\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.873750 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-scripts\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.975428 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-internal-tls-certs\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.975493 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-config-data\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.975554 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-combined-ca-bundle\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.975597 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xk8t2\" (UniqueName: \"kubernetes.io/projected/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-kube-api-access-xk8t2\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.975633 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-public-tls-certs\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.975665 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-logs\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.975719 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-scripts\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.977395 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-logs\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.982150 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-config-data\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.982717 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-scripts\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.983237 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-public-tls-certs\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.983868 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-internal-tls-certs\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.986649 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-combined-ca-bundle\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:14 crc kubenswrapper[4632]: I0313 10:26:14.997469 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xk8t2\" (UniqueName: \"kubernetes.io/projected/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-kube-api-access-xk8t2\") pod \"placement-7dd5c7bdcd-4969b\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:15 crc kubenswrapper[4632]: I0313 10:26:15.111807 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:15 crc kubenswrapper[4632]: I0313 10:26:15.400370 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 10:26:15 crc kubenswrapper[4632]: I0313 10:26:15.858588 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-689764498d-rg7vt" podUID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 13 10:26:17 crc kubenswrapper[4632]: I0313 10:26:17.783304 4632 generic.go:334] "Generic (PLEG): container finished" podID="d8d0f662-d180-4137-8107-e465c5fb0621" containerID="3b5385b113397b9418c59a941d2a27f232c7b0df4b245db65886e55380c57297" exitCode=0 Mar 13 10:26:17 crc kubenswrapper[4632]: I0313 10:26:17.783638 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x8tq8" event={"ID":"d8d0f662-d180-4137-8107-e465c5fb0621","Type":"ContainerDied","Data":"3b5385b113397b9418c59a941d2a27f232c7b0df4b245db65886e55380c57297"} Mar 13 10:26:18 crc kubenswrapper[4632]: I0313 10:26:18.799225 4632 generic.go:334] "Generic (PLEG): container finished" podID="418cb883-abd1-46b4-957f-0a40f3e62297" containerID="3672f721f5cc963fe48f19a0fe26275ae0f1cbd82fd44ed2d6b14dcbb240be1d" exitCode=0 Mar 13 10:26:18 crc kubenswrapper[4632]: I0313 10:26:18.799318 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zdgpw" event={"ID":"418cb883-abd1-46b4-957f-0a40f3e62297","Type":"ContainerDied","Data":"3672f721f5cc963fe48f19a0fe26275ae0f1cbd82fd44ed2d6b14dcbb240be1d"} Mar 13 10:26:19 crc kubenswrapper[4632]: I0313 10:26:19.184424 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 10:26:19 crc kubenswrapper[4632]: I0313 10:26:19.184854 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 10:26:19 crc kubenswrapper[4632]: I0313 10:26:19.246489 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 10:26:19 crc kubenswrapper[4632]: I0313 10:26:19.254692 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 10:26:19 crc kubenswrapper[4632]: I0313 10:26:19.284904 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 10:26:19 crc kubenswrapper[4632]: I0313 10:26:19.284994 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 10:26:19 crc kubenswrapper[4632]: I0313 10:26:19.356911 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 10:26:19 crc kubenswrapper[4632]: I0313 10:26:19.388832 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 10:26:19 crc kubenswrapper[4632]: I0313 10:26:19.816813 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 10:26:19 crc kubenswrapper[4632]: I0313 10:26:19.816851 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 10:26:19 crc kubenswrapper[4632]: I0313 10:26:19.816866 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 10:26:19 crc kubenswrapper[4632]: I0313 10:26:19.817059 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 10:26:20 crc kubenswrapper[4632]: I0313 10:26:20.826423 4632 generic.go:334] "Generic (PLEG): container finished" podID="d722ddd7-e65d-44f7-a02d-18ddf126ccf5" containerID="3ef3ce34ce4d2a0d8d000d31874aca20b10c953ddde87f68a0b04979e69b8bae" exitCode=0 Mar 13 10:26:20 crc kubenswrapper[4632]: I0313 10:26:20.827880 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-7fvlk" event={"ID":"d722ddd7-e65d-44f7-a02d-18ddf126ccf5","Type":"ContainerDied","Data":"3ef3ce34ce4d2a0d8d000d31874aca20b10c953ddde87f68a0b04979e69b8bae"} Mar 13 10:26:21 crc kubenswrapper[4632]: I0313 10:26:21.837968 4632 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 10:26:21 crc kubenswrapper[4632]: I0313 10:26:21.837998 4632 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 10:26:21 crc kubenswrapper[4632]: I0313 10:26:21.838136 4632 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 10:26:21 crc kubenswrapper[4632]: I0313 10:26:21.838147 4632 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.004554 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.117637 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-combined-ca-bundle\") pod \"d8d0f662-d180-4137-8107-e465c5fb0621\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.117748 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-scripts\") pod \"d8d0f662-d180-4137-8107-e465c5fb0621\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.117786 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-fernet-keys\") pod \"d8d0f662-d180-4137-8107-e465c5fb0621\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.117824 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-credential-keys\") pod \"d8d0f662-d180-4137-8107-e465c5fb0621\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.118017 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-config-data\") pod \"d8d0f662-d180-4137-8107-e465c5fb0621\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.118111 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ln474\" (UniqueName: \"kubernetes.io/projected/d8d0f662-d180-4137-8107-e465c5fb0621-kube-api-access-ln474\") pod \"d8d0f662-d180-4137-8107-e465c5fb0621\" (UID: \"d8d0f662-d180-4137-8107-e465c5fb0621\") " Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.129878 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8d0f662-d180-4137-8107-e465c5fb0621-kube-api-access-ln474" (OuterVolumeSpecName: "kube-api-access-ln474") pod "d8d0f662-d180-4137-8107-e465c5fb0621" (UID: "d8d0f662-d180-4137-8107-e465c5fb0621"). InnerVolumeSpecName "kube-api-access-ln474". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.139030 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-scripts" (OuterVolumeSpecName: "scripts") pod "d8d0f662-d180-4137-8107-e465c5fb0621" (UID: "d8d0f662-d180-4137-8107-e465c5fb0621"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.174542 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "d8d0f662-d180-4137-8107-e465c5fb0621" (UID: "d8d0f662-d180-4137-8107-e465c5fb0621"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.180327 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "d8d0f662-d180-4137-8107-e465c5fb0621" (UID: "d8d0f662-d180-4137-8107-e465c5fb0621"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.207248 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8d0f662-d180-4137-8107-e465c5fb0621" (UID: "d8d0f662-d180-4137-8107-e465c5fb0621"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.239085 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ln474\" (UniqueName: \"kubernetes.io/projected/d8d0f662-d180-4137-8107-e465c5fb0621-kube-api-access-ln474\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.239118 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.239127 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.239135 4632 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.239146 4632 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.245527 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-config-data" (OuterVolumeSpecName: "config-data") pod "d8d0f662-d180-4137-8107-e465c5fb0621" (UID: "d8d0f662-d180-4137-8107-e465c5fb0621"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.293243 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zdgpw" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.340985 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8d0f662-d180-4137-8107-e465c5fb0621-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.393174 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-7fvlk" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.442548 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/418cb883-abd1-46b4-957f-0a40f3e62297-db-sync-config-data\") pod \"418cb883-abd1-46b4-957f-0a40f3e62297\" (UID: \"418cb883-abd1-46b4-957f-0a40f3e62297\") " Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.442990 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgn44\" (UniqueName: \"kubernetes.io/projected/418cb883-abd1-46b4-957f-0a40f3e62297-kube-api-access-zgn44\") pod \"418cb883-abd1-46b4-957f-0a40f3e62297\" (UID: \"418cb883-abd1-46b4-957f-0a40f3e62297\") " Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.443061 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418cb883-abd1-46b4-957f-0a40f3e62297-combined-ca-bundle\") pod \"418cb883-abd1-46b4-957f-0a40f3e62297\" (UID: \"418cb883-abd1-46b4-957f-0a40f3e62297\") " Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.449357 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/418cb883-abd1-46b4-957f-0a40f3e62297-kube-api-access-zgn44" (OuterVolumeSpecName: "kube-api-access-zgn44") pod "418cb883-abd1-46b4-957f-0a40f3e62297" (UID: "418cb883-abd1-46b4-957f-0a40f3e62297"). InnerVolumeSpecName "kube-api-access-zgn44". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.476934 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/418cb883-abd1-46b4-957f-0a40f3e62297-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "418cb883-abd1-46b4-957f-0a40f3e62297" (UID: "418cb883-abd1-46b4-957f-0a40f3e62297"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.527241 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/418cb883-abd1-46b4-957f-0a40f3e62297-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "418cb883-abd1-46b4-957f-0a40f3e62297" (UID: "418cb883-abd1-46b4-957f-0a40f3e62297"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.544539 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-combined-ca-bundle\") pod \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\" (UID: \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\") " Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.544722 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-config-data\") pod \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\" (UID: \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\") " Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.544851 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n5rzh\" (UniqueName: \"kubernetes.io/projected/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-kube-api-access-n5rzh\") pod \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\" (UID: \"d722ddd7-e65d-44f7-a02d-18ddf126ccf5\") " Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.545351 4632 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/418cb883-abd1-46b4-957f-0a40f3e62297-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.545375 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgn44\" (UniqueName: \"kubernetes.io/projected/418cb883-abd1-46b4-957f-0a40f3e62297-kube-api-access-zgn44\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.545386 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/418cb883-abd1-46b4-957f-0a40f3e62297-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.558236 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-kube-api-access-n5rzh" (OuterVolumeSpecName: "kube-api-access-n5rzh") pod "d722ddd7-e65d-44f7-a02d-18ddf126ccf5" (UID: "d722ddd7-e65d-44f7-a02d-18ddf126ccf5"). InnerVolumeSpecName "kube-api-access-n5rzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.582477 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d722ddd7-e65d-44f7-a02d-18ddf126ccf5" (UID: "d722ddd7-e65d-44f7-a02d-18ddf126ccf5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.654491 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n5rzh\" (UniqueName: \"kubernetes.io/projected/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-kube-api-access-n5rzh\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.654809 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.704908 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-config-data" (OuterVolumeSpecName: "config-data") pod "d722ddd7-e65d-44f7-a02d-18ddf126ccf5" (UID: "d722ddd7-e65d-44f7-a02d-18ddf126ccf5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.756917 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d722ddd7-e65d-44f7-a02d-18ddf126ccf5-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.859400 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-7dd5c7bdcd-4969b"] Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.866964 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-x8tq8" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.866961 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-x8tq8" event={"ID":"d8d0f662-d180-4137-8107-e465c5fb0621","Type":"ContainerDied","Data":"58b9dfc8050bba09291b639a8d3d5cc84a9643af0afb63feda9e26973d06a678"} Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.867667 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="58b9dfc8050bba09291b639a8d3d5cc84a9643af0afb63feda9e26973d06a678" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.877530 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zdgpw" event={"ID":"418cb883-abd1-46b4-957f-0a40f3e62297","Type":"ContainerDied","Data":"47e1c2b826ae3f1aaa52b7a4210b405df85537a8c7de35fb1657923a6d754982"} Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.877580 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47e1c2b826ae3f1aaa52b7a4210b405df85537a8c7de35fb1657923a6d754982" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.877867 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zdgpw" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.892313 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-7fvlk" event={"ID":"d722ddd7-e65d-44f7-a02d-18ddf126ccf5","Type":"ContainerDied","Data":"76d57552a9eced6e283cb6dee93cf8db23032b8fbb20e4a910d615de236f52d7"} Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.892592 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76d57552a9eced6e283cb6dee93cf8db23032b8fbb20e4a910d615de236f52d7" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.892707 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-7fvlk" Mar 13 10:26:23 crc kubenswrapper[4632]: I0313 10:26:23.912341 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"270ebc10-986f-4473-8a5e-9094de34ae98","Type":"ContainerStarted","Data":"d138d976167695fe9d299247eefcff55845f7ad27e84fc81cc086274294f2e51"} Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.182236 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-f664b756d-8fxf4"] Mar 13 10:26:24 crc kubenswrapper[4632]: E0313 10:26:24.190749 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8d0f662-d180-4137-8107-e465c5fb0621" containerName="keystone-bootstrap" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.190792 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8d0f662-d180-4137-8107-e465c5fb0621" containerName="keystone-bootstrap" Mar 13 10:26:24 crc kubenswrapper[4632]: E0313 10:26:24.190860 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="418cb883-abd1-46b4-957f-0a40f3e62297" containerName="barbican-db-sync" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.190876 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="418cb883-abd1-46b4-957f-0a40f3e62297" containerName="barbican-db-sync" Mar 13 10:26:24 crc kubenswrapper[4632]: E0313 10:26:24.190895 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d722ddd7-e65d-44f7-a02d-18ddf126ccf5" containerName="heat-db-sync" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.190904 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d722ddd7-e65d-44f7-a02d-18ddf126ccf5" containerName="heat-db-sync" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.230273 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="418cb883-abd1-46b4-957f-0a40f3e62297" containerName="barbican-db-sync" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.230522 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="d722ddd7-e65d-44f7-a02d-18ddf126ccf5" containerName="heat-db-sync" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.230594 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8d0f662-d180-4137-8107-e465c5fb0621" containerName="keystone-bootstrap" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.231402 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.231907 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f664b756d-8fxf4"] Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.235405 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.235870 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.236220 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.236456 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-llpcf" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.236654 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.244201 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.293690 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-scripts\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.293766 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-combined-ca-bundle\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.293789 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-public-tls-certs\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.293822 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-fernet-keys\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.293842 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrvpt\" (UniqueName: \"kubernetes.io/projected/df64dbf7-8526-4fab-950a-4afefe47ec77-kube-api-access-rrvpt\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.293860 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-internal-tls-certs\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.293901 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-credential-keys\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.293929 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-config-data\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.298276 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-5c86b4b888-l9574" podUID="6d73a499-d334-4a7a-9783-640b98760672" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.298785 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-5c86b4b888-l9574" podUID="6d73a499-d334-4a7a-9783-640b98760672" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.307869 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-5c86b4b888-l9574" podUID="6d73a499-d334-4a7a-9783-640b98760672" containerName="neutron-api" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.396640 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-config-data\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.396756 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-scripts\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.396790 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-combined-ca-bundle\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.396815 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-public-tls-certs\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.396857 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-fernet-keys\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.396891 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrvpt\" (UniqueName: \"kubernetes.io/projected/df64dbf7-8526-4fab-950a-4afefe47ec77-kube-api-access-rrvpt\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.396916 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-internal-tls-certs\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.397000 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-credential-keys\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.404084 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-combined-ca-bundle\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.412501 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-fernet-keys\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.412633 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-scripts\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.420020 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-credential-keys\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.420353 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-public-tls-certs\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.421003 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-internal-tls-certs\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.424696 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df64dbf7-8526-4fab-950a-4afefe47ec77-config-data\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.427214 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrvpt\" (UniqueName: \"kubernetes.io/projected/df64dbf7-8526-4fab-950a-4afefe47ec77-kube-api-access-rrvpt\") pod \"keystone-f664b756d-8fxf4\" (UID: \"df64dbf7-8526-4fab-950a-4afefe47ec77\") " pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.567000 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.659722 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5fc9b6f5b5-6ps9m"] Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.665190 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.678086 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.678261 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-m45mn" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.678402 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.716642 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5fc9b6f5b5-6ps9m"] Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.759403 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51b847ef-ada2-456f-819d-0084fbb17185-config-data-custom\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.759462 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51b847ef-ada2-456f-819d-0084fbb17185-logs\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.759560 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24dz4\" (UniqueName: \"kubernetes.io/projected/51b847ef-ada2-456f-819d-0084fbb17185-kube-api-access-24dz4\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.759592 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b847ef-ada2-456f-819d-0084fbb17185-config-data\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.759633 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b847ef-ada2-456f-819d-0084fbb17185-combined-ca-bundle\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.759778 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-6c97cdfb86-z2dqq"] Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.785216 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.809401 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.856121 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6c97cdfb86-z2dqq"] Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.861146 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-logs\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.861215 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-combined-ca-bundle\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.861255 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsmbv\" (UniqueName: \"kubernetes.io/projected/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-kube-api-access-lsmbv\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.861298 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24dz4\" (UniqueName: \"kubernetes.io/projected/51b847ef-ada2-456f-819d-0084fbb17185-kube-api-access-24dz4\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.861326 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b847ef-ada2-456f-819d-0084fbb17185-config-data\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.861350 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-config-data\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.861399 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b847ef-ada2-456f-819d-0084fbb17185-combined-ca-bundle\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.861462 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-config-data-custom\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.861532 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51b847ef-ada2-456f-819d-0084fbb17185-config-data-custom\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.861579 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51b847ef-ada2-456f-819d-0084fbb17185-logs\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.862127 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/51b847ef-ada2-456f-819d-0084fbb17185-logs\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.889173 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cc676b85c-q67wf"] Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.890924 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.949378 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cc676b85c-q67wf"] Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.953721 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/51b847ef-ada2-456f-819d-0084fbb17185-combined-ca-bundle\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.956439 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/51b847ef-ada2-456f-819d-0084fbb17185-config-data\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.963615 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-dns-swift-storage-0\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.963806 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-logs\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.963894 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-combined-ca-bundle\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.963991 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq8dx\" (UniqueName: \"kubernetes.io/projected/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-kube-api-access-sq8dx\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.964065 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-dns-svc\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.964159 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-ovsdbserver-sb\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.964237 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsmbv\" (UniqueName: \"kubernetes.io/projected/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-kube-api-access-lsmbv\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.964335 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-config-data\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.964424 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-config\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.964497 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-ovsdbserver-nb\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.964586 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-config-data-custom\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.965616 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/51b847ef-ada2-456f-819d-0084fbb17185-config-data-custom\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.967272 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24dz4\" (UniqueName: \"kubernetes.io/projected/51b847ef-ada2-456f-819d-0084fbb17185-kube-api-access-24dz4\") pod \"barbican-worker-5fc9b6f5b5-6ps9m\" (UID: \"51b847ef-ada2-456f-819d-0084fbb17185\") " pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.968255 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-logs\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.977418 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7dd5c7bdcd-4969b" event={"ID":"5abe7bf3-d44d-4ee5-b568-2d497868f1e5","Type":"ContainerStarted","Data":"0b15584f3607b654abe16b00ac290d1bc5ee6f763bd08234d8697e7f5b5b20bb"} Mar 13 10:26:24 crc kubenswrapper[4632]: I0313 10:26:24.981307 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7dd5c7bdcd-4969b" event={"ID":"5abe7bf3-d44d-4ee5-b568-2d497868f1e5","Type":"ContainerStarted","Data":"604b160eb4cd534ac8def868fbcdab1d748e8bc2952c85fe7198dc4a2b05d7f7"} Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.000684 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-config-data-custom\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.013044 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-config-data\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.020551 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsmbv\" (UniqueName: \"kubernetes.io/projected/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-kube-api-access-lsmbv\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.020601 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58332dcc-b1a6-4550-9c8b-8bbb82c04ff0-combined-ca-bundle\") pod \"barbican-keystone-listener-6c97cdfb86-z2dqq\" (UID: \"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0\") " pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.059575 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.068198 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq8dx\" (UniqueName: \"kubernetes.io/projected/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-kube-api-access-sq8dx\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.076466 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-dns-svc\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.076517 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-ovsdbserver-sb\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.076629 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-config\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.076655 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-ovsdbserver-nb\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.076851 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-dns-swift-storage-0\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.078817 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-dns-svc\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.079907 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-ovsdbserver-sb\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.081244 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-config\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.081912 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-ovsdbserver-nb\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.083864 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-dns-swift-storage-0\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.156839 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq8dx\" (UniqueName: \"kubernetes.io/projected/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-kube-api-access-sq8dx\") pod \"dnsmasq-dns-7cc676b85c-q67wf\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.175184 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.211363 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-548c8b4b94-2dglr"] Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.231137 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-548c8b4b94-2dglr"] Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.231272 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.239274 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.272015 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.394871 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.411139 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-combined-ca-bundle\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.411243 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-config-data-custom\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.411317 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fa310f1-40ef-4e74-9647-d3ea87858f11-logs\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.411399 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-config-data\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.411434 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pmkj\" (UniqueName: \"kubernetes.io/projected/6fa310f1-40ef-4e74-9647-d3ea87858f11-kube-api-access-8pmkj\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.513911 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-config-data\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.513989 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pmkj\" (UniqueName: \"kubernetes.io/projected/6fa310f1-40ef-4e74-9647-d3ea87858f11-kube-api-access-8pmkj\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.514042 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-combined-ca-bundle\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.514081 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-config-data-custom\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.514150 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fa310f1-40ef-4e74-9647-d3ea87858f11-logs\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.514665 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fa310f1-40ef-4e74-9647-d3ea87858f11-logs\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.544875 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-combined-ca-bundle\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.549072 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-config-data\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.565218 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pmkj\" (UniqueName: \"kubernetes.io/projected/6fa310f1-40ef-4e74-9647-d3ea87858f11-kube-api-access-8pmkj\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.586479 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-config-data-custom\") pod \"barbican-api-548c8b4b94-2dglr\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.587100 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-f664b756d-8fxf4"] Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.600234 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:25 crc kubenswrapper[4632]: I0313 10:26:25.860815 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-689764498d-rg7vt" podUID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 13 10:26:26 crc kubenswrapper[4632]: I0313 10:26:26.006692 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5fc9b6f5b5-6ps9m"] Mar 13 10:26:26 crc kubenswrapper[4632]: I0313 10:26:26.068780 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:26 crc kubenswrapper[4632]: I0313 10:26:26.069141 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:26 crc kubenswrapper[4632]: I0313 10:26:26.069156 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7dd5c7bdcd-4969b" event={"ID":"5abe7bf3-d44d-4ee5-b568-2d497868f1e5","Type":"ContainerStarted","Data":"2cfe7ebd70fe3427d7ef352e87ea88bca1736af36e0c260541ced9066c436503"} Mar 13 10:26:26 crc kubenswrapper[4632]: I0313 10:26:26.069173 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f664b756d-8fxf4" event={"ID":"df64dbf7-8526-4fab-950a-4afefe47ec77","Type":"ContainerStarted","Data":"80faef951510ea08c374210ef40d4f2ccc3e8cf3ed32946c01da45487a2e4258"} Mar 13 10:26:26 crc kubenswrapper[4632]: I0313 10:26:26.142259 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-7dd5c7bdcd-4969b" podStartSLOduration=12.142236459 podStartE2EDuration="12.142236459s" podCreationTimestamp="2026-03-13 10:26:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:26.092273858 +0000 UTC m=+1360.114804011" watchObservedRunningTime="2026-03-13 10:26:26.142236459 +0000 UTC m=+1360.164766592" Mar 13 10:26:26 crc kubenswrapper[4632]: I0313 10:26:26.440894 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-6c97cdfb86-z2dqq"] Mar 13 10:26:26 crc kubenswrapper[4632]: I0313 10:26:26.455879 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cc676b85c-q67wf"] Mar 13 10:26:26 crc kubenswrapper[4632]: I0313 10:26:26.680721 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-548c8b4b94-2dglr"] Mar 13 10:26:27 crc kubenswrapper[4632]: I0313 10:26:27.152061 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" event={"ID":"6cbca281-a753-4810-ab5f-a2d5a5e9c41d","Type":"ContainerStarted","Data":"c35af663b9fb4942a62d07ad5236c27fcd9454c66f1967196835246c5924112a"} Mar 13 10:26:27 crc kubenswrapper[4632]: I0313 10:26:27.154216 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" event={"ID":"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0","Type":"ContainerStarted","Data":"715c026f3a82423d483c078d594d783c0987f11d6bee52b752444c42e096d445"} Mar 13 10:26:27 crc kubenswrapper[4632]: I0313 10:26:27.197440 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-f664b756d-8fxf4" event={"ID":"df64dbf7-8526-4fab-950a-4afefe47ec77","Type":"ContainerStarted","Data":"e1bb600f766cd508b5b0989591cf2811ce2ae4a392a60f30c9317e38a3c5276e"} Mar 13 10:26:27 crc kubenswrapper[4632]: I0313 10:26:27.197522 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:26:27 crc kubenswrapper[4632]: I0313 10:26:27.225001 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-548c8b4b94-2dglr" event={"ID":"6fa310f1-40ef-4e74-9647-d3ea87858f11","Type":"ContainerStarted","Data":"4545ac42523c98f674d28d5d0acc10645d2b1e7d8486b7068d13c265711710a4"} Mar 13 10:26:27 crc kubenswrapper[4632]: I0313 10:26:27.230074 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" event={"ID":"51b847ef-ada2-456f-819d-0084fbb17185","Type":"ContainerStarted","Data":"8bf20e8f4a839fa6fe33923720402b2b880de8a7752f1d4c0b7f5335f0df2afa"} Mar 13 10:26:27 crc kubenswrapper[4632]: I0313 10:26:27.236552 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-f664b756d-8fxf4" podStartSLOduration=3.236531996 podStartE2EDuration="3.236531996s" podCreationTimestamp="2026-03-13 10:26:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:27.228491026 +0000 UTC m=+1361.251021159" watchObservedRunningTime="2026-03-13 10:26:27.236531996 +0000 UTC m=+1361.259062139" Mar 13 10:26:27 crc kubenswrapper[4632]: I0313 10:26:27.356665 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-64bdffbb5c-mpfvf" podUID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 13 10:26:27 crc kubenswrapper[4632]: I0313 10:26:27.356701 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/neutron-64bdffbb5c-mpfvf" podUID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" containerName="neutron-api" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 13 10:26:27 crc kubenswrapper[4632]: I0313 10:26:27.370264 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-64bdffbb5c-mpfvf" podUID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" containerName="neutron-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 13 10:26:28 crc kubenswrapper[4632]: I0313 10:26:28.269147 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-548c8b4b94-2dglr" event={"ID":"6fa310f1-40ef-4e74-9647-d3ea87858f11","Type":"ContainerStarted","Data":"309fa94df210d44c275999bad3e9b781bb4f9646e038b1a9463656385d210cf3"} Mar 13 10:26:28 crc kubenswrapper[4632]: I0313 10:26:28.269466 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-548c8b4b94-2dglr" event={"ID":"6fa310f1-40ef-4e74-9647-d3ea87858f11","Type":"ContainerStarted","Data":"67882325af120e97844e1aef36a358fdd186b89ba1f3def214e49a353ec793aa"} Mar 13 10:26:28 crc kubenswrapper[4632]: I0313 10:26:28.269484 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:28 crc kubenswrapper[4632]: I0313 10:26:28.269497 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:28 crc kubenswrapper[4632]: I0313 10:26:28.289675 4632 generic.go:334] "Generic (PLEG): container finished" podID="6cbca281-a753-4810-ab5f-a2d5a5e9c41d" containerID="de515c666eef870de947ba353b522abb4ca9136dc9f866c10cc7d68d392957ce" exitCode=0 Mar 13 10:26:28 crc kubenswrapper[4632]: I0313 10:26:28.290477 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" event={"ID":"6cbca281-a753-4810-ab5f-a2d5a5e9c41d","Type":"ContainerDied","Data":"de515c666eef870de947ba353b522abb4ca9136dc9f866c10cc7d68d392957ce"} Mar 13 10:26:28 crc kubenswrapper[4632]: I0313 10:26:28.293928 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-548c8b4b94-2dglr" podStartSLOduration=3.29391573 podStartE2EDuration="3.29391573s" podCreationTimestamp="2026-03-13 10:26:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:28.291818391 +0000 UTC m=+1362.314348524" watchObservedRunningTime="2026-03-13 10:26:28.29391573 +0000 UTC m=+1362.316445863" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.312368 4632 generic.go:334] "Generic (PLEG): container finished" podID="95fe9a38-2b32-411e-9121-ad4cc32f159e" containerID="d9db78843b825b24c0eab6345b91a7657d2b3f0bb64d65b5dcc125b1edeeb022" exitCode=137 Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.313234 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67d6b4b8f7-nrxn8" event={"ID":"95fe9a38-2b32-411e-9121-ad4cc32f159e","Type":"ContainerDied","Data":"d9db78843b825b24c0eab6345b91a7657d2b3f0bb64d65b5dcc125b1edeeb022"} Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.609749 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-756c4b86c6-rm274"] Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.611581 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.627542 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.647744 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-756c4b86c6-rm274"] Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.652870 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.703983 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-config-data\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.704277 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltgf8\" (UniqueName: \"kubernetes.io/projected/dbc1c989-5fa1-46dc-818e-8d609c069e34-kube-api-access-ltgf8\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.704326 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-config-data-custom\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.704352 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-combined-ca-bundle\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.704393 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-public-tls-certs\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.704432 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-internal-tls-certs\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.704454 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbc1c989-5fa1-46dc-818e-8d609c069e34-logs\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.807052 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltgf8\" (UniqueName: \"kubernetes.io/projected/dbc1c989-5fa1-46dc-818e-8d609c069e34-kube-api-access-ltgf8\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.807139 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-config-data-custom\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.807169 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-combined-ca-bundle\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.807228 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-public-tls-certs\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.807270 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-internal-tls-certs\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.807294 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbc1c989-5fa1-46dc-818e-8d609c069e34-logs\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.807406 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-config-data\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.809327 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/dbc1c989-5fa1-46dc-818e-8d609c069e34-logs\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.820074 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-config-data\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.823698 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-combined-ca-bundle\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.826420 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-internal-tls-certs\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.831487 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-config-data-custom\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.837718 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/dbc1c989-5fa1-46dc-818e-8d609c069e34-public-tls-certs\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.841341 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltgf8\" (UniqueName: \"kubernetes.io/projected/dbc1c989-5fa1-46dc-818e-8d609c069e34-kube-api-access-ltgf8\") pod \"barbican-api-756c4b86c6-rm274\" (UID: \"dbc1c989-5fa1-46dc-818e-8d609c069e34\") " pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:29 crc kubenswrapper[4632]: I0313 10:26:29.932209 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:30 crc kubenswrapper[4632]: I0313 10:26:30.341416 4632 generic.go:334] "Generic (PLEG): container finished" podID="95fe9a38-2b32-411e-9121-ad4cc32f159e" containerID="f8238ac2122bdce07c274a4f41c5a0d859a4162d57594e52444f5d2a425d1e7b" exitCode=137 Mar 13 10:26:30 crc kubenswrapper[4632]: I0313 10:26:30.341499 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67d6b4b8f7-nrxn8" event={"ID":"95fe9a38-2b32-411e-9121-ad4cc32f159e","Type":"ContainerDied","Data":"f8238ac2122bdce07c274a4f41c5a0d859a4162d57594e52444f5d2a425d1e7b"} Mar 13 10:26:30 crc kubenswrapper[4632]: I0313 10:26:30.556585 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 10:26:30 crc kubenswrapper[4632]: I0313 10:26:30.556714 4632 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 10:26:30 crc kubenswrapper[4632]: I0313 10:26:30.581672 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 10:26:31 crc kubenswrapper[4632]: I0313 10:26:31.017432 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 10:26:31 crc kubenswrapper[4632]: I0313 10:26:31.017793 4632 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 10:26:31 crc kubenswrapper[4632]: I0313 10:26:31.116573 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 10:26:31 crc kubenswrapper[4632]: I0313 10:26:31.392907 4632 generic.go:334] "Generic (PLEG): container finished" podID="8f916c05-f172-42b6-9b13-0c8d2058bfb1" containerID="6d5ac5d7a6aab5517e4300c2e14808710d4f8cfa4977c9841f6552b262144012" exitCode=0 Mar 13 10:26:31 crc kubenswrapper[4632]: I0313 10:26:31.394548 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kq8lc" event={"ID":"8f916c05-f172-42b6-9b13-0c8d2058bfb1","Type":"ContainerDied","Data":"6d5ac5d7a6aab5517e4300c2e14808710d4f8cfa4977c9841f6552b262144012"} Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.546408 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.715867 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxkl9\" (UniqueName: \"kubernetes.io/projected/95fe9a38-2b32-411e-9121-ad4cc32f159e-kube-api-access-zxkl9\") pod \"95fe9a38-2b32-411e-9121-ad4cc32f159e\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.716474 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95fe9a38-2b32-411e-9121-ad4cc32f159e-scripts\") pod \"95fe9a38-2b32-411e-9121-ad4cc32f159e\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.716565 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/95fe9a38-2b32-411e-9121-ad4cc32f159e-horizon-secret-key\") pod \"95fe9a38-2b32-411e-9121-ad4cc32f159e\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.716686 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/95fe9a38-2b32-411e-9121-ad4cc32f159e-config-data\") pod \"95fe9a38-2b32-411e-9121-ad4cc32f159e\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.716736 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95fe9a38-2b32-411e-9121-ad4cc32f159e-logs\") pod \"95fe9a38-2b32-411e-9121-ad4cc32f159e\" (UID: \"95fe9a38-2b32-411e-9121-ad4cc32f159e\") " Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.724926 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95fe9a38-2b32-411e-9121-ad4cc32f159e-logs" (OuterVolumeSpecName: "logs") pod "95fe9a38-2b32-411e-9121-ad4cc32f159e" (UID: "95fe9a38-2b32-411e-9121-ad4cc32f159e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.736247 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95fe9a38-2b32-411e-9121-ad4cc32f159e-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "95fe9a38-2b32-411e-9121-ad4cc32f159e" (UID: "95fe9a38-2b32-411e-9121-ad4cc32f159e"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.750321 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95fe9a38-2b32-411e-9121-ad4cc32f159e-kube-api-access-zxkl9" (OuterVolumeSpecName: "kube-api-access-zxkl9") pod "95fe9a38-2b32-411e-9121-ad4cc32f159e" (UID: "95fe9a38-2b32-411e-9121-ad4cc32f159e"). InnerVolumeSpecName "kube-api-access-zxkl9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.792602 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95fe9a38-2b32-411e-9121-ad4cc32f159e-config-data" (OuterVolumeSpecName: "config-data") pod "95fe9a38-2b32-411e-9121-ad4cc32f159e" (UID: "95fe9a38-2b32-411e-9121-ad4cc32f159e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.814574 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95fe9a38-2b32-411e-9121-ad4cc32f159e-scripts" (OuterVolumeSpecName: "scripts") pod "95fe9a38-2b32-411e-9121-ad4cc32f159e" (UID: "95fe9a38-2b32-411e-9121-ad4cc32f159e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.819621 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxkl9\" (UniqueName: \"kubernetes.io/projected/95fe9a38-2b32-411e-9121-ad4cc32f159e-kube-api-access-zxkl9\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.819791 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/95fe9a38-2b32-411e-9121-ad4cc32f159e-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.819901 4632 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/95fe9a38-2b32-411e-9121-ad4cc32f159e-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.820132 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/95fe9a38-2b32-411e-9121-ad4cc32f159e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:32 crc kubenswrapper[4632]: I0313 10:26:32.820237 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/95fe9a38-2b32-411e-9121-ad4cc32f159e-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.042454 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.125297 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-scripts\") pod \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.125459 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-db-sync-config-data\") pod \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.125586 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5m5bn\" (UniqueName: \"kubernetes.io/projected/8f916c05-f172-42b6-9b13-0c8d2058bfb1-kube-api-access-5m5bn\") pod \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.125717 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-combined-ca-bundle\") pod \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.125755 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-config-data\") pod \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.126867 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f916c05-f172-42b6-9b13-0c8d2058bfb1-etc-machine-id\") pod \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\" (UID: \"8f916c05-f172-42b6-9b13-0c8d2058bfb1\") " Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.127043 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8f916c05-f172-42b6-9b13-0c8d2058bfb1-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8f916c05-f172-42b6-9b13-0c8d2058bfb1" (UID: "8f916c05-f172-42b6-9b13-0c8d2058bfb1"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.129121 4632 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8f916c05-f172-42b6-9b13-0c8d2058bfb1-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.168853 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f916c05-f172-42b6-9b13-0c8d2058bfb1-kube-api-access-5m5bn" (OuterVolumeSpecName: "kube-api-access-5m5bn") pod "8f916c05-f172-42b6-9b13-0c8d2058bfb1" (UID: "8f916c05-f172-42b6-9b13-0c8d2058bfb1"). InnerVolumeSpecName "kube-api-access-5m5bn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.172061 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-scripts" (OuterVolumeSpecName: "scripts") pod "8f916c05-f172-42b6-9b13-0c8d2058bfb1" (UID: "8f916c05-f172-42b6-9b13-0c8d2058bfb1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.172570 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8f916c05-f172-42b6-9b13-0c8d2058bfb1" (UID: "8f916c05-f172-42b6-9b13-0c8d2058bfb1"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.206005 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f916c05-f172-42b6-9b13-0c8d2058bfb1" (UID: "8f916c05-f172-42b6-9b13-0c8d2058bfb1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.230645 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.230688 4632 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.230706 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5m5bn\" (UniqueName: \"kubernetes.io/projected/8f916c05-f172-42b6-9b13-0c8d2058bfb1-kube-api-access-5m5bn\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.230718 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.298273 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-756c4b86c6-rm274"] Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.315053 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-config-data" (OuterVolumeSpecName: "config-data") pod "8f916c05-f172-42b6-9b13-0c8d2058bfb1" (UID: "8f916c05-f172-42b6-9b13-0c8d2058bfb1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:33 crc kubenswrapper[4632]: W0313 10:26:33.318917 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbc1c989_5fa1_46dc_818e_8d609c069e34.slice/crio-eda6b117169809aad9d8b5baaee672caebe4bfc7c3a6fbea05ec0231894d4fd5 WatchSource:0}: Error finding container eda6b117169809aad9d8b5baaee672caebe4bfc7c3a6fbea05ec0231894d4fd5: Status 404 returned error can't find the container with id eda6b117169809aad9d8b5baaee672caebe4bfc7c3a6fbea05ec0231894d4fd5 Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.332294 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f916c05-f172-42b6-9b13-0c8d2058bfb1-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.424149 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-kq8lc" event={"ID":"8f916c05-f172-42b6-9b13-0c8d2058bfb1","Type":"ContainerDied","Data":"8ac8055b0e5fc8cb1135e4ae559dd9794358a9f9dfb68fd20402b62c57115f00"} Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.424217 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ac8055b0e5fc8cb1135e4ae559dd9794358a9f9dfb68fd20402b62c57115f00" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.424407 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-kq8lc" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.426799 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" event={"ID":"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0","Type":"ContainerStarted","Data":"e64692995de9a1a74114696f6777f70f0296d603809ccbd570a856f3f597856f"} Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.426837 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" event={"ID":"58332dcc-b1a6-4550-9c8b-8bbb82c04ff0","Type":"ContainerStarted","Data":"4f9bc45bc10e2684aa506712975284aebd47a9a1ff8843e9364f36ca6efa6a60"} Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.434277 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-756c4b86c6-rm274" event={"ID":"dbc1c989-5fa1-46dc-818e-8d609c069e34","Type":"ContainerStarted","Data":"eda6b117169809aad9d8b5baaee672caebe4bfc7c3a6fbea05ec0231894d4fd5"} Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.440926 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" event={"ID":"51b847ef-ada2-456f-819d-0084fbb17185","Type":"ContainerStarted","Data":"bd7c8c722be2532e1ae19202c0e58d8318178f85688079a6dffd9812ff5db7bc"} Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.463176 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-6c97cdfb86-z2dqq" podStartSLOduration=3.39914094 podStartE2EDuration="9.46315514s" podCreationTimestamp="2026-03-13 10:26:24 +0000 UTC" firstStartedPulling="2026-03-13 10:26:26.479731416 +0000 UTC m=+1360.502261549" lastFinishedPulling="2026-03-13 10:26:32.543745616 +0000 UTC m=+1366.566275749" observedRunningTime="2026-03-13 10:26:33.450225064 +0000 UTC m=+1367.472755207" watchObservedRunningTime="2026-03-13 10:26:33.46315514 +0000 UTC m=+1367.485685273" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.475268 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-67d6b4b8f7-nrxn8" event={"ID":"95fe9a38-2b32-411e-9121-ad4cc32f159e","Type":"ContainerDied","Data":"23d0d6f6bc6174b2a86ec905a9477b2974881387bec66374cfa55dca37114aec"} Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.475330 4632 scope.go:117] "RemoveContainer" containerID="f8238ac2122bdce07c274a4f41c5a0d859a4162d57594e52444f5d2a425d1e7b" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.475435 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-67d6b4b8f7-nrxn8" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.484203 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" event={"ID":"6cbca281-a753-4810-ab5f-a2d5a5e9c41d","Type":"ContainerStarted","Data":"d4574f8c63575790d1486fa1f8221ce400d19901d3317c976e702416503cba64"} Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.506352 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.584705 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" podStartSLOduration=9.584681912 podStartE2EDuration="9.584681912s" podCreationTimestamp="2026-03-13 10:26:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:33.552551173 +0000 UTC m=+1367.575081296" watchObservedRunningTime="2026-03-13 10:26:33.584681912 +0000 UTC m=+1367.607212045" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.648164 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-67d6b4b8f7-nrxn8"] Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.659737 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-67d6b4b8f7-nrxn8"] Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.822681 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 10:26:33 crc kubenswrapper[4632]: E0313 10:26:33.823103 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95fe9a38-2b32-411e-9121-ad4cc32f159e" containerName="horizon" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.823121 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="95fe9a38-2b32-411e-9121-ad4cc32f159e" containerName="horizon" Mar 13 10:26:33 crc kubenswrapper[4632]: E0313 10:26:33.823132 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95fe9a38-2b32-411e-9121-ad4cc32f159e" containerName="horizon-log" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.823138 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="95fe9a38-2b32-411e-9121-ad4cc32f159e" containerName="horizon-log" Mar 13 10:26:33 crc kubenswrapper[4632]: E0313 10:26:33.823150 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f916c05-f172-42b6-9b13-0c8d2058bfb1" containerName="cinder-db-sync" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.823157 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f916c05-f172-42b6-9b13-0c8d2058bfb1" containerName="cinder-db-sync" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.823323 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="95fe9a38-2b32-411e-9121-ad4cc32f159e" containerName="horizon" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.823344 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="95fe9a38-2b32-411e-9121-ad4cc32f159e" containerName="horizon-log" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.823354 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f916c05-f172-42b6-9b13-0c8d2058bfb1" containerName="cinder-db-sync" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.824289 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.849839 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.849975 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.850254 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-j7c52" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.850442 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.853562 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.943843 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cc676b85c-q67wf"] Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.954513 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.954773 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-scripts\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.954857 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.954976 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.955116 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkdch\" (UniqueName: \"kubernetes.io/projected/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-kube-api-access-lkdch\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:33 crc kubenswrapper[4632]: I0313 10:26:33.955268 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-config-data\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:33.999064 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f6d96bd7f-txx79"] Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.010035 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.056993 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-config-data\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.057072 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.057100 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-scripts\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.057132 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.057175 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.057220 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkdch\" (UniqueName: \"kubernetes.io/projected/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-kube-api-access-lkdch\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.057561 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.070203 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-config-data\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.070534 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.070931 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-scripts\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.071793 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95fe9a38-2b32-411e-9121-ad4cc32f159e" path="/var/lib/kubelet/pods/95fe9a38-2b32-411e-9121-ad4cc32f159e/volumes" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.078629 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.090058 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f6d96bd7f-txx79"] Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.091225 4632 scope.go:117] "RemoveContainer" containerID="d9db78843b825b24c0eab6345b91a7657d2b3f0bb64d65b5dcc125b1edeeb022" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.120758 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkdch\" (UniqueName: \"kubernetes.io/projected/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-kube-api-access-lkdch\") pod \"cinder-scheduler-0\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.162323 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-dns-swift-storage-0\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.162409 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-ovsdbserver-sb\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.162441 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-dns-svc\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.162484 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7sjz\" (UniqueName: \"kubernetes.io/projected/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-kube-api-access-g7sjz\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.162518 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-config\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.162605 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-ovsdbserver-nb\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.172895 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.264498 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-ovsdbserver-nb\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.264561 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-dns-swift-storage-0\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.264599 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-ovsdbserver-sb\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.264620 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-dns-svc\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.264648 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7sjz\" (UniqueName: \"kubernetes.io/projected/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-kube-api-access-g7sjz\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.264671 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-config\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.265652 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-config\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.266209 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-ovsdbserver-nb\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.266919 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-dns-swift-storage-0\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.267480 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-ovsdbserver-sb\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.267992 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-dns-svc\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.280029 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.281421 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.300158 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.306357 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7sjz\" (UniqueName: \"kubernetes.io/projected/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-kube-api-access-g7sjz\") pod \"dnsmasq-dns-f6d96bd7f-txx79\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.310213 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.341441 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.367608 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.367766 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-scripts\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.367846 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58677e2e-9fc6-4e50-b342-e912afa8d969-etc-machine-id\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.367875 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58677e2e-9fc6-4e50-b342-e912afa8d969-logs\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.368082 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-config-data\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.368161 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-config-data-custom\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.368213 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd7tw\" (UniqueName: \"kubernetes.io/projected/58677e2e-9fc6-4e50-b342-e912afa8d969-kube-api-access-vd7tw\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.472301 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-scripts\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.472372 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58677e2e-9fc6-4e50-b342-e912afa8d969-etc-machine-id\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.472404 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58677e2e-9fc6-4e50-b342-e912afa8d969-logs\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.472460 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-config-data\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.472507 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-config-data-custom\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.472542 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd7tw\" (UniqueName: \"kubernetes.io/projected/58677e2e-9fc6-4e50-b342-e912afa8d969-kube-api-access-vd7tw\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.472578 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.473393 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58677e2e-9fc6-4e50-b342-e912afa8d969-logs\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.481338 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-config-data\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.481421 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58677e2e-9fc6-4e50-b342-e912afa8d969-etc-machine-id\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.489677 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-config-data-custom\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.490029 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-scripts\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.505688 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.521684 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd7tw\" (UniqueName: \"kubernetes.io/projected/58677e2e-9fc6-4e50-b342-e912afa8d969-kube-api-access-vd7tw\") pod \"cinder-api-0\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.529625 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-756c4b86c6-rm274" event={"ID":"dbc1c989-5fa1-46dc-818e-8d609c069e34","Type":"ContainerStarted","Data":"0d4dd1918d120f5fc3a0ac6484d4e6405bafa0d4720b5488a0d85e9bad80fcf2"} Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.555014 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" event={"ID":"51b847ef-ada2-456f-819d-0084fbb17185","Type":"ContainerStarted","Data":"789e66a026cc5bb74f59bfa18247bcf771599597baee3a6d15f80b61a7a1fb39"} Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.631501 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.651498 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5fc9b6f5b5-6ps9m" podStartSLOduration=4.217235668 podStartE2EDuration="10.651472969s" podCreationTimestamp="2026-03-13 10:26:24 +0000 UTC" firstStartedPulling="2026-03-13 10:26:26.02929894 +0000 UTC m=+1360.051829073" lastFinishedPulling="2026-03-13 10:26:32.463536251 +0000 UTC m=+1366.486066374" observedRunningTime="2026-03-13 10:26:34.60962354 +0000 UTC m=+1368.632153693" watchObservedRunningTime="2026-03-13 10:26:34.651472969 +0000 UTC m=+1368.674003122" Mar 13 10:26:34 crc kubenswrapper[4632]: I0313 10:26:34.774751 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="1761ca69-46fd-4375-af60-22b3e77c19a2" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.279288 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.411097 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.411537 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.412331 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"dc4a058f6feb7822333693352f32f5677ff03988b7b5b71005c85c4bf733b402"} pod="openstack/horizon-7bdb5f7878-ng2k2" containerMessage="Container horizon failed startup probe, will be restarted" Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.412377 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" containerID="cri-o://dc4a058f6feb7822333693352f32f5677ff03988b7b5b71005c85c4bf733b402" gracePeriod=30 Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.637494 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f6d96bd7f-txx79"] Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.648221 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" event={"ID":"55712a50-9dcf-44ce-8bac-9aa3ecf65db4","Type":"ContainerStarted","Data":"749192ea37afcdb5bad8f984bb1339eb6de202d1531a18803ce98189920ca65c"} Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.685380 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6e715bfb-1bd5-4c21-ac77-df48fa58a69c","Type":"ContainerStarted","Data":"39e3d26a54ef5266d55d8b5bd910a7757a4fa838dc128c8cd4f4e7a4524e6288"} Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.706817 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" podUID="6cbca281-a753-4810-ab5f-a2d5a5e9c41d" containerName="dnsmasq-dns" containerID="cri-o://d4574f8c63575790d1486fa1f8221ce400d19901d3317c976e702416503cba64" gracePeriod=10 Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.708801 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-756c4b86c6-rm274" event={"ID":"dbc1c989-5fa1-46dc-818e-8d609c069e34","Type":"ContainerStarted","Data":"674891ab2fd374236aa142133f7ffa3c2586bcb2078e41af46e08b0c49dd29e8"} Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.708879 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.708908 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.743094 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.762182 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-756c4b86c6-rm274" podStartSLOduration=6.762159903 podStartE2EDuration="6.762159903s" podCreationTimestamp="2026-03-13 10:26:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:35.752599177 +0000 UTC m=+1369.775129330" watchObservedRunningTime="2026-03-13 10:26:35.762159903 +0000 UTC m=+1369.784690026" Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.858866 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-689764498d-rg7vt" podUID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.858998 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.860040 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"8ce0185281fb59d0c6bda2b2c484ad3711b4bd3b729b4b8677e75ca6b8e1f739"} pod="openstack/horizon-689764498d-rg7vt" containerMessage="Container horizon failed startup probe, will be restarted" Mar 13 10:26:35 crc kubenswrapper[4632]: I0313 10:26:35.860108 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-689764498d-rg7vt" podUID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerName="horizon" containerID="cri-o://8ce0185281fb59d0c6bda2b2c484ad3711b4bd3b729b4b8677e75ca6b8e1f739" gracePeriod=30 Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.665635 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.766105 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-ovsdbserver-nb\") pod \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.766369 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq8dx\" (UniqueName: \"kubernetes.io/projected/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-kube-api-access-sq8dx\") pod \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.766761 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-ovsdbserver-sb\") pod \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.766813 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-config\") pod \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.766837 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-dns-swift-storage-0\") pod \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.767139 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-dns-svc\") pod \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\" (UID: \"6cbca281-a753-4810-ab5f-a2d5a5e9c41d\") " Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.826883 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-kube-api-access-sq8dx" (OuterVolumeSpecName: "kube-api-access-sq8dx") pod "6cbca281-a753-4810-ab5f-a2d5a5e9c41d" (UID: "6cbca281-a753-4810-ab5f-a2d5a5e9c41d"). InnerVolumeSpecName "kube-api-access-sq8dx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.827112 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58677e2e-9fc6-4e50-b342-e912afa8d969","Type":"ContainerStarted","Data":"8037b401a0baaaa45f09498066b3b722d38c4aef73b4ab3874c935fbc21eac6e"} Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.872419 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq8dx\" (UniqueName: \"kubernetes.io/projected/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-kube-api-access-sq8dx\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.903199 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "6cbca281-a753-4810-ab5f-a2d5a5e9c41d" (UID: "6cbca281-a753-4810-ab5f-a2d5a5e9c41d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.928154 4632 generic.go:334] "Generic (PLEG): container finished" podID="6cbca281-a753-4810-ab5f-a2d5a5e9c41d" containerID="d4574f8c63575790d1486fa1f8221ce400d19901d3317c976e702416503cba64" exitCode=0 Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.928360 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.929145 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" event={"ID":"6cbca281-a753-4810-ab5f-a2d5a5e9c41d","Type":"ContainerDied","Data":"d4574f8c63575790d1486fa1f8221ce400d19901d3317c976e702416503cba64"} Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.929176 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cc676b85c-q67wf" event={"ID":"6cbca281-a753-4810-ab5f-a2d5a5e9c41d","Type":"ContainerDied","Data":"c35af663b9fb4942a62d07ad5236c27fcd9454c66f1967196835246c5924112a"} Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.929193 4632 scope.go:117] "RemoveContainer" containerID="d4574f8c63575790d1486fa1f8221ce400d19901d3317c976e702416503cba64" Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.931012 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "6cbca281-a753-4810-ab5f-a2d5a5e9c41d" (UID: "6cbca281-a753-4810-ab5f-a2d5a5e9c41d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.936995 4632 generic.go:334] "Generic (PLEG): container finished" podID="55712a50-9dcf-44ce-8bac-9aa3ecf65db4" containerID="af3fa8988b343c225a97a2143774e273237597ed7c92bf90057d129267e74a5e" exitCode=0 Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.938278 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" event={"ID":"55712a50-9dcf-44ce-8bac-9aa3ecf65db4","Type":"ContainerDied","Data":"af3fa8988b343c225a97a2143774e273237597ed7c92bf90057d129267e74a5e"} Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.989822 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:36 crc kubenswrapper[4632]: I0313 10:26:36.989852 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.003586 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "6cbca281-a753-4810-ab5f-a2d5a5e9c41d" (UID: "6cbca281-a753-4810-ab5f-a2d5a5e9c41d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.009628 4632 scope.go:117] "RemoveContainer" containerID="de515c666eef870de947ba353b522abb4ca9136dc9f866c10cc7d68d392957ce" Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.027669 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-config" (OuterVolumeSpecName: "config") pod "6cbca281-a753-4810-ab5f-a2d5a5e9c41d" (UID: "6cbca281-a753-4810-ab5f-a2d5a5e9c41d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.034449 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "6cbca281-a753-4810-ab5f-a2d5a5e9c41d" (UID: "6cbca281-a753-4810-ab5f-a2d5a5e9c41d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.048341 4632 scope.go:117] "RemoveContainer" containerID="d4574f8c63575790d1486fa1f8221ce400d19901d3317c976e702416503cba64" Mar 13 10:26:37 crc kubenswrapper[4632]: E0313 10:26:37.053091 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4574f8c63575790d1486fa1f8221ce400d19901d3317c976e702416503cba64\": container with ID starting with d4574f8c63575790d1486fa1f8221ce400d19901d3317c976e702416503cba64 not found: ID does not exist" containerID="d4574f8c63575790d1486fa1f8221ce400d19901d3317c976e702416503cba64" Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.053155 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4574f8c63575790d1486fa1f8221ce400d19901d3317c976e702416503cba64"} err="failed to get container status \"d4574f8c63575790d1486fa1f8221ce400d19901d3317c976e702416503cba64\": rpc error: code = NotFound desc = could not find container \"d4574f8c63575790d1486fa1f8221ce400d19901d3317c976e702416503cba64\": container with ID starting with d4574f8c63575790d1486fa1f8221ce400d19901d3317c976e702416503cba64 not found: ID does not exist" Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.053184 4632 scope.go:117] "RemoveContainer" containerID="de515c666eef870de947ba353b522abb4ca9136dc9f866c10cc7d68d392957ce" Mar 13 10:26:37 crc kubenswrapper[4632]: E0313 10:26:37.053667 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de515c666eef870de947ba353b522abb4ca9136dc9f866c10cc7d68d392957ce\": container with ID starting with de515c666eef870de947ba353b522abb4ca9136dc9f866c10cc7d68d392957ce not found: ID does not exist" containerID="de515c666eef870de947ba353b522abb4ca9136dc9f866c10cc7d68d392957ce" Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.053728 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de515c666eef870de947ba353b522abb4ca9136dc9f866c10cc7d68d392957ce"} err="failed to get container status \"de515c666eef870de947ba353b522abb4ca9136dc9f866c10cc7d68d392957ce\": rpc error: code = NotFound desc = could not find container \"de515c666eef870de947ba353b522abb4ca9136dc9f866c10cc7d68d392957ce\": container with ID starting with de515c666eef870de947ba353b522abb4ca9136dc9f866c10cc7d68d392957ce not found: ID does not exist" Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.091170 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.091203 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.091213 4632 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/6cbca281-a753-4810-ab5f-a2d5a5e9c41d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.270424 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cc676b85c-q67wf"] Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.283972 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cc676b85c-q67wf"] Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.812404 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 10:26:37 crc kubenswrapper[4632]: I0313 10:26:37.979453 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58677e2e-9fc6-4e50-b342-e912afa8d969","Type":"ContainerStarted","Data":"a9d0bc7751d471197cb532c1a7e500502d2e1e74a150ed57680796972e393189"} Mar 13 10:26:38 crc kubenswrapper[4632]: I0313 10:26:38.007580 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" event={"ID":"55712a50-9dcf-44ce-8bac-9aa3ecf65db4","Type":"ContainerStarted","Data":"fd4487114042316df9fc87c4e68537674ac28dafb4672e7c807d655817ad05cf"} Mar 13 10:26:38 crc kubenswrapper[4632]: I0313 10:26:38.008828 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:38 crc kubenswrapper[4632]: I0313 10:26:38.100225 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" podStartSLOduration=5.100201899 podStartE2EDuration="5.100201899s" podCreationTimestamp="2026-03-13 10:26:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:38.080818761 +0000 UTC m=+1372.103348894" watchObservedRunningTime="2026-03-13 10:26:38.100201899 +0000 UTC m=+1372.122732052" Mar 13 10:26:38 crc kubenswrapper[4632]: I0313 10:26:38.121224 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cbca281-a753-4810-ab5f-a2d5a5e9c41d" path="/var/lib/kubelet/pods/6cbca281-a753-4810-ab5f-a2d5a5e9c41d/volumes" Mar 13 10:26:38 crc kubenswrapper[4632]: I0313 10:26:38.121866 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6e715bfb-1bd5-4c21-ac77-df48fa58a69c","Type":"ContainerStarted","Data":"ef04a85e3f43e78fc0a1f042d4fba78426911b179ed385ad01f997acf3e9c595"} Mar 13 10:26:39 crc kubenswrapper[4632]: I0313 10:26:39.714196 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-548c8b4b94-2dglr" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.168:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:26:39 crc kubenswrapper[4632]: I0313 10:26:39.714226 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-548c8b4b94-2dglr" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.168:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:26:40 crc kubenswrapper[4632]: I0313 10:26:40.106660 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58677e2e-9fc6-4e50-b342-e912afa8d969","Type":"ContainerStarted","Data":"3a8d9431bb58dc2e36bce7009280ffed0639f98e73ca93dba3c41c03d94fb14f"} Mar 13 10:26:40 crc kubenswrapper[4632]: I0313 10:26:40.106871 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="58677e2e-9fc6-4e50-b342-e912afa8d969" containerName="cinder-api-log" containerID="cri-o://a9d0bc7751d471197cb532c1a7e500502d2e1e74a150ed57680796972e393189" gracePeriod=30 Mar 13 10:26:40 crc kubenswrapper[4632]: I0313 10:26:40.107294 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 13 10:26:40 crc kubenswrapper[4632]: I0313 10:26:40.107643 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="58677e2e-9fc6-4e50-b342-e912afa8d969" containerName="cinder-api" containerID="cri-o://3a8d9431bb58dc2e36bce7009280ffed0639f98e73ca93dba3c41c03d94fb14f" gracePeriod=30 Mar 13 10:26:40 crc kubenswrapper[4632]: I0313 10:26:40.117374 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6e715bfb-1bd5-4c21-ac77-df48fa58a69c","Type":"ContainerStarted","Data":"2d245775b1eef98309f47f0b4c5c25147b2fa7c1b34bae4b51ee329f49499e55"} Mar 13 10:26:40 crc kubenswrapper[4632]: I0313 10:26:40.159690 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=6.159674761 podStartE2EDuration="6.159674761s" podCreationTimestamp="2026-03-13 10:26:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:40.147597415 +0000 UTC m=+1374.170127548" watchObservedRunningTime="2026-03-13 10:26:40.159674761 +0000 UTC m=+1374.182204894" Mar 13 10:26:40 crc kubenswrapper[4632]: I0313 10:26:40.190781 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=6.258534269 podStartE2EDuration="7.190759095s" podCreationTimestamp="2026-03-13 10:26:33 +0000 UTC" firstStartedPulling="2026-03-13 10:26:35.311884579 +0000 UTC m=+1369.334414712" lastFinishedPulling="2026-03-13 10:26:36.244109415 +0000 UTC m=+1370.266639538" observedRunningTime="2026-03-13 10:26:40.178923326 +0000 UTC m=+1374.201453469" watchObservedRunningTime="2026-03-13 10:26:40.190759095 +0000 UTC m=+1374.213289228" Mar 13 10:26:40 crc kubenswrapper[4632]: I0313 10:26:40.648439 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-548c8b4b94-2dglr" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.168:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:26:40 crc kubenswrapper[4632]: I0313 10:26:40.697363 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-548c8b4b94-2dglr" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.168:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:26:41 crc kubenswrapper[4632]: I0313 10:26:41.138267 4632 generic.go:334] "Generic (PLEG): container finished" podID="58677e2e-9fc6-4e50-b342-e912afa8d969" containerID="a9d0bc7751d471197cb532c1a7e500502d2e1e74a150ed57680796972e393189" exitCode=143 Mar 13 10:26:41 crc kubenswrapper[4632]: I0313 10:26:41.138400 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58677e2e-9fc6-4e50-b342-e912afa8d969","Type":"ContainerDied","Data":"a9d0bc7751d471197cb532c1a7e500502d2e1e74a150ed57680796972e393189"} Mar 13 10:26:43 crc kubenswrapper[4632]: I0313 10:26:43.610409 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:44 crc kubenswrapper[4632]: I0313 10:26:44.174674 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 10:26:44 crc kubenswrapper[4632]: I0313 10:26:44.178138 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="6e715bfb-1bd5-4c21-ac77-df48fa58a69c" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.170:8080/\": dial tcp 10.217.0.170:8080: connect: connection refused" Mar 13 10:26:44 crc kubenswrapper[4632]: I0313 10:26:44.346562 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:26:44 crc kubenswrapper[4632]: I0313 10:26:44.447482 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5776d95bfc-hl9dv"] Mar 13 10:26:44 crc kubenswrapper[4632]: I0313 10:26:44.447727 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" podUID="ff547198-2736-4059-8e66-e63ea9ce7345" containerName="dnsmasq-dns" containerID="cri-o://1076485d4d02b6cacd1f94b4c459b88d5309d73c47777ad04b4bed1ee81eb7ff" gracePeriod=10 Mar 13 10:26:44 crc kubenswrapper[4632]: I0313 10:26:44.757713 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-548c8b4b94-2dglr" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.168:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:26:45 crc kubenswrapper[4632]: I0313 10:26:45.197147 4632 generic.go:334] "Generic (PLEG): container finished" podID="ff547198-2736-4059-8e66-e63ea9ce7345" containerID="1076485d4d02b6cacd1f94b4c459b88d5309d73c47777ad04b4bed1ee81eb7ff" exitCode=0 Mar 13 10:26:45 crc kubenswrapper[4632]: I0313 10:26:45.197200 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" event={"ID":"ff547198-2736-4059-8e66-e63ea9ce7345","Type":"ContainerDied","Data":"1076485d4d02b6cacd1f94b4c459b88d5309d73c47777ad04b4bed1ee81eb7ff"} Mar 13 10:26:45 crc kubenswrapper[4632]: I0313 10:26:45.690156 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-548c8b4b94-2dglr" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.168:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:26:45 crc kubenswrapper[4632]: I0313 10:26:45.739238 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-548c8b4b94-2dglr" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.168:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:26:45 crc kubenswrapper[4632]: I0313 10:26:45.739287 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:45 crc kubenswrapper[4632]: I0313 10:26:45.753301 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:46 crc kubenswrapper[4632]: I0313 10:26:46.939132 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-756c4b86c6-rm274" podUID="dbc1c989-5fa1-46dc-818e-8d609c069e34" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.169:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:26:48 crc kubenswrapper[4632]: I0313 10:26:48.615179 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-756c4b86c6-rm274" podUID="dbc1c989-5fa1-46dc-818e-8d609c069e34" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.169:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.099314 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.110888 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.420235 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6db55c595b-pwgcg"] Mar 13 10:26:49 crc kubenswrapper[4632]: E0313 10:26:49.420608 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cbca281-a753-4810-ab5f-a2d5a5e9c41d" containerName="dnsmasq-dns" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.420625 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cbca281-a753-4810-ab5f-a2d5a5e9c41d" containerName="dnsmasq-dns" Mar 13 10:26:49 crc kubenswrapper[4632]: E0313 10:26:49.420641 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cbca281-a753-4810-ab5f-a2d5a5e9c41d" containerName="init" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.420647 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cbca281-a753-4810-ab5f-a2d5a5e9c41d" containerName="init" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.420849 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cbca281-a753-4810-ab5f-a2d5a5e9c41d" containerName="dnsmasq-dns" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.421810 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.457558 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" podUID="ff547198-2736-4059-8e66-e63ea9ce7345" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.158:5353: connect: connection refused" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.482079 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6db55c595b-pwgcg"] Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.549414 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-scripts\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.549466 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-combined-ca-bundle\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.549510 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jms5c\" (UniqueName: \"kubernetes.io/projected/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-kube-api-access-jms5c\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.549570 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-config-data\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.549608 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-internal-tls-certs\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.549630 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-logs\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.549683 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-public-tls-certs\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.653497 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jms5c\" (UniqueName: \"kubernetes.io/projected/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-kube-api-access-jms5c\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.653670 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-config-data\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.653750 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-internal-tls-certs\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.653808 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-logs\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.653916 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-public-tls-certs\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.653975 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-scripts\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.654008 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-combined-ca-bundle\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.656373 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-logs\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.676245 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="58677e2e-9fc6-4e50-b342-e912afa8d969" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.172:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.677833 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-public-tls-certs\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.679246 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-combined-ca-bundle\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.680896 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-config-data\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.689389 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-scripts\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.690042 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-internal-tls-certs\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.713285 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jms5c\" (UniqueName: \"kubernetes.io/projected/ab896d5b-a5b6-46a3-84d8-c3a8c968eac0-kube-api-access-jms5c\") pod \"placement-6db55c595b-pwgcg\" (UID: \"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0\") " pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:49 crc kubenswrapper[4632]: I0313 10:26:49.785557 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:50 crc kubenswrapper[4632]: I0313 10:26:50.041813 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 10:26:50 crc kubenswrapper[4632]: I0313 10:26:50.123736 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 10:26:50 crc kubenswrapper[4632]: I0313 10:26:50.273810 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6e715bfb-1bd5-4c21-ac77-df48fa58a69c" containerName="cinder-scheduler" containerID="cri-o://ef04a85e3f43e78fc0a1f042d4fba78426911b179ed385ad01f997acf3e9c595" gracePeriod=30 Mar 13 10:26:50 crc kubenswrapper[4632]: I0313 10:26:50.273978 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="6e715bfb-1bd5-4c21-ac77-df48fa58a69c" containerName="probe" containerID="cri-o://2d245775b1eef98309f47f0b4c5c25147b2fa7c1b34bae4b51ee329f49499e55" gracePeriod=30 Mar 13 10:26:51 crc kubenswrapper[4632]: E0313 10:26:51.155262 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6e715bfb_1bd5_4c21_ac77_df48fa58a69c.slice/crio-conmon-2d245775b1eef98309f47f0b4c5c25147b2fa7c1b34bae4b51ee329f49499e55.scope\": RecentStats: unable to find data in memory cache]" Mar 13 10:26:51 crc kubenswrapper[4632]: I0313 10:26:51.287223 4632 generic.go:334] "Generic (PLEG): container finished" podID="6e715bfb-1bd5-4c21-ac77-df48fa58a69c" containerID="2d245775b1eef98309f47f0b4c5c25147b2fa7c1b34bae4b51ee329f49499e55" exitCode=0 Mar 13 10:26:51 crc kubenswrapper[4632]: I0313 10:26:51.287607 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6e715bfb-1bd5-4c21-ac77-df48fa58a69c","Type":"ContainerDied","Data":"2d245775b1eef98309f47f0b4c5c25147b2fa7c1b34bae4b51ee329f49499e55"} Mar 13 10:26:51 crc kubenswrapper[4632]: I0313 10:26:51.316026 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-756c4b86c6-rm274" Mar 13 10:26:51 crc kubenswrapper[4632]: I0313 10:26:51.402358 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-548c8b4b94-2dglr"] Mar 13 10:26:51 crc kubenswrapper[4632]: I0313 10:26:51.411197 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-548c8b4b94-2dglr" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api-log" containerID="cri-o://67882325af120e97844e1aef36a358fdd186b89ba1f3def214e49a353ec793aa" gracePeriod=30 Mar 13 10:26:51 crc kubenswrapper[4632]: I0313 10:26:51.412279 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-548c8b4b94-2dglr" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api" containerID="cri-o://309fa94df210d44c275999bad3e9b781bb4f9646e038b1a9463656385d210cf3" gracePeriod=30 Mar 13 10:26:52 crc kubenswrapper[4632]: I0313 10:26:52.302551 4632 generic.go:334] "Generic (PLEG): container finished" podID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerID="67882325af120e97844e1aef36a358fdd186b89ba1f3def214e49a353ec793aa" exitCode=143 Mar 13 10:26:52 crc kubenswrapper[4632]: I0313 10:26:52.302657 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-548c8b4b94-2dglr" event={"ID":"6fa310f1-40ef-4e74-9647-d3ea87858f11","Type":"ContainerDied","Data":"67882325af120e97844e1aef36a358fdd186b89ba1f3def214e49a353ec793aa"} Mar 13 10:26:52 crc kubenswrapper[4632]: I0313 10:26:52.306666 4632 generic.go:334] "Generic (PLEG): container finished" podID="6e715bfb-1bd5-4c21-ac77-df48fa58a69c" containerID="ef04a85e3f43e78fc0a1f042d4fba78426911b179ed385ad01f997acf3e9c595" exitCode=0 Mar 13 10:26:52 crc kubenswrapper[4632]: I0313 10:26:52.306721 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6e715bfb-1bd5-4c21-ac77-df48fa58a69c","Type":"ContainerDied","Data":"ef04a85e3f43e78fc0a1f042d4fba78426911b179ed385ad01f997acf3e9c595"} Mar 13 10:26:52 crc kubenswrapper[4632]: I0313 10:26:52.310961 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 13 10:26:53 crc kubenswrapper[4632]: E0313 10:26:53.488103 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/ubi9/httpd-24:latest" Mar 13 10:26:53 crc kubenswrapper[4632]: E0313 10:26:53.488574 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:proxy-httpd,Image:registry.redhat.io/ubi9/httpd-24:latest,Command:[/usr/sbin/httpd],Args:[-DFOREGROUND],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:proxy-httpd,HostPort:0,ContainerPort:3000,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf/httpd.conf,SubPath:httpd.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/httpd/conf.d/ssl.conf,SubPath:ssl.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:run-httpd,ReadOnly:false,MountPath:/run/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:log-httpd,ReadOnly:false,MountPath:/var/log/httpd,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jz5nk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/,Port:{0 3000 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:30,PeriodSeconds:30,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(270ebc10-986f-4473-8a5e-9094de34ae98): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 10:26:53 crc kubenswrapper[4632]: E0313 10:26:53.489799 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"proxy-httpd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="270ebc10-986f-4473-8a5e-9094de34ae98" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.023900 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.116210 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.160474 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-ovsdbserver-sb\") pod \"ff547198-2736-4059-8e66-e63ea9ce7345\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.160520 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-dns-swift-storage-0\") pod \"ff547198-2736-4059-8e66-e63ea9ce7345\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.160565 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkdch\" (UniqueName: \"kubernetes.io/projected/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-kube-api-access-lkdch\") pod \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.160603 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-config\") pod \"ff547198-2736-4059-8e66-e63ea9ce7345\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.160639 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v4rzv\" (UniqueName: \"kubernetes.io/projected/ff547198-2736-4059-8e66-e63ea9ce7345-kube-api-access-v4rzv\") pod \"ff547198-2736-4059-8e66-e63ea9ce7345\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.160675 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-config-data-custom\") pod \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.160777 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-scripts\") pod \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.160835 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-dns-svc\") pod \"ff547198-2736-4059-8e66-e63ea9ce7345\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.160862 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-etc-machine-id\") pod \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.160909 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-config-data\") pod \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.160926 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-combined-ca-bundle\") pod \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.160975 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-ovsdbserver-nb\") pod \"ff547198-2736-4059-8e66-e63ea9ce7345\" (UID: \"ff547198-2736-4059-8e66-e63ea9ce7345\") " Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.164314 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6e715bfb-1bd5-4c21-ac77-df48fa58a69c" (UID: "6e715bfb-1bd5-4c21-ac77-df48fa58a69c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.172493 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6e715bfb-1bd5-4c21-ac77-df48fa58a69c" (UID: "6e715bfb-1bd5-4c21-ac77-df48fa58a69c"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.172593 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-kube-api-access-lkdch" (OuterVolumeSpecName: "kube-api-access-lkdch") pod "6e715bfb-1bd5-4c21-ac77-df48fa58a69c" (UID: "6e715bfb-1bd5-4c21-ac77-df48fa58a69c"). InnerVolumeSpecName "kube-api-access-lkdch". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.183804 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff547198-2736-4059-8e66-e63ea9ce7345-kube-api-access-v4rzv" (OuterVolumeSpecName: "kube-api-access-v4rzv") pod "ff547198-2736-4059-8e66-e63ea9ce7345" (UID: "ff547198-2736-4059-8e66-e63ea9ce7345"). InnerVolumeSpecName "kube-api-access-v4rzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.184804 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-scripts" (OuterVolumeSpecName: "scripts") pod "6e715bfb-1bd5-4c21-ac77-df48fa58a69c" (UID: "6e715bfb-1bd5-4c21-ac77-df48fa58a69c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.262134 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e715bfb-1bd5-4c21-ac77-df48fa58a69c" (UID: "6e715bfb-1bd5-4c21-ac77-df48fa58a69c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.262578 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-combined-ca-bundle\") pod \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\" (UID: \"6e715bfb-1bd5-4c21-ac77-df48fa58a69c\") " Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.263107 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkdch\" (UniqueName: \"kubernetes.io/projected/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-kube-api-access-lkdch\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.263132 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v4rzv\" (UniqueName: \"kubernetes.io/projected/ff547198-2736-4059-8e66-e63ea9ce7345-kube-api-access-v4rzv\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.263144 4632 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.263155 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.263166 4632 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:54 crc kubenswrapper[4632]: W0313 10:26:54.263245 4632 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/6e715bfb-1bd5-4c21-ac77-df48fa58a69c/volumes/kubernetes.io~secret/combined-ca-bundle Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.263255 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6e715bfb-1bd5-4c21-ac77-df48fa58a69c" (UID: "6e715bfb-1bd5-4c21-ac77-df48fa58a69c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.268215 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6db55c595b-pwgcg"] Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.275832 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ff547198-2736-4059-8e66-e63ea9ce7345" (UID: "ff547198-2736-4059-8e66-e63ea9ce7345"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.277304 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ff547198-2736-4059-8e66-e63ea9ce7345" (UID: "ff547198-2736-4059-8e66-e63ea9ce7345"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.293837 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.294206 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ff547198-2736-4059-8e66-e63ea9ce7345" (UID: "ff547198-2736-4059-8e66-e63ea9ce7345"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.331195 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ff547198-2736-4059-8e66-e63ea9ce7345" (UID: "ff547198-2736-4059-8e66-e63ea9ce7345"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.337024 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" event={"ID":"ff547198-2736-4059-8e66-e63ea9ce7345","Type":"ContainerDied","Data":"8db6fac31f3928e6490a77faa8cf72ab51791153ec4fce9dafd1cd9fb950c31f"} Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.337063 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5776d95bfc-hl9dv" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.337091 4632 scope.go:117] "RemoveContainer" containerID="1076485d4d02b6cacd1f94b4c459b88d5309d73c47777ad04b4bed1ee81eb7ff" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.340438 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"6e715bfb-1bd5-4c21-ac77-df48fa58a69c","Type":"ContainerDied","Data":"39e3d26a54ef5266d55d8b5bd910a7757a4fa838dc128c8cd4f4e7a4524e6288"} Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.340554 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.347535 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="270ebc10-986f-4473-8a5e-9094de34ae98" containerName="ceilometer-central-agent" containerID="cri-o://27c121915dbbdfc336d1bc55bed50eb5edaf76e1bc92f4f6b5e249f4ffe5098a" gracePeriod=30 Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.347834 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6db55c595b-pwgcg" event={"ID":"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0","Type":"ContainerStarted","Data":"cd52fc862b4f56e1e6058f382a29e8086f42b20e58a9788e92c543c0f3389cd6"} Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.347892 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="270ebc10-986f-4473-8a5e-9094de34ae98" containerName="sg-core" containerID="cri-o://d138d976167695fe9d299247eefcff55845f7ad27e84fc81cc086274294f2e51" gracePeriod=30 Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.347966 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="270ebc10-986f-4473-8a5e-9094de34ae98" containerName="ceilometer-notification-agent" containerID="cri-o://b63cc4f80efbb7b17b044808a5b6c8d5aa98b9e2ae8e38ab95a55c4e3ba911d1" gracePeriod=30 Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.365253 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-config" (OuterVolumeSpecName: "config") pod "ff547198-2736-4059-8e66-e63ea9ce7345" (UID: "ff547198-2736-4059-8e66-e63ea9ce7345"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.365922 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.365956 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.365972 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.365984 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.365995 4632 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.366006 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ff547198-2736-4059-8e66-e63ea9ce7345-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.385398 4632 scope.go:117] "RemoveContainer" containerID="dec2a00b325c16f4a1d001f23d5e8b1ffdb30f4c935f90c479b4c2928a1f9cbd" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.425994 4632 scope.go:117] "RemoveContainer" containerID="2d245775b1eef98309f47f0b4c5c25147b2fa7c1b34bae4b51ee329f49499e55" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.446060 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-config-data" (OuterVolumeSpecName: "config-data") pod "6e715bfb-1bd5-4c21-ac77-df48fa58a69c" (UID: "6e715bfb-1bd5-4c21-ac77-df48fa58a69c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.467347 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e715bfb-1bd5-4c21-ac77-df48fa58a69c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.521034 4632 scope.go:117] "RemoveContainer" containerID="ef04a85e3f43e78fc0a1f042d4fba78426911b179ed385ad01f997acf3e9c595" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.645714 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-64bdffbb5c-mpfvf"] Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.646081 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-64bdffbb5c-mpfvf" podUID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" containerName="neutron-api" containerID="cri-o://027b2c4436a3d137f7ef6a7921904bf128e17aa7812143af60d4d11a546759da" gracePeriod=30 Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.646252 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-64bdffbb5c-mpfvf" podUID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" containerName="neutron-httpd" containerID="cri-o://bbc256375bc79a61ff656574ec8a596aed3314e7ad4cd2f7fcf6a7462aee3274" gracePeriod=30 Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.723205 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6588559b77-6f4bf"] Mar 13 10:26:54 crc kubenswrapper[4632]: E0313 10:26:54.731329 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff547198-2736-4059-8e66-e63ea9ce7345" containerName="dnsmasq-dns" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.731541 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff547198-2736-4059-8e66-e63ea9ce7345" containerName="dnsmasq-dns" Mar 13 10:26:54 crc kubenswrapper[4632]: E0313 10:26:54.731617 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e715bfb-1bd5-4c21-ac77-df48fa58a69c" containerName="cinder-scheduler" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.731675 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e715bfb-1bd5-4c21-ac77-df48fa58a69c" containerName="cinder-scheduler" Mar 13 10:26:54 crc kubenswrapper[4632]: E0313 10:26:54.731751 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e715bfb-1bd5-4c21-ac77-df48fa58a69c" containerName="probe" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.731809 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e715bfb-1bd5-4c21-ac77-df48fa58a69c" containerName="probe" Mar 13 10:26:54 crc kubenswrapper[4632]: E0313 10:26:54.731873 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff547198-2736-4059-8e66-e63ea9ce7345" containerName="init" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.731924 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff547198-2736-4059-8e66-e63ea9ce7345" containerName="init" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.732151 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff547198-2736-4059-8e66-e63ea9ce7345" containerName="dnsmasq-dns" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.732225 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e715bfb-1bd5-4c21-ac77-df48fa58a69c" containerName="probe" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.732287 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e715bfb-1bd5-4c21-ac77-df48fa58a69c" containerName="cinder-scheduler" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.733323 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.745245 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.782457 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-httpd-config\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.782507 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-combined-ca-bundle\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.782592 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-config\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.782684 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-public-tls-certs\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.782716 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-internal-tls-certs\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.782764 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g98kg\" (UniqueName: \"kubernetes.io/projected/79498b99-6b5c-4a95-8558-5d615fc7abba-kube-api-access-g98kg\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.782823 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-ovndb-tls-certs\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.800161 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6588559b77-6f4bf"] Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.837335 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5776d95bfc-hl9dv"] Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.855485 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5776d95bfc-hl9dv"] Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.875126 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.886251 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-ovndb-tls-certs\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.886328 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-httpd-config\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.886356 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-combined-ca-bundle\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.886403 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-config\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.886489 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-public-tls-certs\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.886528 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-internal-tls-certs\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.886576 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g98kg\" (UniqueName: \"kubernetes.io/projected/79498b99-6b5c-4a95-8558-5d615fc7abba-kube-api-access-g98kg\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.894405 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.916383 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-internal-tls-certs\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.916907 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-public-tls-certs\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.917647 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-httpd-config\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.920813 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-ovndb-tls-certs\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.926091 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-config\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.945800 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-combined-ca-bundle\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.954009 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g98kg\" (UniqueName: \"kubernetes.io/projected/79498b99-6b5c-4a95-8558-5d615fc7abba-kube-api-access-g98kg\") pod \"neutron-6588559b77-6f4bf\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.974006 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.975459 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 10:26:54 crc kubenswrapper[4632]: I0313 10:26:54.980262 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.082650 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.091579 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2c1c19b-95a5-4db1-8e54-36fe83704b25-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.091621 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2c1c19b-95a5-4db1-8e54-36fe83704b25-scripts\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.091674 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn7fb\" (UniqueName: \"kubernetes.io/projected/d2c1c19b-95a5-4db1-8e54-36fe83704b25-kube-api-access-qn7fb\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.091697 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2c1c19b-95a5-4db1-8e54-36fe83704b25-config-data\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.091758 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2c1c19b-95a5-4db1-8e54-36fe83704b25-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.091791 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2c1c19b-95a5-4db1-8e54-36fe83704b25-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.112086 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.194366 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2c1c19b-95a5-4db1-8e54-36fe83704b25-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.194416 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2c1c19b-95a5-4db1-8e54-36fe83704b25-scripts\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.194474 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn7fb\" (UniqueName: \"kubernetes.io/projected/d2c1c19b-95a5-4db1-8e54-36fe83704b25-kube-api-access-qn7fb\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.194881 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2c1c19b-95a5-4db1-8e54-36fe83704b25-config-data\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.194956 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2c1c19b-95a5-4db1-8e54-36fe83704b25-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.194984 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2c1c19b-95a5-4db1-8e54-36fe83704b25-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.197489 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d2c1c19b-95a5-4db1-8e54-36fe83704b25-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.215790 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d2c1c19b-95a5-4db1-8e54-36fe83704b25-scripts\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.216300 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d2c1c19b-95a5-4db1-8e54-36fe83704b25-config-data\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.216653 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d2c1c19b-95a5-4db1-8e54-36fe83704b25-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.218428 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn7fb\" (UniqueName: \"kubernetes.io/projected/d2c1c19b-95a5-4db1-8e54-36fe83704b25-kube-api-access-qn7fb\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.222464 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d2c1c19b-95a5-4db1-8e54-36fe83704b25-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"d2c1c19b-95a5-4db1-8e54-36fe83704b25\") " pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.325348 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.381322 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6db55c595b-pwgcg" event={"ID":"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0","Type":"ContainerStarted","Data":"53916000669a99bf8cb4a1d3bdcfa4c5fdf4d945c1bc57cbcc0a66a10b039644"} Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.419356 4632 generic.go:334] "Generic (PLEG): container finished" podID="270ebc10-986f-4473-8a5e-9094de34ae98" containerID="d138d976167695fe9d299247eefcff55845f7ad27e84fc81cc086274294f2e51" exitCode=2 Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.419389 4632 generic.go:334] "Generic (PLEG): container finished" podID="270ebc10-986f-4473-8a5e-9094de34ae98" containerID="27c121915dbbdfc336d1bc55bed50eb5edaf76e1bc92f4f6b5e249f4ffe5098a" exitCode=0 Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.419448 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"270ebc10-986f-4473-8a5e-9094de34ae98","Type":"ContainerDied","Data":"d138d976167695fe9d299247eefcff55845f7ad27e84fc81cc086274294f2e51"} Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.419494 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"270ebc10-986f-4473-8a5e-9094de34ae98","Type":"ContainerDied","Data":"27c121915dbbdfc336d1bc55bed50eb5edaf76e1bc92f4f6b5e249f4ffe5098a"} Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.423490 4632 generic.go:334] "Generic (PLEG): container finished" podID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerID="309fa94df210d44c275999bad3e9b781bb4f9646e038b1a9463656385d210cf3" exitCode=0 Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.423567 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-548c8b4b94-2dglr" event={"ID":"6fa310f1-40ef-4e74-9647-d3ea87858f11","Type":"ContainerDied","Data":"309fa94df210d44c275999bad3e9b781bb4f9646e038b1a9463656385d210cf3"} Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.723428 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.811143 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-config-data-custom\") pod \"6fa310f1-40ef-4e74-9647-d3ea87858f11\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.811242 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fa310f1-40ef-4e74-9647-d3ea87858f11-logs\") pod \"6fa310f1-40ef-4e74-9647-d3ea87858f11\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.811320 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pmkj\" (UniqueName: \"kubernetes.io/projected/6fa310f1-40ef-4e74-9647-d3ea87858f11-kube-api-access-8pmkj\") pod \"6fa310f1-40ef-4e74-9647-d3ea87858f11\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.811350 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-config-data\") pod \"6fa310f1-40ef-4e74-9647-d3ea87858f11\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.811368 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-combined-ca-bundle\") pod \"6fa310f1-40ef-4e74-9647-d3ea87858f11\" (UID: \"6fa310f1-40ef-4e74-9647-d3ea87858f11\") " Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.811988 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6fa310f1-40ef-4e74-9647-d3ea87858f11-logs" (OuterVolumeSpecName: "logs") pod "6fa310f1-40ef-4e74-9647-d3ea87858f11" (UID: "6fa310f1-40ef-4e74-9647-d3ea87858f11"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.837478 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6fa310f1-40ef-4e74-9647-d3ea87858f11" (UID: "6fa310f1-40ef-4e74-9647-d3ea87858f11"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.837890 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6fa310f1-40ef-4e74-9647-d3ea87858f11-kube-api-access-8pmkj" (OuterVolumeSpecName: "kube-api-access-8pmkj") pod "6fa310f1-40ef-4e74-9647-d3ea87858f11" (UID: "6fa310f1-40ef-4e74-9647-d3ea87858f11"). InnerVolumeSpecName "kube-api-access-8pmkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.894239 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6fa310f1-40ef-4e74-9647-d3ea87858f11" (UID: "6fa310f1-40ef-4e74-9647-d3ea87858f11"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.916982 4632 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.917014 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6fa310f1-40ef-4e74-9647-d3ea87858f11-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.917024 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pmkj\" (UniqueName: \"kubernetes.io/projected/6fa310f1-40ef-4e74-9647-d3ea87858f11-kube-api-access-8pmkj\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.917035 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:55 crc kubenswrapper[4632]: I0313 10:26:55.963330 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-config-data" (OuterVolumeSpecName: "config-data") pod "6fa310f1-40ef-4e74-9647-d3ea87858f11" (UID: "6fa310f1-40ef-4e74-9647-d3ea87858f11"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.021014 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6fa310f1-40ef-4e74-9647-d3ea87858f11-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.085596 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e715bfb-1bd5-4c21-ac77-df48fa58a69c" path="/var/lib/kubelet/pods/6e715bfb-1bd5-4c21-ac77-df48fa58a69c/volumes" Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.087455 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff547198-2736-4059-8e66-e63ea9ce7345" path="/var/lib/kubelet/pods/ff547198-2736-4059-8e66-e63ea9ce7345/volumes" Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.095631 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6588559b77-6f4bf"] Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.114844 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 10:26:56 crc kubenswrapper[4632]: W0313 10:26:56.121528 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79498b99_6b5c_4a95_8558_5d615fc7abba.slice/crio-cc3aa5e44b0dc25bbbe479e7210125c65a55be4449da25f59fdeef0322a73ed3 WatchSource:0}: Error finding container cc3aa5e44b0dc25bbbe479e7210125c65a55be4449da25f59fdeef0322a73ed3: Status 404 returned error can't find the container with id cc3aa5e44b0dc25bbbe479e7210125c65a55be4449da25f59fdeef0322a73ed3 Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.453794 4632 generic.go:334] "Generic (PLEG): container finished" podID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" containerID="bbc256375bc79a61ff656574ec8a596aed3314e7ad4cd2f7fcf6a7462aee3274" exitCode=0 Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.453869 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64bdffbb5c-mpfvf" event={"ID":"6c867fc1-05ed-46c3-99dc-71ef8a09dad3","Type":"ContainerDied","Data":"bbc256375bc79a61ff656574ec8a596aed3314e7ad4cd2f7fcf6a7462aee3274"} Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.455527 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d2c1c19b-95a5-4db1-8e54-36fe83704b25","Type":"ContainerStarted","Data":"9fdece3800c287c4e18e7f493209526ed799ae59bb9b012c7f57b96117c81a49"} Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.456744 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6588559b77-6f4bf" event={"ID":"79498b99-6b5c-4a95-8558-5d615fc7abba","Type":"ContainerStarted","Data":"cc3aa5e44b0dc25bbbe479e7210125c65a55be4449da25f59fdeef0322a73ed3"} Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.459015 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6db55c595b-pwgcg" event={"ID":"ab896d5b-a5b6-46a3-84d8-c3a8c968eac0","Type":"ContainerStarted","Data":"3f503ce1a453ee49856abaa4a3d77ea00e6382fc8095631e76a224e4d7cf8ac2"} Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.459528 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.459569 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.478187 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-548c8b4b94-2dglr" event={"ID":"6fa310f1-40ef-4e74-9647-d3ea87858f11","Type":"ContainerDied","Data":"4545ac42523c98f674d28d5d0acc10645d2b1e7d8486b7068d13c265711710a4"} Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.478246 4632 scope.go:117] "RemoveContainer" containerID="309fa94df210d44c275999bad3e9b781bb4f9646e038b1a9463656385d210cf3" Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.478258 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-548c8b4b94-2dglr" Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.492173 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6db55c595b-pwgcg" podStartSLOduration=7.492142267 podStartE2EDuration="7.492142267s" podCreationTimestamp="2026-03-13 10:26:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:56.489266179 +0000 UTC m=+1390.511796312" watchObservedRunningTime="2026-03-13 10:26:56.492142267 +0000 UTC m=+1390.514672400" Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.573716 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-548c8b4b94-2dglr"] Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.588170 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-548c8b4b94-2dglr"] Mar 13 10:26:56 crc kubenswrapper[4632]: I0313 10:26:56.621190 4632 scope.go:117] "RemoveContainer" containerID="67882325af120e97844e1aef36a358fdd186b89ba1f3def214e49a353ec793aa" Mar 13 10:26:57 crc kubenswrapper[4632]: I0313 10:26:57.304590 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-64bdffbb5c-mpfvf" podUID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.159:9696/\": dial tcp 10.217.0.159:9696: connect: connection refused" Mar 13 10:26:57 crc kubenswrapper[4632]: I0313 10:26:57.530011 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d2c1c19b-95a5-4db1-8e54-36fe83704b25","Type":"ContainerStarted","Data":"67516bb124d863acdb93cbafff12001c1c53c2a821587b0e3e99f6135ee28e92"} Mar 13 10:26:57 crc kubenswrapper[4632]: I0313 10:26:57.544034 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6588559b77-6f4bf" event={"ID":"79498b99-6b5c-4a95-8558-5d615fc7abba","Type":"ContainerStarted","Data":"2e4dbe726a115e20d5697b52cbd987856c78465356a65ffaf180382482e42ad0"} Mar 13 10:26:57 crc kubenswrapper[4632]: I0313 10:26:57.544098 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6588559b77-6f4bf" event={"ID":"79498b99-6b5c-4a95-8558-5d615fc7abba","Type":"ContainerStarted","Data":"a37056b823559676b78bbdad36e07fb68a02ab13bf670546d16508926857a154"} Mar 13 10:26:57 crc kubenswrapper[4632]: I0313 10:26:57.544717 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:26:57 crc kubenswrapper[4632]: I0313 10:26:57.569279 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6588559b77-6f4bf" podStartSLOduration=3.569255977 podStartE2EDuration="3.569255977s" podCreationTimestamp="2026-03-13 10:26:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:57.562276862 +0000 UTC m=+1391.584807015" watchObservedRunningTime="2026-03-13 10:26:57.569255977 +0000 UTC m=+1391.591786120" Mar 13 10:26:58 crc kubenswrapper[4632]: I0313 10:26:58.055710 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" path="/var/lib/kubelet/pods/6fa310f1-40ef-4e74-9647-d3ea87858f11/volumes" Mar 13 10:26:59 crc kubenswrapper[4632]: I0313 10:26:59.579065 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d2c1c19b-95a5-4db1-8e54-36fe83704b25","Type":"ContainerStarted","Data":"5eee9458e5025ac2fc90f250d59097faf32248e3b261634e845944e47ef32ad2"} Mar 13 10:26:59 crc kubenswrapper[4632]: I0313 10:26:59.630182 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.630158762 podStartE2EDuration="5.630158762s" podCreationTimestamp="2026-03-13 10:26:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:26:59.625310737 +0000 UTC m=+1393.647840880" watchObservedRunningTime="2026-03-13 10:26:59.630158762 +0000 UTC m=+1393.652688885" Mar 13 10:26:59 crc kubenswrapper[4632]: I0313 10:26:59.735569 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-f664b756d-8fxf4" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.325881 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.525781 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 10:27:00 crc kubenswrapper[4632]: E0313 10:27:00.526646 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.526673 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api" Mar 13 10:27:00 crc kubenswrapper[4632]: E0313 10:27:00.526695 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api-log" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.526704 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api-log" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.526991 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.527037 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api-log" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.527830 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.530162 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.530679 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.536862 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-gvm2d" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.553049 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.601090 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-548c8b4b94-2dglr" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.168:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.603345 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-548c8b4b94-2dglr" podUID="6fa310f1-40ef-4e74-9647-d3ea87858f11" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.168:9311/healthcheck\": dial tcp 10.217.0.168:9311: i/o timeout (Client.Timeout exceeded while awaiting headers)" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.625576 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8adc4254-ad10-4335-a365-876324d1af24-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " pod="openstack/openstackclient" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.625658 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzl69\" (UniqueName: \"kubernetes.io/projected/8adc4254-ad10-4335-a365-876324d1af24-kube-api-access-pzl69\") pod \"openstackclient\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " pod="openstack/openstackclient" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.625690 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8adc4254-ad10-4335-a365-876324d1af24-openstack-config-secret\") pod \"openstackclient\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " pod="openstack/openstackclient" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.625764 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8adc4254-ad10-4335-a365-876324d1af24-openstack-config\") pod \"openstackclient\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " pod="openstack/openstackclient" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.727226 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8adc4254-ad10-4335-a365-876324d1af24-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " pod="openstack/openstackclient" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.727505 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzl69\" (UniqueName: \"kubernetes.io/projected/8adc4254-ad10-4335-a365-876324d1af24-kube-api-access-pzl69\") pod \"openstackclient\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " pod="openstack/openstackclient" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.727524 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8adc4254-ad10-4335-a365-876324d1af24-openstack-config-secret\") pod \"openstackclient\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " pod="openstack/openstackclient" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.727588 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8adc4254-ad10-4335-a365-876324d1af24-openstack-config\") pod \"openstackclient\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " pod="openstack/openstackclient" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.730123 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8adc4254-ad10-4335-a365-876324d1af24-openstack-config\") pod \"openstackclient\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " pod="openstack/openstackclient" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.743582 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8adc4254-ad10-4335-a365-876324d1af24-combined-ca-bundle\") pod \"openstackclient\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " pod="openstack/openstackclient" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.744771 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8adc4254-ad10-4335-a365-876324d1af24-openstack-config-secret\") pod \"openstackclient\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " pod="openstack/openstackclient" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.751875 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzl69\" (UniqueName: \"kubernetes.io/projected/8adc4254-ad10-4335-a365-876324d1af24-kube-api-access-pzl69\") pod \"openstackclient\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " pod="openstack/openstackclient" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.857509 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.946582 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Mar 13 10:27:00 crc kubenswrapper[4632]: I0313 10:27:00.964294 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.015233 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.017438 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.037290 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.137330 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmjms\" (UniqueName: \"kubernetes.io/projected/aef9680f-df77-4e2e-ac53-9d7530c2270c-kube-api-access-cmjms\") pod \"openstackclient\" (UID: \"aef9680f-df77-4e2e-ac53-9d7530c2270c\") " pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.137423 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aef9680f-df77-4e2e-ac53-9d7530c2270c-openstack-config-secret\") pod \"openstackclient\" (UID: \"aef9680f-df77-4e2e-ac53-9d7530c2270c\") " pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.137491 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aef9680f-df77-4e2e-ac53-9d7530c2270c-openstack-config\") pod \"openstackclient\" (UID: \"aef9680f-df77-4e2e-ac53-9d7530c2270c\") " pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.137546 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef9680f-df77-4e2e-ac53-9d7530c2270c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aef9680f-df77-4e2e-ac53-9d7530c2270c\") " pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.240033 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aef9680f-df77-4e2e-ac53-9d7530c2270c-openstack-config-secret\") pod \"openstackclient\" (UID: \"aef9680f-df77-4e2e-ac53-9d7530c2270c\") " pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.240431 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aef9680f-df77-4e2e-ac53-9d7530c2270c-openstack-config\") pod \"openstackclient\" (UID: \"aef9680f-df77-4e2e-ac53-9d7530c2270c\") " pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.240494 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef9680f-df77-4e2e-ac53-9d7530c2270c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aef9680f-df77-4e2e-ac53-9d7530c2270c\") " pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.240598 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmjms\" (UniqueName: \"kubernetes.io/projected/aef9680f-df77-4e2e-ac53-9d7530c2270c-kube-api-access-cmjms\") pod \"openstackclient\" (UID: \"aef9680f-df77-4e2e-ac53-9d7530c2270c\") " pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.242273 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/aef9680f-df77-4e2e-ac53-9d7530c2270c-openstack-config\") pod \"openstackclient\" (UID: \"aef9680f-df77-4e2e-ac53-9d7530c2270c\") " pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.247206 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef9680f-df77-4e2e-ac53-9d7530c2270c-combined-ca-bundle\") pod \"openstackclient\" (UID: \"aef9680f-df77-4e2e-ac53-9d7530c2270c\") " pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: E0313 10:27:01.263624 4632 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 13 10:27:01 crc kubenswrapper[4632]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_8adc4254-ad10-4335-a365-876324d1af24_0(fd110ce1a0d6e4d5ad13c3bd776d304884dcdd79911e2024e5cf26000b535d37): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"fd110ce1a0d6e4d5ad13c3bd776d304884dcdd79911e2024e5cf26000b535d37" Netns:"/var/run/netns/4de97584-a76a-4eaa-8a9c-1af1c056a2a4" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=fd110ce1a0d6e4d5ad13c3bd776d304884dcdd79911e2024e5cf26000b535d37;K8S_POD_UID=8adc4254-ad10-4335-a365-876324d1af24" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/8adc4254-ad10-4335-a365-876324d1af24]: expected pod UID "8adc4254-ad10-4335-a365-876324d1af24" but got "aef9680f-df77-4e2e-ac53-9d7530c2270c" from Kube API Mar 13 10:27:01 crc kubenswrapper[4632]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 10:27:01 crc kubenswrapper[4632]: > Mar 13 10:27:01 crc kubenswrapper[4632]: E0313 10:27:01.263736 4632 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 13 10:27:01 crc kubenswrapper[4632]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_8adc4254-ad10-4335-a365-876324d1af24_0(fd110ce1a0d6e4d5ad13c3bd776d304884dcdd79911e2024e5cf26000b535d37): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"fd110ce1a0d6e4d5ad13c3bd776d304884dcdd79911e2024e5cf26000b535d37" Netns:"/var/run/netns/4de97584-a76a-4eaa-8a9c-1af1c056a2a4" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=fd110ce1a0d6e4d5ad13c3bd776d304884dcdd79911e2024e5cf26000b535d37;K8S_POD_UID=8adc4254-ad10-4335-a365-876324d1af24" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/8adc4254-ad10-4335-a365-876324d1af24]: expected pod UID "8adc4254-ad10-4335-a365-876324d1af24" but got "aef9680f-df77-4e2e-ac53-9d7530c2270c" from Kube API Mar 13 10:27:01 crc kubenswrapper[4632]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 10:27:01 crc kubenswrapper[4632]: > pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.264318 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/aef9680f-df77-4e2e-ac53-9d7530c2270c-openstack-config-secret\") pod \"openstackclient\" (UID: \"aef9680f-df77-4e2e-ac53-9d7530c2270c\") " pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.281711 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmjms\" (UniqueName: \"kubernetes.io/projected/aef9680f-df77-4e2e-ac53-9d7530c2270c-kube-api-access-cmjms\") pod \"openstackclient\" (UID: \"aef9680f-df77-4e2e-ac53-9d7530c2270c\") " pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.343200 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.659169 4632 generic.go:334] "Generic (PLEG): container finished" podID="270ebc10-986f-4473-8a5e-9094de34ae98" containerID="b63cc4f80efbb7b17b044808a5b6c8d5aa98b9e2ae8e38ab95a55c4e3ba911d1" exitCode=0 Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.659615 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"270ebc10-986f-4473-8a5e-9094de34ae98","Type":"ContainerDied","Data":"b63cc4f80efbb7b17b044808a5b6c8d5aa98b9e2ae8e38ab95a55c4e3ba911d1"} Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.663637 4632 generic.go:334] "Generic (PLEG): container finished" podID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" containerID="027b2c4436a3d137f7ef6a7921904bf128e17aa7812143af60d4d11a546759da" exitCode=0 Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.663732 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.664357 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64bdffbb5c-mpfvf" event={"ID":"6c867fc1-05ed-46c3-99dc-71ef8a09dad3","Type":"ContainerDied","Data":"027b2c4436a3d137f7ef6a7921904bf128e17aa7812143af60d4d11a546759da"} Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.670738 4632 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="8adc4254-ad10-4335-a365-876324d1af24" podUID="aef9680f-df77-4e2e-ac53-9d7530c2270c" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.682125 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.753466 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8adc4254-ad10-4335-a365-876324d1af24-openstack-config-secret\") pod \"8adc4254-ad10-4335-a365-876324d1af24\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.754728 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8adc4254-ad10-4335-a365-876324d1af24-combined-ca-bundle\") pod \"8adc4254-ad10-4335-a365-876324d1af24\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.754862 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8adc4254-ad10-4335-a365-876324d1af24-openstack-config\") pod \"8adc4254-ad10-4335-a365-876324d1af24\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.755088 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pzl69\" (UniqueName: \"kubernetes.io/projected/8adc4254-ad10-4335-a365-876324d1af24-kube-api-access-pzl69\") pod \"8adc4254-ad10-4335-a365-876324d1af24\" (UID: \"8adc4254-ad10-4335-a365-876324d1af24\") " Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.758378 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8adc4254-ad10-4335-a365-876324d1af24-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "8adc4254-ad10-4335-a365-876324d1af24" (UID: "8adc4254-ad10-4335-a365-876324d1af24"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.762223 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8adc4254-ad10-4335-a365-876324d1af24-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "8adc4254-ad10-4335-a365-876324d1af24" (UID: "8adc4254-ad10-4335-a365-876324d1af24"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.765225 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8adc4254-ad10-4335-a365-876324d1af24-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8adc4254-ad10-4335-a365-876324d1af24" (UID: "8adc4254-ad10-4335-a365-876324d1af24"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.768136 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8adc4254-ad10-4335-a365-876324d1af24-kube-api-access-pzl69" (OuterVolumeSpecName: "kube-api-access-pzl69") pod "8adc4254-ad10-4335-a365-876324d1af24" (UID: "8adc4254-ad10-4335-a365-876324d1af24"). InnerVolumeSpecName "kube-api-access-pzl69". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.859115 4632 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/8adc4254-ad10-4335-a365-876324d1af24-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.859417 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8adc4254-ad10-4335-a365-876324d1af24-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.859427 4632 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/8adc4254-ad10-4335-a365-876324d1af24-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.859436 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pzl69\" (UniqueName: \"kubernetes.io/projected/8adc4254-ad10-4335-a365-876324d1af24-kube-api-access-pzl69\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:01 crc kubenswrapper[4632]: I0313 10:27:01.920378 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 10:27:01 crc kubenswrapper[4632]: W0313 10:27:01.934092 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaef9680f_df77_4e2e_ac53_9d7530c2270c.slice/crio-d29ea12656ced9b9702cd194ca8790d076046a2079e8c6f2febbe1128ab32a1a WatchSource:0}: Error finding container d29ea12656ced9b9702cd194ca8790d076046a2079e8c6f2febbe1128ab32a1a: Status 404 returned error can't find the container with id d29ea12656ced9b9702cd194ca8790d076046a2079e8c6f2febbe1128ab32a1a Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.082083 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8adc4254-ad10-4335-a365-876324d1af24" path="/var/lib/kubelet/pods/8adc4254-ad10-4335-a365-876324d1af24/volumes" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.299051 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.310381 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.397607 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-config-data\") pod \"270ebc10-986f-4473-8a5e-9094de34ae98\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.397656 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-combined-ca-bundle\") pod \"270ebc10-986f-4473-8a5e-9094de34ae98\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.397674 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-httpd-config\") pod \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.397694 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-scripts\") pod \"270ebc10-986f-4473-8a5e-9094de34ae98\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.397713 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/270ebc10-986f-4473-8a5e-9094de34ae98-log-httpd\") pod \"270ebc10-986f-4473-8a5e-9094de34ae98\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.397760 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jz5nk\" (UniqueName: \"kubernetes.io/projected/270ebc10-986f-4473-8a5e-9094de34ae98-kube-api-access-jz5nk\") pod \"270ebc10-986f-4473-8a5e-9094de34ae98\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.397818 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/270ebc10-986f-4473-8a5e-9094de34ae98-run-httpd\") pod \"270ebc10-986f-4473-8a5e-9094de34ae98\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.397857 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-internal-tls-certs\") pod \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.397885 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-combined-ca-bundle\") pod \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.397905 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbbzk\" (UniqueName: \"kubernetes.io/projected/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-kube-api-access-rbbzk\") pod \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.397928 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-ovndb-tls-certs\") pod \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.397984 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-public-tls-certs\") pod \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.398032 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-config\") pod \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\" (UID: \"6c867fc1-05ed-46c3-99dc-71ef8a09dad3\") " Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.398054 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-sg-core-conf-yaml\") pod \"270ebc10-986f-4473-8a5e-9094de34ae98\" (UID: \"270ebc10-986f-4473-8a5e-9094de34ae98\") " Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.399164 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/270ebc10-986f-4473-8a5e-9094de34ae98-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "270ebc10-986f-4473-8a5e-9094de34ae98" (UID: "270ebc10-986f-4473-8a5e-9094de34ae98"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.403441 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/270ebc10-986f-4473-8a5e-9094de34ae98-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "270ebc10-986f-4473-8a5e-9094de34ae98" (UID: "270ebc10-986f-4473-8a5e-9094de34ae98"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.407091 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-scripts" (OuterVolumeSpecName: "scripts") pod "270ebc10-986f-4473-8a5e-9094de34ae98" (UID: "270ebc10-986f-4473-8a5e-9094de34ae98"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.409397 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-kube-api-access-rbbzk" (OuterVolumeSpecName: "kube-api-access-rbbzk") pod "6c867fc1-05ed-46c3-99dc-71ef8a09dad3" (UID: "6c867fc1-05ed-46c3-99dc-71ef8a09dad3"). InnerVolumeSpecName "kube-api-access-rbbzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.412417 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/270ebc10-986f-4473-8a5e-9094de34ae98-kube-api-access-jz5nk" (OuterVolumeSpecName: "kube-api-access-jz5nk") pod "270ebc10-986f-4473-8a5e-9094de34ae98" (UID: "270ebc10-986f-4473-8a5e-9094de34ae98"). InnerVolumeSpecName "kube-api-access-jz5nk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.413345 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "6c867fc1-05ed-46c3-99dc-71ef8a09dad3" (UID: "6c867fc1-05ed-46c3-99dc-71ef8a09dad3"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.488578 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "270ebc10-986f-4473-8a5e-9094de34ae98" (UID: "270ebc10-986f-4473-8a5e-9094de34ae98"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.497634 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-config-data" (OuterVolumeSpecName: "config-data") pod "270ebc10-986f-4473-8a5e-9094de34ae98" (UID: "270ebc10-986f-4473-8a5e-9094de34ae98"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.499859 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbbzk\" (UniqueName: \"kubernetes.io/projected/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-kube-api-access-rbbzk\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.499896 4632 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.499906 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.499914 4632 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.499922 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.499931 4632 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/270ebc10-986f-4473-8a5e-9094de34ae98-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.499942 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jz5nk\" (UniqueName: \"kubernetes.io/projected/270ebc10-986f-4473-8a5e-9094de34ae98-kube-api-access-jz5nk\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.499961 4632 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/270ebc10-986f-4473-8a5e-9094de34ae98-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.524182 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c867fc1-05ed-46c3-99dc-71ef8a09dad3" (UID: "6c867fc1-05ed-46c3-99dc-71ef8a09dad3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.543881 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-config" (OuterVolumeSpecName: "config") pod "6c867fc1-05ed-46c3-99dc-71ef8a09dad3" (UID: "6c867fc1-05ed-46c3-99dc-71ef8a09dad3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.553097 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "270ebc10-986f-4473-8a5e-9094de34ae98" (UID: "270ebc10-986f-4473-8a5e-9094de34ae98"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.565863 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "6c867fc1-05ed-46c3-99dc-71ef8a09dad3" (UID: "6c867fc1-05ed-46c3-99dc-71ef8a09dad3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.572005 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6c867fc1-05ed-46c3-99dc-71ef8a09dad3" (UID: "6c867fc1-05ed-46c3-99dc-71ef8a09dad3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.601673 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/270ebc10-986f-4473-8a5e-9094de34ae98-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.601716 4632 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.601732 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.601743 4632 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.601757 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.612746 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "6c867fc1-05ed-46c3-99dc-71ef8a09dad3" (UID: "6c867fc1-05ed-46c3-99dc-71ef8a09dad3"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.682481 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"270ebc10-986f-4473-8a5e-9094de34ae98","Type":"ContainerDied","Data":"c4fcb786f7a33daa32bea87a76b7b56e9f86402051990ca301fe80823cca805f"} Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.682542 4632 scope.go:117] "RemoveContainer" containerID="d138d976167695fe9d299247eefcff55845f7ad27e84fc81cc086274294f2e51" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.682568 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.684539 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"aef9680f-df77-4e2e-ac53-9d7530c2270c","Type":"ContainerStarted","Data":"d29ea12656ced9b9702cd194ca8790d076046a2079e8c6f2febbe1128ab32a1a"} Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.688717 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.689496 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-64bdffbb5c-mpfvf" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.692351 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-64bdffbb5c-mpfvf" event={"ID":"6c867fc1-05ed-46c3-99dc-71ef8a09dad3","Type":"ContainerDied","Data":"bb71081b64258f79a4055c8e129128f47654fe94235aa2a730194da521f70fe1"} Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.702293 4632 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="8adc4254-ad10-4335-a365-876324d1af24" podUID="aef9680f-df77-4e2e-ac53-9d7530c2270c" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.703239 4632 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6c867fc1-05ed-46c3-99dc-71ef8a09dad3-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.722857 4632 scope.go:117] "RemoveContainer" containerID="b63cc4f80efbb7b17b044808a5b6c8d5aa98b9e2ae8e38ab95a55c4e3ba911d1" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.816239 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.830710 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.840126 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:27:02 crc kubenswrapper[4632]: E0313 10:27:02.840623 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" containerName="neutron-httpd" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.840646 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" containerName="neutron-httpd" Mar 13 10:27:02 crc kubenswrapper[4632]: E0313 10:27:02.840665 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" containerName="neutron-api" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.840674 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" containerName="neutron-api" Mar 13 10:27:02 crc kubenswrapper[4632]: E0313 10:27:02.840718 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ebc10-986f-4473-8a5e-9094de34ae98" containerName="ceilometer-central-agent" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.840726 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ebc10-986f-4473-8a5e-9094de34ae98" containerName="ceilometer-central-agent" Mar 13 10:27:02 crc kubenswrapper[4632]: E0313 10:27:02.840738 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ebc10-986f-4473-8a5e-9094de34ae98" containerName="sg-core" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.840746 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ebc10-986f-4473-8a5e-9094de34ae98" containerName="sg-core" Mar 13 10:27:02 crc kubenswrapper[4632]: E0313 10:27:02.840768 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ebc10-986f-4473-8a5e-9094de34ae98" containerName="ceilometer-notification-agent" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.840776 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ebc10-986f-4473-8a5e-9094de34ae98" containerName="ceilometer-notification-agent" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.841018 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ebc10-986f-4473-8a5e-9094de34ae98" containerName="ceilometer-notification-agent" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.841042 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ebc10-986f-4473-8a5e-9094de34ae98" containerName="ceilometer-central-agent" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.841057 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" containerName="neutron-api" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.841067 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ebc10-986f-4473-8a5e-9094de34ae98" containerName="sg-core" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.841083 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" containerName="neutron-httpd" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.850316 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-64bdffbb5c-mpfvf"] Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.850494 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.853257 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.853528 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.857794 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-64bdffbb5c-mpfvf"] Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.862079 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.881683 4632 scope.go:117] "RemoveContainer" containerID="27c121915dbbdfc336d1bc55bed50eb5edaf76e1bc92f4f6b5e249f4ffe5098a" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.907284 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-run-httpd\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.907352 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-log-httpd\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.907407 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.907430 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.907472 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xcjwv\" (UniqueName: \"kubernetes.io/projected/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-kube-api-access-xcjwv\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.908329 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-config-data\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.908368 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-scripts\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.935210 4632 scope.go:117] "RemoveContainer" containerID="bbc256375bc79a61ff656574ec8a596aed3314e7ad4cd2f7fcf6a7462aee3274" Mar 13 10:27:02 crc kubenswrapper[4632]: I0313 10:27:02.977896 4632 scope.go:117] "RemoveContainer" containerID="027b2c4436a3d137f7ef6a7921904bf128e17aa7812143af60d4d11a546759da" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.009811 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-config-data\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.009903 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-scripts\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.010002 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-run-httpd\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.010028 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-log-httpd\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.010072 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.010088 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.010119 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xcjwv\" (UniqueName: \"kubernetes.io/projected/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-kube-api-access-xcjwv\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.011429 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-log-httpd\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.015431 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-run-httpd\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.021407 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.021482 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-scripts\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.023620 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-config-data\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.049913 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.054423 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xcjwv\" (UniqueName: \"kubernetes.io/projected/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-kube-api-access-xcjwv\") pod \"ceilometer-0\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.178217 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.334731 4632 scope.go:117] "RemoveContainer" containerID="746bd1f1584c6b468985171d618d35f15871608c045fd5e9f4070c7ace66e505" Mar 13 10:27:03 crc kubenswrapper[4632]: I0313 10:27:03.825200 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:27:04 crc kubenswrapper[4632]: I0313 10:27:04.059153 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="270ebc10-986f-4473-8a5e-9094de34ae98" path="/var/lib/kubelet/pods/270ebc10-986f-4473-8a5e-9094de34ae98/volumes" Mar 13 10:27:04 crc kubenswrapper[4632]: I0313 10:27:04.060143 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c867fc1-05ed-46c3-99dc-71ef8a09dad3" path="/var/lib/kubelet/pods/6c867fc1-05ed-46c3-99dc-71ef8a09dad3/volumes" Mar 13 10:27:04 crc kubenswrapper[4632]: I0313 10:27:04.719487 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c","Type":"ContainerStarted","Data":"bb00bf460a4849cc1a7c1bad8a739981e87c032a18a9222632d57abbccea8858"} Mar 13 10:27:04 crc kubenswrapper[4632]: I0313 10:27:04.720113 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c","Type":"ContainerStarted","Data":"df15286148314ab907f4a05031eefbb838636621ee25cc2f368e3d56ae19621b"} Mar 13 10:27:05 crc kubenswrapper[4632]: I0313 10:27:05.736622 4632 generic.go:334] "Generic (PLEG): container finished" podID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerID="dc4a058f6feb7822333693352f32f5677ff03988b7b5b71005c85c4bf733b402" exitCode=137 Mar 13 10:27:05 crc kubenswrapper[4632]: I0313 10:27:05.737003 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb5f7878-ng2k2" event={"ID":"3e4afb91-ce26-4325-89c9-2542da2ec48a","Type":"ContainerDied","Data":"dc4a058f6feb7822333693352f32f5677ff03988b7b5b71005c85c4bf733b402"} Mar 13 10:27:05 crc kubenswrapper[4632]: I0313 10:27:05.758043 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c","Type":"ContainerStarted","Data":"aa0a9edf7c00bb4d08cf1a3f2565b5016be14ce1312e093e1c44112d2d594f42"} Mar 13 10:27:05 crc kubenswrapper[4632]: I0313 10:27:05.758104 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c","Type":"ContainerStarted","Data":"c7001d72ce189e15496046472a90b656a4129de71ad96c6f49a1d6b92862a990"} Mar 13 10:27:05 crc kubenswrapper[4632]: I0313 10:27:05.813834 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 10:27:06 crc kubenswrapper[4632]: I0313 10:27:06.773423 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb5f7878-ng2k2" event={"ID":"3e4afb91-ce26-4325-89c9-2542da2ec48a","Type":"ContainerStarted","Data":"c9dfdd84c36e6ac95b45a488b62e176636bdecfbe3a88d3f5d2058d92ebbacdd"} Mar 13 10:27:06 crc kubenswrapper[4632]: I0313 10:27:06.804393 4632 generic.go:334] "Generic (PLEG): container finished" podID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerID="8ce0185281fb59d0c6bda2b2c484ad3711b4bd3b729b4b8677e75ca6b8e1f739" exitCode=137 Mar 13 10:27:06 crc kubenswrapper[4632]: I0313 10:27:06.808692 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689764498d-rg7vt" event={"ID":"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c","Type":"ContainerDied","Data":"8ce0185281fb59d0c6bda2b2c484ad3711b4bd3b729b4b8677e75ca6b8e1f739"} Mar 13 10:27:06 crc kubenswrapper[4632]: I0313 10:27:06.809572 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689764498d-rg7vt" event={"ID":"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c","Type":"ContainerStarted","Data":"433c9aa5a02161c4bc7228b52cc460020479cbbb899bc6549755a59b8ad796f4"} Mar 13 10:27:09 crc kubenswrapper[4632]: I0313 10:27:09.847825 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c","Type":"ContainerStarted","Data":"d67a8223c96cadeaa871fcaaaad472258eb768daca2821f6757940c48f3eafd6"} Mar 13 10:27:09 crc kubenswrapper[4632]: I0313 10:27:09.875322 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.701995889 podStartE2EDuration="7.875300214s" podCreationTimestamp="2026-03-13 10:27:02 +0000 UTC" firstStartedPulling="2026-03-13 10:27:03.831548964 +0000 UTC m=+1397.854079097" lastFinishedPulling="2026-03-13 10:27:09.004853289 +0000 UTC m=+1403.027383422" observedRunningTime="2026-03-13 10:27:09.868418862 +0000 UTC m=+1403.890948995" watchObservedRunningTime="2026-03-13 10:27:09.875300214 +0000 UTC m=+1403.897830347" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.244016 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-7dbf8b9ddc-6p5vh"] Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.245639 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.266743 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.267601 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.269400 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.309144 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7dbf8b9ddc-6p5vh"] Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.384552 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2gqw\" (UniqueName: \"kubernetes.io/projected/03ca050c-63a7-4b37-91fe-fe5c322cca78-kube-api-access-p2gqw\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.384882 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ca050c-63a7-4b37-91fe-fe5c322cca78-combined-ca-bundle\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.385124 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ca050c-63a7-4b37-91fe-fe5c322cca78-internal-tls-certs\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.385352 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/03ca050c-63a7-4b37-91fe-fe5c322cca78-etc-swift\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.385488 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ca050c-63a7-4b37-91fe-fe5c322cca78-public-tls-certs\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.385639 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ca050c-63a7-4b37-91fe-fe5c322cca78-config-data\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.385796 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03ca050c-63a7-4b37-91fe-fe5c322cca78-run-httpd\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.385966 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03ca050c-63a7-4b37-91fe-fe5c322cca78-log-httpd\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.461762 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.461822 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.487779 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ca050c-63a7-4b37-91fe-fe5c322cca78-public-tls-certs\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.487839 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ca050c-63a7-4b37-91fe-fe5c322cca78-config-data\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.487875 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03ca050c-63a7-4b37-91fe-fe5c322cca78-run-httpd\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.487929 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03ca050c-63a7-4b37-91fe-fe5c322cca78-log-httpd\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.488191 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2gqw\" (UniqueName: \"kubernetes.io/projected/03ca050c-63a7-4b37-91fe-fe5c322cca78-kube-api-access-p2gqw\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.488227 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ca050c-63a7-4b37-91fe-fe5c322cca78-combined-ca-bundle\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.488253 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ca050c-63a7-4b37-91fe-fe5c322cca78-internal-tls-certs\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.488303 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/03ca050c-63a7-4b37-91fe-fe5c322cca78-etc-swift\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.489438 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03ca050c-63a7-4b37-91fe-fe5c322cca78-run-httpd\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.489499 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/03ca050c-63a7-4b37-91fe-fe5c322cca78-log-httpd\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.496192 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ca050c-63a7-4b37-91fe-fe5c322cca78-internal-tls-certs\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.497349 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/03ca050c-63a7-4b37-91fe-fe5c322cca78-etc-swift\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.497546 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/03ca050c-63a7-4b37-91fe-fe5c322cca78-public-tls-certs\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.518188 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2gqw\" (UniqueName: \"kubernetes.io/projected/03ca050c-63a7-4b37-91fe-fe5c322cca78-kube-api-access-p2gqw\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.519470 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/03ca050c-63a7-4b37-91fe-fe5c322cca78-config-data\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.523779 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03ca050c-63a7-4b37-91fe-fe5c322cca78-combined-ca-bundle\") pod \"swift-proxy-7dbf8b9ddc-6p5vh\" (UID: \"03ca050c-63a7-4b37-91fe-fe5c322cca78\") " pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.612042 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.805426 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.864249 4632 generic.go:334] "Generic (PLEG): container finished" podID="58677e2e-9fc6-4e50-b342-e912afa8d969" containerID="3a8d9431bb58dc2e36bce7009280ffed0639f98e73ca93dba3c41c03d94fb14f" exitCode=137 Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.864305 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58677e2e-9fc6-4e50-b342-e912afa8d969","Type":"ContainerDied","Data":"3a8d9431bb58dc2e36bce7009280ffed0639f98e73ca93dba3c41c03d94fb14f"} Mar 13 10:27:10 crc kubenswrapper[4632]: I0313 10:27:10.864534 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 10:27:11 crc kubenswrapper[4632]: I0313 10:27:11.875449 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="ceilometer-central-agent" containerID="cri-o://bb00bf460a4849cc1a7c1bad8a739981e87c032a18a9222632d57abbccea8858" gracePeriod=30 Mar 13 10:27:11 crc kubenswrapper[4632]: I0313 10:27:11.875508 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="sg-core" containerID="cri-o://aa0a9edf7c00bb4d08cf1a3f2565b5016be14ce1312e093e1c44112d2d594f42" gracePeriod=30 Mar 13 10:27:11 crc kubenswrapper[4632]: I0313 10:27:11.875543 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="ceilometer-notification-agent" containerID="cri-o://c7001d72ce189e15496046472a90b656a4129de71ad96c6f49a1d6b92862a990" gracePeriod=30 Mar 13 10:27:11 crc kubenswrapper[4632]: I0313 10:27:11.875530 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="proxy-httpd" containerID="cri-o://d67a8223c96cadeaa871fcaaaad472258eb768daca2821f6757940c48f3eafd6" gracePeriod=30 Mar 13 10:27:12 crc kubenswrapper[4632]: I0313 10:27:12.899186 4632 generic.go:334] "Generic (PLEG): container finished" podID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerID="d67a8223c96cadeaa871fcaaaad472258eb768daca2821f6757940c48f3eafd6" exitCode=0 Mar 13 10:27:12 crc kubenswrapper[4632]: I0313 10:27:12.899543 4632 generic.go:334] "Generic (PLEG): container finished" podID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerID="aa0a9edf7c00bb4d08cf1a3f2565b5016be14ce1312e093e1c44112d2d594f42" exitCode=2 Mar 13 10:27:12 crc kubenswrapper[4632]: I0313 10:27:12.899562 4632 generic.go:334] "Generic (PLEG): container finished" podID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerID="c7001d72ce189e15496046472a90b656a4129de71ad96c6f49a1d6b92862a990" exitCode=0 Mar 13 10:27:12 crc kubenswrapper[4632]: I0313 10:27:12.899572 4632 generic.go:334] "Generic (PLEG): container finished" podID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerID="bb00bf460a4849cc1a7c1bad8a739981e87c032a18a9222632d57abbccea8858" exitCode=0 Mar 13 10:27:12 crc kubenswrapper[4632]: I0313 10:27:12.899363 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c","Type":"ContainerDied","Data":"d67a8223c96cadeaa871fcaaaad472258eb768daca2821f6757940c48f3eafd6"} Mar 13 10:27:12 crc kubenswrapper[4632]: I0313 10:27:12.899616 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c","Type":"ContainerDied","Data":"aa0a9edf7c00bb4d08cf1a3f2565b5016be14ce1312e093e1c44112d2d594f42"} Mar 13 10:27:12 crc kubenswrapper[4632]: I0313 10:27:12.899636 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c","Type":"ContainerDied","Data":"c7001d72ce189e15496046472a90b656a4129de71ad96c6f49a1d6b92862a990"} Mar 13 10:27:12 crc kubenswrapper[4632]: I0313 10:27:12.899651 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c","Type":"ContainerDied","Data":"bb00bf460a4849cc1a7c1bad8a739981e87c032a18a9222632d57abbccea8858"} Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.208431 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7f9df5b5b5-q6dp2"] Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.213146 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.221826 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.229462 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-vbbdq" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.229923 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.243984 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7f9df5b5b5-q6dp2"] Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.353555 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-config-data\") pod \"heat-engine-7f9df5b5b5-q6dp2\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.353606 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-config-data-custom\") pod \"heat-engine-7f9df5b5b5-q6dp2\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.353655 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-combined-ca-bundle\") pod \"heat-engine-7f9df5b5b5-q6dp2\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.353842 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5895\" (UniqueName: \"kubernetes.io/projected/757b852e-068c-4885-99b8-af2e6f23e445-kube-api-access-d5895\") pod \"heat-engine-7f9df5b5b5-q6dp2\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.439791 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7888df55c7-mw5p4"] Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.441812 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.455263 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5895\" (UniqueName: \"kubernetes.io/projected/757b852e-068c-4885-99b8-af2e6f23e445-kube-api-access-d5895\") pod \"heat-engine-7f9df5b5b5-q6dp2\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.455350 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-config-data\") pod \"heat-engine-7f9df5b5b5-q6dp2\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.455377 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-config-data-custom\") pod \"heat-engine-7f9df5b5b5-q6dp2\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.455430 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-combined-ca-bundle\") pod \"heat-engine-7f9df5b5b5-q6dp2\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.465335 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-config-data\") pod \"heat-engine-7f9df5b5b5-q6dp2\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.466570 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-combined-ca-bundle\") pod \"heat-engine-7f9df5b5b5-q6dp2\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.548834 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7888df55c7-mw5p4"] Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.549806 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-config-data-custom\") pod \"heat-engine-7f9df5b5b5-q6dp2\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.567014 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-ovsdbserver-nb\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.567168 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-dns-svc\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.567264 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-dns-swift-storage-0\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.567304 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-ovsdbserver-sb\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.567380 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-config\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.567446 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6mzz\" (UniqueName: \"kubernetes.io/projected/904f04cd-8110-4637-8bb4-67c4b83e189b-kube-api-access-k6mzz\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.651316 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5895\" (UniqueName: \"kubernetes.io/projected/757b852e-068c-4885-99b8-af2e6f23e445-kube-api-access-d5895\") pod \"heat-engine-7f9df5b5b5-q6dp2\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.674689 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6mzz\" (UniqueName: \"kubernetes.io/projected/904f04cd-8110-4637-8bb4-67c4b83e189b-kube-api-access-k6mzz\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.674751 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-ovsdbserver-nb\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.674868 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-dns-svc\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.674962 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-dns-swift-storage-0\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.674989 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-ovsdbserver-sb\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.675041 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-config\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.683905 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-config\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.685092 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-dns-swift-storage-0\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.687727 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-ovsdbserver-sb\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.690830 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-dns-svc\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.703874 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-d856c56c-cmd2q"] Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.711612 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-ovsdbserver-nb\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.730687 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.743806 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.744734 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6mzz\" (UniqueName: \"kubernetes.io/projected/904f04cd-8110-4637-8bb4-67c4b83e189b-kube-api-access-k6mzz\") pod \"dnsmasq-dns-7888df55c7-mw5p4\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.825078 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-d856c56c-cmd2q"] Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.835135 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.858922 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-b547848c4-bn5vs"] Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.867702 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.875314 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.881516 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prc4f\" (UniqueName: \"kubernetes.io/projected/5d10747e-ba77-4986-9d4b-636fcbf823ab-kube-api-access-prc4f\") pod \"heat-cfnapi-d856c56c-cmd2q\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.881573 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-config-data-custom\") pod \"heat-cfnapi-d856c56c-cmd2q\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.881655 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-config-data\") pod \"heat-cfnapi-d856c56c-cmd2q\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.881748 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-combined-ca-bundle\") pod \"heat-cfnapi-d856c56c-cmd2q\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.897226 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-b547848c4-bn5vs"] Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.967458 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.983603 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prc4f\" (UniqueName: \"kubernetes.io/projected/5d10747e-ba77-4986-9d4b-636fcbf823ab-kube-api-access-prc4f\") pod \"heat-cfnapi-d856c56c-cmd2q\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.986694 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-config-data-custom\") pod \"heat-cfnapi-d856c56c-cmd2q\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.986840 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nn98r\" (UniqueName: \"kubernetes.io/projected/07914020-653d-4509-9f60-22726224c7c6-kube-api-access-nn98r\") pod \"heat-api-b547848c4-bn5vs\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.986883 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-config-data\") pod \"heat-cfnapi-d856c56c-cmd2q\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.986968 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-config-data-custom\") pod \"heat-api-b547848c4-bn5vs\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.987037 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-config-data\") pod \"heat-api-b547848c4-bn5vs\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.987137 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-combined-ca-bundle\") pod \"heat-cfnapi-d856c56c-cmd2q\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.987210 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-combined-ca-bundle\") pod \"heat-api-b547848c4-bn5vs\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.997881 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-config-data-custom\") pod \"heat-cfnapi-d856c56c-cmd2q\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.998401 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-config-data\") pod \"heat-cfnapi-d856c56c-cmd2q\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:13 crc kubenswrapper[4632]: I0313 10:27:13.998856 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-combined-ca-bundle\") pod \"heat-cfnapi-d856c56c-cmd2q\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:14 crc kubenswrapper[4632]: I0313 10:27:14.004576 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prc4f\" (UniqueName: \"kubernetes.io/projected/5d10747e-ba77-4986-9d4b-636fcbf823ab-kube-api-access-prc4f\") pod \"heat-cfnapi-d856c56c-cmd2q\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:14 crc kubenswrapper[4632]: I0313 10:27:14.091294 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nn98r\" (UniqueName: \"kubernetes.io/projected/07914020-653d-4509-9f60-22726224c7c6-kube-api-access-nn98r\") pod \"heat-api-b547848c4-bn5vs\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:14 crc kubenswrapper[4632]: I0313 10:27:14.091367 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-config-data-custom\") pod \"heat-api-b547848c4-bn5vs\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:14 crc kubenswrapper[4632]: I0313 10:27:14.091407 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-config-data\") pod \"heat-api-b547848c4-bn5vs\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:14 crc kubenswrapper[4632]: I0313 10:27:14.091477 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-combined-ca-bundle\") pod \"heat-api-b547848c4-bn5vs\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:14 crc kubenswrapper[4632]: I0313 10:27:14.103533 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-combined-ca-bundle\") pod \"heat-api-b547848c4-bn5vs\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:14 crc kubenswrapper[4632]: I0313 10:27:14.103739 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-config-data-custom\") pod \"heat-api-b547848c4-bn5vs\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:14 crc kubenswrapper[4632]: I0313 10:27:14.104481 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-config-data\") pod \"heat-api-b547848c4-bn5vs\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:14 crc kubenswrapper[4632]: I0313 10:27:14.115484 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nn98r\" (UniqueName: \"kubernetes.io/projected/07914020-653d-4509-9f60-22726224c7c6-kube-api-access-nn98r\") pod \"heat-api-b547848c4-bn5vs\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:14 crc kubenswrapper[4632]: I0313 10:27:14.143112 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:14 crc kubenswrapper[4632]: I0313 10:27:14.226033 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:14 crc kubenswrapper[4632]: I0313 10:27:14.633228 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="58677e2e-9fc6-4e50-b342-e912afa8d969" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.172:8776/healthcheck\": dial tcp 10.217.0.172:8776: connect: connection refused" Mar 13 10:27:15 crc kubenswrapper[4632]: I0313 10:27:15.394762 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:27:15 crc kubenswrapper[4632]: I0313 10:27:15.395671 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:27:15 crc kubenswrapper[4632]: I0313 10:27:15.395964 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 10:27:15 crc kubenswrapper[4632]: I0313 10:27:15.857265 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:27:15 crc kubenswrapper[4632]: I0313 10:27:15.857574 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.716393 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-wgv42"] Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.717823 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wgv42" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.746963 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-wgv42"] Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.760913 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsxhh\" (UniqueName: \"kubernetes.io/projected/f0c32ed5-c3b0-45ea-99de-87c45cb1ba77-kube-api-access-nsxhh\") pod \"nova-api-db-create-wgv42\" (UID: \"f0c32ed5-c3b0-45ea-99de-87c45cb1ba77\") " pod="openstack/nova-api-db-create-wgv42" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.761098 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0c32ed5-c3b0-45ea-99de-87c45cb1ba77-operator-scripts\") pod \"nova-api-db-create-wgv42\" (UID: \"f0c32ed5-c3b0-45ea-99de-87c45cb1ba77\") " pod="openstack/nova-api-db-create-wgv42" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.817524 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-kswhw"] Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.821349 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kswhw" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.832201 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-kswhw"] Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.863544 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdg5s\" (UniqueName: \"kubernetes.io/projected/8e0fb1fc-c94a-44f0-a269-e7211c6fcfba-kube-api-access-jdg5s\") pod \"nova-cell0-db-create-kswhw\" (UID: \"8e0fb1fc-c94a-44f0-a269-e7211c6fcfba\") " pod="openstack/nova-cell0-db-create-kswhw" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.863623 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0c32ed5-c3b0-45ea-99de-87c45cb1ba77-operator-scripts\") pod \"nova-api-db-create-wgv42\" (UID: \"f0c32ed5-c3b0-45ea-99de-87c45cb1ba77\") " pod="openstack/nova-api-db-create-wgv42" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.863662 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e0fb1fc-c94a-44f0-a269-e7211c6fcfba-operator-scripts\") pod \"nova-cell0-db-create-kswhw\" (UID: \"8e0fb1fc-c94a-44f0-a269-e7211c6fcfba\") " pod="openstack/nova-cell0-db-create-kswhw" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.872120 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsxhh\" (UniqueName: \"kubernetes.io/projected/f0c32ed5-c3b0-45ea-99de-87c45cb1ba77-kube-api-access-nsxhh\") pod \"nova-api-db-create-wgv42\" (UID: \"f0c32ed5-c3b0-45ea-99de-87c45cb1ba77\") " pod="openstack/nova-api-db-create-wgv42" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.873463 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0c32ed5-c3b0-45ea-99de-87c45cb1ba77-operator-scripts\") pod \"nova-api-db-create-wgv42\" (UID: \"f0c32ed5-c3b0-45ea-99de-87c45cb1ba77\") " pod="openstack/nova-api-db-create-wgv42" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.905983 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsxhh\" (UniqueName: \"kubernetes.io/projected/f0c32ed5-c3b0-45ea-99de-87c45cb1ba77-kube-api-access-nsxhh\") pod \"nova-api-db-create-wgv42\" (UID: \"f0c32ed5-c3b0-45ea-99de-87c45cb1ba77\") " pod="openstack/nova-api-db-create-wgv42" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.954573 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-fshjb"] Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.956205 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fshjb" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.973680 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e0fb1fc-c94a-44f0-a269-e7211c6fcfba-operator-scripts\") pod \"nova-cell0-db-create-kswhw\" (UID: \"8e0fb1fc-c94a-44f0-a269-e7211c6fcfba\") " pod="openstack/nova-cell0-db-create-kswhw" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.974077 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdg5s\" (UniqueName: \"kubernetes.io/projected/8e0fb1fc-c94a-44f0-a269-e7211c6fcfba-kube-api-access-jdg5s\") pod \"nova-cell0-db-create-kswhw\" (UID: \"8e0fb1fc-c94a-44f0-a269-e7211c6fcfba\") " pod="openstack/nova-cell0-db-create-kswhw" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.975761 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e0fb1fc-c94a-44f0-a269-e7211c6fcfba-operator-scripts\") pod \"nova-cell0-db-create-kswhw\" (UID: \"8e0fb1fc-c94a-44f0-a269-e7211c6fcfba\") " pod="openstack/nova-cell0-db-create-kswhw" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.978158 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-f3f1-account-create-update-29g8s"] Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.979522 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f3f1-account-create-update-29g8s" Mar 13 10:27:18 crc kubenswrapper[4632]: I0313 10:27:18.981582 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.002124 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-fshjb"] Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.026752 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f3f1-account-create-update-29g8s"] Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.037759 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdg5s\" (UniqueName: \"kubernetes.io/projected/8e0fb1fc-c94a-44f0-a269-e7211c6fcfba-kube-api-access-jdg5s\") pod \"nova-cell0-db-create-kswhw\" (UID: \"8e0fb1fc-c94a-44f0-a269-e7211c6fcfba\") " pod="openstack/nova-cell0-db-create-kswhw" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.054558 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wgv42" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.093869 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09bd98be-9d10-4a53-8ef6-c4718b05c3f6-operator-scripts\") pod \"nova-cell1-db-create-fshjb\" (UID: \"09bd98be-9d10-4a53-8ef6-c4718b05c3f6\") " pod="openstack/nova-cell1-db-create-fshjb" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.093953 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qgng\" (UniqueName: \"kubernetes.io/projected/234a900d-887b-448c-8336-010107726c1e-kube-api-access-9qgng\") pod \"nova-api-f3f1-account-create-update-29g8s\" (UID: \"234a900d-887b-448c-8336-010107726c1e\") " pod="openstack/nova-api-f3f1-account-create-update-29g8s" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.094046 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cl442\" (UniqueName: \"kubernetes.io/projected/09bd98be-9d10-4a53-8ef6-c4718b05c3f6-kube-api-access-cl442\") pod \"nova-cell1-db-create-fshjb\" (UID: \"09bd98be-9d10-4a53-8ef6-c4718b05c3f6\") " pod="openstack/nova-cell1-db-create-fshjb" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.095880 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/234a900d-887b-448c-8336-010107726c1e-operator-scripts\") pod \"nova-api-f3f1-account-create-update-29g8s\" (UID: \"234a900d-887b-448c-8336-010107726c1e\") " pod="openstack/nova-api-f3f1-account-create-update-29g8s" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.134654 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-86d4-account-create-update-5c7rj"] Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.142277 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.145229 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.160956 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-86d4-account-create-update-5c7rj"] Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.168582 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kswhw" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.199435 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5fx7d\" (UniqueName: \"kubernetes.io/projected/bbaf5a79-1c34-4518-afb9-19703fe6c45b-kube-api-access-5fx7d\") pod \"nova-cell0-86d4-account-create-update-5c7rj\" (UID: \"bbaf5a79-1c34-4518-afb9-19703fe6c45b\") " pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.199552 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09bd98be-9d10-4a53-8ef6-c4718b05c3f6-operator-scripts\") pod \"nova-cell1-db-create-fshjb\" (UID: \"09bd98be-9d10-4a53-8ef6-c4718b05c3f6\") " pod="openstack/nova-cell1-db-create-fshjb" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.199617 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qgng\" (UniqueName: \"kubernetes.io/projected/234a900d-887b-448c-8336-010107726c1e-kube-api-access-9qgng\") pod \"nova-api-f3f1-account-create-update-29g8s\" (UID: \"234a900d-887b-448c-8336-010107726c1e\") " pod="openstack/nova-api-f3f1-account-create-update-29g8s" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.199701 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbaf5a79-1c34-4518-afb9-19703fe6c45b-operator-scripts\") pod \"nova-cell0-86d4-account-create-update-5c7rj\" (UID: \"bbaf5a79-1c34-4518-afb9-19703fe6c45b\") " pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.199750 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cl442\" (UniqueName: \"kubernetes.io/projected/09bd98be-9d10-4a53-8ef6-c4718b05c3f6-kube-api-access-cl442\") pod \"nova-cell1-db-create-fshjb\" (UID: \"09bd98be-9d10-4a53-8ef6-c4718b05c3f6\") " pod="openstack/nova-cell1-db-create-fshjb" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.199895 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/234a900d-887b-448c-8336-010107726c1e-operator-scripts\") pod \"nova-api-f3f1-account-create-update-29g8s\" (UID: \"234a900d-887b-448c-8336-010107726c1e\") " pod="openstack/nova-api-f3f1-account-create-update-29g8s" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.200476 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09bd98be-9d10-4a53-8ef6-c4718b05c3f6-operator-scripts\") pod \"nova-cell1-db-create-fshjb\" (UID: \"09bd98be-9d10-4a53-8ef6-c4718b05c3f6\") " pod="openstack/nova-cell1-db-create-fshjb" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.200706 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/234a900d-887b-448c-8336-010107726c1e-operator-scripts\") pod \"nova-api-f3f1-account-create-update-29g8s\" (UID: \"234a900d-887b-448c-8336-010107726c1e\") " pod="openstack/nova-api-f3f1-account-create-update-29g8s" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.230447 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qgng\" (UniqueName: \"kubernetes.io/projected/234a900d-887b-448c-8336-010107726c1e-kube-api-access-9qgng\") pod \"nova-api-f3f1-account-create-update-29g8s\" (UID: \"234a900d-887b-448c-8336-010107726c1e\") " pod="openstack/nova-api-f3f1-account-create-update-29g8s" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.245645 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cl442\" (UniqueName: \"kubernetes.io/projected/09bd98be-9d10-4a53-8ef6-c4718b05c3f6-kube-api-access-cl442\") pod \"nova-cell1-db-create-fshjb\" (UID: \"09bd98be-9d10-4a53-8ef6-c4718b05c3f6\") " pod="openstack/nova-cell1-db-create-fshjb" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.302202 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5fx7d\" (UniqueName: \"kubernetes.io/projected/bbaf5a79-1c34-4518-afb9-19703fe6c45b-kube-api-access-5fx7d\") pod \"nova-cell0-86d4-account-create-update-5c7rj\" (UID: \"bbaf5a79-1c34-4518-afb9-19703fe6c45b\") " pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.302346 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbaf5a79-1c34-4518-afb9-19703fe6c45b-operator-scripts\") pod \"nova-cell0-86d4-account-create-update-5c7rj\" (UID: \"bbaf5a79-1c34-4518-afb9-19703fe6c45b\") " pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.303282 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbaf5a79-1c34-4518-afb9-19703fe6c45b-operator-scripts\") pod \"nova-cell0-86d4-account-create-update-5c7rj\" (UID: \"bbaf5a79-1c34-4518-afb9-19703fe6c45b\") " pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.315585 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fshjb" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.327315 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f3f1-account-create-update-29g8s" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.341523 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5fx7d\" (UniqueName: \"kubernetes.io/projected/bbaf5a79-1c34-4518-afb9-19703fe6c45b-kube-api-access-5fx7d\") pod \"nova-cell0-86d4-account-create-update-5c7rj\" (UID: \"bbaf5a79-1c34-4518-afb9-19703fe6c45b\") " pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.355373 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-2f8c-account-create-update-g4b8g"] Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.356932 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.360637 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.391075 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2f8c-account-create-update-g4b8g"] Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.409515 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrgfq\" (UniqueName: \"kubernetes.io/projected/8462be25-a577-476d-b54a-73790a8aa189-kube-api-access-rrgfq\") pod \"nova-cell1-2f8c-account-create-update-g4b8g\" (UID: \"8462be25-a577-476d-b54a-73790a8aa189\") " pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.409607 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8462be25-a577-476d-b54a-73790a8aa189-operator-scripts\") pod \"nova-cell1-2f8c-account-create-update-g4b8g\" (UID: \"8462be25-a577-476d-b54a-73790a8aa189\") " pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.463445 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.512202 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrgfq\" (UniqueName: \"kubernetes.io/projected/8462be25-a577-476d-b54a-73790a8aa189-kube-api-access-rrgfq\") pod \"nova-cell1-2f8c-account-create-update-g4b8g\" (UID: \"8462be25-a577-476d-b54a-73790a8aa189\") " pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.512363 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8462be25-a577-476d-b54a-73790a8aa189-operator-scripts\") pod \"nova-cell1-2f8c-account-create-update-g4b8g\" (UID: \"8462be25-a577-476d-b54a-73790a8aa189\") " pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.513598 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8462be25-a577-476d-b54a-73790a8aa189-operator-scripts\") pod \"nova-cell1-2f8c-account-create-update-g4b8g\" (UID: \"8462be25-a577-476d-b54a-73790a8aa189\") " pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.530924 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrgfq\" (UniqueName: \"kubernetes.io/projected/8462be25-a577-476d-b54a-73790a8aa189-kube-api-access-rrgfq\") pod \"nova-cell1-2f8c-account-create-update-g4b8g\" (UID: \"8462be25-a577-476d-b54a-73790a8aa189\") " pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" Mar 13 10:27:19 crc kubenswrapper[4632]: E0313 10:27:19.555902 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-openstackclient:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:27:19 crc kubenswrapper[4632]: E0313 10:27:19.556208 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-openstackclient:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:27:19 crc kubenswrapper[4632]: E0313 10:27:19.556445 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-openstackclient:e43235cb19da04699a53f42b6a75afe9,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n55dh85h67dh9h89h686h667h58ch57hc4h5c8h556hf9h567h5f5h66dh65dhc6hfh56bh655h67fh88h555h65h5dfh5b8h5d6h65fh684hdh6q,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cmjms,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(aef9680f-df77-4e2e-ac53-9d7530c2270c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:27:19 crc kubenswrapper[4632]: E0313 10:27:19.559317 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="aef9680f-df77-4e2e-ac53-9d7530c2270c" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.632971 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="58677e2e-9fc6-4e50-b342-e912afa8d969" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.172:8776/healthcheck\": dial tcp 10.217.0.172:8776: connect: connection refused" Mar 13 10:27:19 crc kubenswrapper[4632]: I0313 10:27:19.737373 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" Mar 13 10:27:20 crc kubenswrapper[4632]: E0313 10:27:20.096552 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-openstackclient:e43235cb19da04699a53f42b6a75afe9\\\"\"" pod="openstack/openstackclient" podUID="aef9680f-df77-4e2e-ac53-9d7530c2270c" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.372100 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.466769 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-config-data\") pod \"58677e2e-9fc6-4e50-b342-e912afa8d969\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.466852 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-combined-ca-bundle\") pod \"58677e2e-9fc6-4e50-b342-e912afa8d969\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.466906 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-config-data-custom\") pod \"58677e2e-9fc6-4e50-b342-e912afa8d969\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.467010 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd7tw\" (UniqueName: \"kubernetes.io/projected/58677e2e-9fc6-4e50-b342-e912afa8d969-kube-api-access-vd7tw\") pod \"58677e2e-9fc6-4e50-b342-e912afa8d969\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.467041 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58677e2e-9fc6-4e50-b342-e912afa8d969-etc-machine-id\") pod \"58677e2e-9fc6-4e50-b342-e912afa8d969\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.467069 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58677e2e-9fc6-4e50-b342-e912afa8d969-logs\") pod \"58677e2e-9fc6-4e50-b342-e912afa8d969\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.467152 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-scripts\") pod \"58677e2e-9fc6-4e50-b342-e912afa8d969\" (UID: \"58677e2e-9fc6-4e50-b342-e912afa8d969\") " Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.470308 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/58677e2e-9fc6-4e50-b342-e912afa8d969-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "58677e2e-9fc6-4e50-b342-e912afa8d969" (UID: "58677e2e-9fc6-4e50-b342-e912afa8d969"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.473608 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58677e2e-9fc6-4e50-b342-e912afa8d969-logs" (OuterVolumeSpecName: "logs") pod "58677e2e-9fc6-4e50-b342-e912afa8d969" (UID: "58677e2e-9fc6-4e50-b342-e912afa8d969"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.492924 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-scripts" (OuterVolumeSpecName: "scripts") pod "58677e2e-9fc6-4e50-b342-e912afa8d969" (UID: "58677e2e-9fc6-4e50-b342-e912afa8d969"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.493090 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "58677e2e-9fc6-4e50-b342-e912afa8d969" (UID: "58677e2e-9fc6-4e50-b342-e912afa8d969"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.493228 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58677e2e-9fc6-4e50-b342-e912afa8d969-kube-api-access-vd7tw" (OuterVolumeSpecName: "kube-api-access-vd7tw") pod "58677e2e-9fc6-4e50-b342-e912afa8d969" (UID: "58677e2e-9fc6-4e50-b342-e912afa8d969"). InnerVolumeSpecName "kube-api-access-vd7tw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.585107 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.585151 4632 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.585166 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd7tw\" (UniqueName: \"kubernetes.io/projected/58677e2e-9fc6-4e50-b342-e912afa8d969-kube-api-access-vd7tw\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.585180 4632 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58677e2e-9fc6-4e50-b342-e912afa8d969-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.585192 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58677e2e-9fc6-4e50-b342-e912afa8d969-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.621868 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58677e2e-9fc6-4e50-b342-e912afa8d969" (UID: "58677e2e-9fc6-4e50-b342-e912afa8d969"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.690209 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.814523 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-config-data" (OuterVolumeSpecName: "config-data") pod "58677e2e-9fc6-4e50-b342-e912afa8d969" (UID: "58677e2e-9fc6-4e50-b342-e912afa8d969"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:20 crc kubenswrapper[4632]: I0313 10:27:20.902783 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58677e2e-9fc6-4e50-b342-e912afa8d969-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.058251 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.190167 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58677e2e-9fc6-4e50-b342-e912afa8d969","Type":"ContainerDied","Data":"8037b401a0baaaa45f09498066b3b722d38c4aef73b4ab3874c935fbc21eac6e"} Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.190231 4632 scope.go:117] "RemoveContainer" containerID="3a8d9431bb58dc2e36bce7009280ffed0639f98e73ca93dba3c41c03d94fb14f" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.190394 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.214564 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-run-httpd\") pod \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.214639 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-log-httpd\") pod \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.214674 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-scripts\") pod \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.214696 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-config-data\") pod \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.214876 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcjwv\" (UniqueName: \"kubernetes.io/projected/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-kube-api-access-xcjwv\") pod \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.215027 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-combined-ca-bundle\") pod \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.215060 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-sg-core-conf-yaml\") pod \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\" (UID: \"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c\") " Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.217758 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" (UID: "c4e63898-c65b-42c6-9ec5-3089ae7a8d8c"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.225900 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7888df55c7-mw5p4"] Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.230867 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" (UID: "c4e63898-c65b-42c6-9ec5-3089ae7a8d8c"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.252428 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-scripts" (OuterVolumeSpecName: "scripts") pod "c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" (UID: "c4e63898-c65b-42c6-9ec5-3089ae7a8d8c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.315782 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c4e63898-c65b-42c6-9ec5-3089ae7a8d8c","Type":"ContainerDied","Data":"df15286148314ab907f4a05031eefbb838636621ee25cc2f368e3d56ae19621b"} Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.316074 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.331003 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-kube-api-access-xcjwv" (OuterVolumeSpecName: "kube-api-access-xcjwv") pod "c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" (UID: "c4e63898-c65b-42c6-9ec5-3089ae7a8d8c"). InnerVolumeSpecName "kube-api-access-xcjwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.332635 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcjwv\" (UniqueName: \"kubernetes.io/projected/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-kube-api-access-xcjwv\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.368932 4632 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.368971 4632 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.368984 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.423384 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" (UID: "c4e63898-c65b-42c6-9ec5-3089ae7a8d8c"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.470327 4632 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.552032 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.631365 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.668092 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 13 10:27:21 crc kubenswrapper[4632]: E0313 10:27:21.668585 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="ceilometer-central-agent" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.668603 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="ceilometer-central-agent" Mar 13 10:27:21 crc kubenswrapper[4632]: E0313 10:27:21.668628 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58677e2e-9fc6-4e50-b342-e912afa8d969" containerName="cinder-api-log" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.668636 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="58677e2e-9fc6-4e50-b342-e912afa8d969" containerName="cinder-api-log" Mar 13 10:27:21 crc kubenswrapper[4632]: E0313 10:27:21.668651 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="ceilometer-notification-agent" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.668659 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="ceilometer-notification-agent" Mar 13 10:27:21 crc kubenswrapper[4632]: E0313 10:27:21.668685 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58677e2e-9fc6-4e50-b342-e912afa8d969" containerName="cinder-api" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.668695 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="58677e2e-9fc6-4e50-b342-e912afa8d969" containerName="cinder-api" Mar 13 10:27:21 crc kubenswrapper[4632]: E0313 10:27:21.668723 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="proxy-httpd" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.668730 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="proxy-httpd" Mar 13 10:27:21 crc kubenswrapper[4632]: E0313 10:27:21.668745 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="sg-core" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.668752 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="sg-core" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.668984 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="proxy-httpd" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.669006 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="58677e2e-9fc6-4e50-b342-e912afa8d969" containerName="cinder-api" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.669022 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="ceilometer-notification-agent" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.669041 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="sg-core" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.669053 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" containerName="ceilometer-central-agent" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.669070 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="58677e2e-9fc6-4e50-b342-e912afa8d969" containerName="cinder-api-log" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.676657 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.683359 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.683607 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.683788 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.720046 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.790274 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-config-data-custom\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.790325 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.790348 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.790370 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-config-data\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.790391 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6785ba8c-a47b-4851-945e-c07ccecb9911-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.790452 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89xkn\" (UniqueName: \"kubernetes.io/projected/6785ba8c-a47b-4851-945e-c07ccecb9911-kube-api-access-89xkn\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.790479 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-scripts\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.790509 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.790545 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6785ba8c-a47b-4851-945e-c07ccecb9911-logs\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.807050 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-kswhw"] Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.807418 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" (UID: "c4e63898-c65b-42c6-9ec5-3089ae7a8d8c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.811167 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-d856c56c-cmd2q"] Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.832489 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-7dbf8b9ddc-6p5vh"] Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.888599 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-config-data" (OuterVolumeSpecName: "config-data") pod "c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" (UID: "c4e63898-c65b-42c6-9ec5-3089ae7a8d8c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.902621 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-config-data-custom\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.902784 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.902855 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.902899 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-config-data\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.902988 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6785ba8c-a47b-4851-945e-c07ccecb9911-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.903065 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89xkn\" (UniqueName: \"kubernetes.io/projected/6785ba8c-a47b-4851-945e-c07ccecb9911-kube-api-access-89xkn\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.903109 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-scripts\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.903184 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.903416 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6785ba8c-a47b-4851-945e-c07ccecb9911-logs\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.904859 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.959091 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.936484 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6785ba8c-a47b-4851-945e-c07ccecb9911-logs\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.905243 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6785ba8c-a47b-4851-945e-c07ccecb9911-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.966799 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-config-data\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:21 crc kubenswrapper[4632]: I0313 10:27:21.970222 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-scripts\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.014419 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.004091 4632 scope.go:117] "RemoveContainer" containerID="a9d0bc7751d471197cb532c1a7e500502d2e1e74a150ed57680796972e393189" Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.027792 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-config-data-custom\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.028796 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89xkn\" (UniqueName: \"kubernetes.io/projected/6785ba8c-a47b-4851-945e-c07ccecb9911-kube-api-access-89xkn\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.031250 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.138578 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/6785ba8c-a47b-4851-945e-c07ccecb9911-public-tls-certs\") pod \"cinder-api-0\" (UID: \"6785ba8c-a47b-4851-945e-c07ccecb9911\") " pod="openstack/cinder-api-0" Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.252460 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58677e2e-9fc6-4e50-b342-e912afa8d969" path="/var/lib/kubelet/pods/58677e2e-9fc6-4e50-b342-e912afa8d969/volumes" Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.268496 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-fshjb"] Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.305834 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-86d4-account-create-update-5c7rj"] Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.343028 4632 scope.go:117] "RemoveContainer" containerID="d67a8223c96cadeaa871fcaaaad472258eb768daca2821f6757940c48f3eafd6" Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.350146 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" event={"ID":"904f04cd-8110-4637-8bb4-67c4b83e189b","Type":"ContainerStarted","Data":"b305d4370882ddeb316b7136e1b6a31fb9b050f68adc94baa9487a0176e85bb7"} Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.381637 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-b547848c4-bn5vs"] Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.384032 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-kswhw" event={"ID":"8e0fb1fc-c94a-44f0-a269-e7211c6fcfba","Type":"ContainerStarted","Data":"d2670f3af135aaf60a4a9f708985b74e740f5e7bc5471b2c38d01bfe606d1cfa"} Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.394248 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.396180 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-d856c56c-cmd2q" event={"ID":"5d10747e-ba77-4986-9d4b-636fcbf823ab","Type":"ContainerStarted","Data":"c4b118bba3eb9eaa2f3d30625225786b624eac290ce33f3a700f116e125abbc7"} Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.410010 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fshjb" event={"ID":"09bd98be-9d10-4a53-8ef6-c4718b05c3f6","Type":"ContainerStarted","Data":"c4ef1230411d68688aa0ca250739c759e42cb6d89e416542cd5bc528c1419eff"} Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.411162 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7f9df5b5b5-q6dp2"] Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.435196 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" event={"ID":"03ca050c-63a7-4b37-91fe-fe5c322cca78","Type":"ContainerStarted","Data":"fdf316a11cb9aca1f8cd6fb110ee0a2edd19f7386a5741d37b2da1b481bd8466"} Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.467509 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-2f8c-account-create-update-g4b8g"] Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.556106 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-wgv42"] Mar 13 10:27:22 crc kubenswrapper[4632]: I0313 10:27:22.595495 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-f3f1-account-create-update-29g8s"] Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:22.796599 4632 scope.go:117] "RemoveContainer" containerID="aa0a9edf7c00bb4d08cf1a3f2565b5016be14ce1312e093e1c44112d2d594f42" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.141335 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.182853 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.215472 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.233788 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.234164 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.250498 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.251339 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 10:27:23 crc kubenswrapper[4632]: E0313 10:27:23.263731 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4e63898_c65b_42c6_9ec5_3089ae7a8d8c.slice/crio-df15286148314ab907f4a05031eefbb838636621ee25cc2f368e3d56ae19621b\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc4e63898_c65b_42c6_9ec5_3089ae7a8d8c.slice\": RecentStats: unable to find data in memory cache]" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.270035 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.342587 4632 scope.go:117] "RemoveContainer" containerID="c7001d72ce189e15496046472a90b656a4129de71ad96c6f49a1d6b92862a990" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.343865 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6db55c595b-pwgcg" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.347021 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-scripts\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.347165 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/536490c7-c218-43ca-b601-84fdf0721b13-run-httpd\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.347205 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/536490c7-c218-43ca-b601-84fdf0721b13-log-httpd\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.347256 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-config-data\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.347401 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.347583 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crtzk\" (UniqueName: \"kubernetes.io/projected/536490c7-c218-43ca-b601-84fdf0721b13-kube-api-access-crtzk\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.347628 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.473451 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-scripts\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.473554 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/536490c7-c218-43ca-b601-84fdf0721b13-run-httpd\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.473586 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/536490c7-c218-43ca-b601-84fdf0721b13-log-httpd\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.473621 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-config-data\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.473746 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.473859 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crtzk\" (UniqueName: \"kubernetes.io/projected/536490c7-c218-43ca-b601-84fdf0721b13-kube-api-access-crtzk\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.473898 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.476071 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/536490c7-c218-43ca-b601-84fdf0721b13-run-httpd\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.480333 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/536490c7-c218-43ca-b601-84fdf0721b13-log-httpd\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.499826 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-scripts\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.512499 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7dd5c7bdcd-4969b"] Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.512814 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-7dd5c7bdcd-4969b" podUID="5abe7bf3-d44d-4ee5-b568-2d497868f1e5" containerName="placement-log" containerID="cri-o://0b15584f3607b654abe16b00ac290d1bc5ee6f763bd08234d8697e7f5b5b20bb" gracePeriod=30 Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.513014 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-7dd5c7bdcd-4969b" podUID="5abe7bf3-d44d-4ee5-b568-2d497868f1e5" containerName="placement-api" containerID="cri-o://2cfe7ebd70fe3427d7ef352e87ea88bca1736af36e0c260541ced9066c436503" gracePeriod=30 Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.532248 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.532549 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-config-data\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.533273 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.569477 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crtzk\" (UniqueName: \"kubernetes.io/projected/536490c7-c218-43ca-b601-84fdf0721b13-kube-api-access-crtzk\") pod \"ceilometer-0\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " pod="openstack/ceilometer-0" Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.577451 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" event={"ID":"8462be25-a577-476d-b54a-73790a8aa189","Type":"ContainerStarted","Data":"d5791dd2ca6757eefbd1007667971c0817daf4611f261b3e01b5a58673b3e353"} Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.578351 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.613223 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f3f1-account-create-update-29g8s" event={"ID":"234a900d-887b-448c-8336-010107726c1e","Type":"ContainerStarted","Data":"6eb1c0223253d25c6c8ddc47c33c06c64e6f5b3a0035afe9508e0276ff9d5aaf"} Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.642181 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-b547848c4-bn5vs" event={"ID":"07914020-653d-4509-9f60-22726224c7c6","Type":"ContainerStarted","Data":"bb01c2352414aa3e5bdfcb4abaaae4c47a152945a1d74d64f5cf1228335558e9"} Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.650120 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7f9df5b5b5-q6dp2" event={"ID":"757b852e-068c-4885-99b8-af2e6f23e445","Type":"ContainerStarted","Data":"db0a88b20ef1358b7cfb558aebb52cdeba5b5f143eee06ddc98fa0acfb3ab01b"} Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.708314 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wgv42" event={"ID":"f0c32ed5-c3b0-45ea-99de-87c45cb1ba77","Type":"ContainerStarted","Data":"645f6056b5af662b78f01666e55121dd00fa1cf0b8aa9bf79ae1ebdf5a74d21d"} Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.711220 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" event={"ID":"bbaf5a79-1c34-4518-afb9-19703fe6c45b","Type":"ContainerStarted","Data":"32dece687daf55d82de95e6edff961fa23c0b30410ca7620ec1c15a1d72b8f64"} Mar 13 10:27:23 crc kubenswrapper[4632]: I0313 10:27:23.750262 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-kswhw" podStartSLOduration=5.750232897 podStartE2EDuration="5.750232897s" podCreationTimestamp="2026-03-13 10:27:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:27:23.737241921 +0000 UTC m=+1417.759772054" watchObservedRunningTime="2026-03-13 10:27:23.750232897 +0000 UTC m=+1417.772763030" Mar 13 10:27:24 crc kubenswrapper[4632]: I0313 10:27:24.025288 4632 scope.go:117] "RemoveContainer" containerID="bb00bf460a4849cc1a7c1bad8a739981e87c032a18a9222632d57abbccea8858" Mar 13 10:27:24 crc kubenswrapper[4632]: I0313 10:27:24.080962 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4e63898-c65b-42c6-9ec5-3089ae7a8d8c" path="/var/lib/kubelet/pods/c4e63898-c65b-42c6-9ec5-3089ae7a8d8c/volumes" Mar 13 10:27:24 crc kubenswrapper[4632]: I0313 10:27:24.314121 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:27:24 crc kubenswrapper[4632]: I0313 10:27:24.584064 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-c959f64fb-hx4t8"] Mar 13 10:27:24 crc kubenswrapper[4632]: I0313 10:27:24.608009 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:24 crc kubenswrapper[4632]: I0313 10:27:24.620864 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-c959f64fb-hx4t8"] Mar 13 10:27:24 crc kubenswrapper[4632]: I0313 10:27:24.652580 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53145947-4584-4cef-b085-a0e0f550dde9-config-data-custom\") pod \"heat-engine-c959f64fb-hx4t8\" (UID: \"53145947-4584-4cef-b085-a0e0f550dde9\") " pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:24 crc kubenswrapper[4632]: I0313 10:27:24.652817 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsdfl\" (UniqueName: \"kubernetes.io/projected/53145947-4584-4cef-b085-a0e0f550dde9-kube-api-access-bsdfl\") pod \"heat-engine-c959f64fb-hx4t8\" (UID: \"53145947-4584-4cef-b085-a0e0f550dde9\") " pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:24 crc kubenswrapper[4632]: I0313 10:27:24.652885 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53145947-4584-4cef-b085-a0e0f550dde9-combined-ca-bundle\") pod \"heat-engine-c959f64fb-hx4t8\" (UID: \"53145947-4584-4cef-b085-a0e0f550dde9\") " pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:24 crc kubenswrapper[4632]: I0313 10:27:24.653228 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53145947-4584-4cef-b085-a0e0f550dde9-config-data\") pod \"heat-engine-c959f64fb-hx4t8\" (UID: \"53145947-4584-4cef-b085-a0e0f550dde9\") " pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:24 crc kubenswrapper[4632]: I0313 10:27:24.693383 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-6f597ccc7c-zgmpr"] Mar 13 10:27:24 crc kubenswrapper[4632]: I0313 10:27:24.694890 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.769309 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6f597ccc7c-zgmpr"] Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.769373 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-66b64f87f7-6z95j"] Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.774115 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53145947-4584-4cef-b085-a0e0f550dde9-config-data\") pod \"heat-engine-c959f64fb-hx4t8\" (UID: \"53145947-4584-4cef-b085-a0e0f550dde9\") " pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.774184 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53145947-4584-4cef-b085-a0e0f550dde9-config-data-custom\") pod \"heat-engine-c959f64fb-hx4t8\" (UID: \"53145947-4584-4cef-b085-a0e0f550dde9\") " pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.774210 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-config-data-custom\") pod \"heat-api-6f597ccc7c-zgmpr\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.774281 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-combined-ca-bundle\") pod \"heat-api-6f597ccc7c-zgmpr\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.774311 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsdfl\" (UniqueName: \"kubernetes.io/projected/53145947-4584-4cef-b085-a0e0f550dde9-kube-api-access-bsdfl\") pod \"heat-engine-c959f64fb-hx4t8\" (UID: \"53145947-4584-4cef-b085-a0e0f550dde9\") " pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.774330 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53145947-4584-4cef-b085-a0e0f550dde9-combined-ca-bundle\") pod \"heat-engine-c959f64fb-hx4t8\" (UID: \"53145947-4584-4cef-b085-a0e0f550dde9\") " pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.774375 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-config-data\") pod \"heat-api-6f597ccc7c-zgmpr\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.774417 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vct9r\" (UniqueName: \"kubernetes.io/projected/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-kube-api-access-vct9r\") pod \"heat-api-6f597ccc7c-zgmpr\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.788050 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53145947-4584-4cef-b085-a0e0f550dde9-config-data\") pod \"heat-engine-c959f64fb-hx4t8\" (UID: \"53145947-4584-4cef-b085-a0e0f550dde9\") " pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.812818 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/53145947-4584-4cef-b085-a0e0f550dde9-config-data-custom\") pod \"heat-engine-c959f64fb-hx4t8\" (UID: \"53145947-4584-4cef-b085-a0e0f550dde9\") " pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.815139 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-66b64f87f7-6z95j"] Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.815245 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.826909 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53145947-4584-4cef-b085-a0e0f550dde9-combined-ca-bundle\") pod \"heat-engine-c959f64fb-hx4t8\" (UID: \"53145947-4584-4cef-b085-a0e0f550dde9\") " pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.835706 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsdfl\" (UniqueName: \"kubernetes.io/projected/53145947-4584-4cef-b085-a0e0f550dde9-kube-api-access-bsdfl\") pod \"heat-engine-c959f64fb-hx4t8\" (UID: \"53145947-4584-4cef-b085-a0e0f550dde9\") " pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.841550 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" event={"ID":"8462be25-a577-476d-b54a-73790a8aa189","Type":"ContainerStarted","Data":"baa73e1779483e615256cb324392bd7ff43cccd507e79b501108b7a61007ed58"} Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.886172 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-config-data-custom\") pod \"heat-api-6f597ccc7c-zgmpr\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.886247 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-combined-ca-bundle\") pod \"heat-api-6f597ccc7c-zgmpr\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.886279 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-config-data\") pod \"heat-cfnapi-66b64f87f7-6z95j\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.886312 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-config-data-custom\") pod \"heat-cfnapi-66b64f87f7-6z95j\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.886345 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-config-data\") pod \"heat-api-6f597ccc7c-zgmpr\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.886372 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-combined-ca-bundle\") pod \"heat-cfnapi-66b64f87f7-6z95j\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.886404 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vct9r\" (UniqueName: \"kubernetes.io/projected/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-kube-api-access-vct9r\") pod \"heat-api-6f597ccc7c-zgmpr\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.886426 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5qrk\" (UniqueName: \"kubernetes.io/projected/8bca285e-17f7-4505-8a25-21f5ee739584-kube-api-access-w5qrk\") pod \"heat-cfnapi-66b64f87f7-6z95j\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.887761 4632 generic.go:334] "Generic (PLEG): container finished" podID="09bd98be-9d10-4a53-8ef6-c4718b05c3f6" containerID="a73d11226d1411728675707324588174ab20222ac0a86a31f153adf5c08496b7" exitCode=0 Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.887814 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fshjb" event={"ID":"09bd98be-9d10-4a53-8ef6-c4718b05c3f6","Type":"ContainerDied","Data":"a73d11226d1411728675707324588174ab20222ac0a86a31f153adf5c08496b7"} Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.901719 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-config-data\") pod \"heat-api-6f597ccc7c-zgmpr\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.903348 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-config-data-custom\") pod \"heat-api-6f597ccc7c-zgmpr\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.903922 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-combined-ca-bundle\") pod \"heat-api-6f597ccc7c-zgmpr\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.909850 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" podStartSLOduration=5.909822227 podStartE2EDuration="5.909822227s" podCreationTimestamp="2026-03-13 10:27:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:27:24.885005621 +0000 UTC m=+1418.907535744" watchObservedRunningTime="2026-03-13 10:27:24.909822227 +0000 UTC m=+1418.932352360" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.925244 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" event={"ID":"03ca050c-63a7-4b37-91fe-fe5c322cca78","Type":"ContainerStarted","Data":"94b786c65a2ca6a08eecb9fac67251053ce759dcb6a34953019d7b0f5ae51054"} Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.930971 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vct9r\" (UniqueName: \"kubernetes.io/projected/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-kube-api-access-vct9r\") pod \"heat-api-6f597ccc7c-zgmpr\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.931899 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6785ba8c-a47b-4851-945e-c07ccecb9911","Type":"ContainerStarted","Data":"b0c8d0ab5767df5dcc1a0b06d386ffcd85be89291952021813dc42c8e426af90"} Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.980482 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.988293 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-config-data\") pod \"heat-cfnapi-66b64f87f7-6z95j\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.988354 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-config-data-custom\") pod \"heat-cfnapi-66b64f87f7-6z95j\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.988392 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-combined-ca-bundle\") pod \"heat-cfnapi-66b64f87f7-6z95j\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.988433 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5qrk\" (UniqueName: \"kubernetes.io/projected/8bca285e-17f7-4505-8a25-21f5ee739584-kube-api-access-w5qrk\") pod \"heat-cfnapi-66b64f87f7-6z95j\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.996959 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-combined-ca-bundle\") pod \"heat-cfnapi-66b64f87f7-6z95j\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.997637 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-config-data-custom\") pod \"heat-cfnapi-66b64f87f7-6z95j\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:24.998751 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-config-data\") pod \"heat-cfnapi-66b64f87f7-6z95j\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.017229 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5qrk\" (UniqueName: \"kubernetes.io/projected/8bca285e-17f7-4505-8a25-21f5ee739584-kube-api-access-w5qrk\") pod \"heat-cfnapi-66b64f87f7-6z95j\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.053030 4632 generic.go:334] "Generic (PLEG): container finished" podID="904f04cd-8110-4637-8bb4-67c4b83e189b" containerID="10ef0805fc14af19dcea5ad4d4426bd1471fa5008be0ab704ad9b901662ea060" exitCode=0 Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.053110 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" event={"ID":"904f04cd-8110-4637-8bb4-67c4b83e189b","Type":"ContainerDied","Data":"10ef0805fc14af19dcea5ad4d4426bd1471fa5008be0ab704ad9b901662ea060"} Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.082237 4632 generic.go:334] "Generic (PLEG): container finished" podID="5abe7bf3-d44d-4ee5-b568-2d497868f1e5" containerID="0b15584f3607b654abe16b00ac290d1bc5ee6f763bd08234d8697e7f5b5b20bb" exitCode=143 Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.082339 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7dd5c7bdcd-4969b" event={"ID":"5abe7bf3-d44d-4ee5-b568-2d497868f1e5","Type":"ContainerDied","Data":"0b15584f3607b654abe16b00ac290d1bc5ee6f763bd08234d8697e7f5b5b20bb"} Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.097032 4632 generic.go:334] "Generic (PLEG): container finished" podID="8e0fb1fc-c94a-44f0-a269-e7211c6fcfba" containerID="f531fb1c9798e5386771f799aeaf5ec81a37e70faa215029f1e44845844c0b7a" exitCode=0 Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.097099 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-kswhw" event={"ID":"8e0fb1fc-c94a-44f0-a269-e7211c6fcfba","Type":"ContainerDied","Data":"f531fb1c9798e5386771f799aeaf5ec81a37e70faa215029f1e44845844c0b7a"} Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.233312 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6588559b77-6f4bf" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.239443 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.255019 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.368050 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5c86b4b888-l9574"] Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.368286 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5c86b4b888-l9574" podUID="6d73a499-d334-4a7a-9783-640b98760672" containerName="neutron-api" containerID="cri-o://8c839401b1db62da93454588496b8ab534c9e6313aa3bcb0003cb9137b63b2ca" gracePeriod=30 Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.368838 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-5c86b4b888-l9574" podUID="6d73a499-d334-4a7a-9783-640b98760672" containerName="neutron-httpd" containerID="cri-o://e005b4f09b297f1fe00efd39c9534b7382173cd69b88dca5466ba89c0f3c0de7" gracePeriod=30 Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.396295 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.793874 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:27:25 crc kubenswrapper[4632]: I0313 10:27:25.858160 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-689764498d-rg7vt" podUID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.120245 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7f9df5b5b5-q6dp2" event={"ID":"757b852e-068c-4885-99b8-af2e6f23e445","Type":"ContainerStarted","Data":"c974019fb18638d059aa9b080871c3232dbdb322c997ebb8d28de7a80fef50a0"} Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.120372 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.122960 4632 generic.go:334] "Generic (PLEG): container finished" podID="6d73a499-d334-4a7a-9783-640b98760672" containerID="e005b4f09b297f1fe00efd39c9534b7382173cd69b88dca5466ba89c0f3c0de7" exitCode=0 Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.122981 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c86b4b888-l9574" event={"ID":"6d73a499-d334-4a7a-9783-640b98760672","Type":"ContainerDied","Data":"e005b4f09b297f1fe00efd39c9534b7382173cd69b88dca5466ba89c0f3c0de7"} Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.126908 4632 generic.go:334] "Generic (PLEG): container finished" podID="f0c32ed5-c3b0-45ea-99de-87c45cb1ba77" containerID="62c66b71b16f2cd37ff478080f4c30eed65f51b807f687725f8ec89f5dd9d0dc" exitCode=0 Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.127004 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wgv42" event={"ID":"f0c32ed5-c3b0-45ea-99de-87c45cb1ba77","Type":"ContainerDied","Data":"62c66b71b16f2cd37ff478080f4c30eed65f51b807f687725f8ec89f5dd9d0dc"} Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.129917 4632 generic.go:334] "Generic (PLEG): container finished" podID="8462be25-a577-476d-b54a-73790a8aa189" containerID="baa73e1779483e615256cb324392bd7ff43cccd507e79b501108b7a61007ed58" exitCode=0 Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.129994 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" event={"ID":"8462be25-a577-476d-b54a-73790a8aa189","Type":"ContainerDied","Data":"baa73e1779483e615256cb324392bd7ff43cccd507e79b501108b7a61007ed58"} Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.134322 4632 generic.go:334] "Generic (PLEG): container finished" podID="234a900d-887b-448c-8336-010107726c1e" containerID="9cee7abc6c76d73494106b5582f85b871d225f179b8f40700ad2248a8daa7c60" exitCode=0 Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.134407 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f3f1-account-create-update-29g8s" event={"ID":"234a900d-887b-448c-8336-010107726c1e","Type":"ContainerDied","Data":"9cee7abc6c76d73494106b5582f85b871d225f179b8f40700ad2248a8daa7c60"} Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.142082 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" event={"ID":"bbaf5a79-1c34-4518-afb9-19703fe6c45b","Type":"ContainerStarted","Data":"d9f2ab5e1a5be1d4939b9fe05ba3a5cdbc725953ea1e78a027cf1f61d4444ba0"} Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.142136 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7f9df5b5b5-q6dp2" podStartSLOduration=13.142116976 podStartE2EDuration="13.142116976s" podCreationTimestamp="2026-03-13 10:27:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:27:26.137892606 +0000 UTC m=+1420.160422759" watchObservedRunningTime="2026-03-13 10:27:26.142116976 +0000 UTC m=+1420.164647099" Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.149248 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" event={"ID":"03ca050c-63a7-4b37-91fe-fe5c322cca78","Type":"ContainerStarted","Data":"d8e95d6e44c7d021f2ace88c9a6c134873fb843b4515ef7629a6970c5d1bd8f9"} Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.149610 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.149738 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.155779 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6785ba8c-a47b-4851-945e-c07ccecb9911","Type":"ContainerStarted","Data":"33649b65fc233ba28987b27f24de1fa0c00851971451968b6f04cf9bbe1f240d"} Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.236656 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" podStartSLOduration=7.236633671 podStartE2EDuration="7.236633671s" podCreationTimestamp="2026-03-13 10:27:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:27:26.227118226 +0000 UTC m=+1420.249648369" watchObservedRunningTime="2026-03-13 10:27:26.236633671 +0000 UTC m=+1420.259163804" Mar 13 10:27:26 crc kubenswrapper[4632]: I0313 10:27:26.277978 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" podStartSLOduration=16.277934496 podStartE2EDuration="16.277934496s" podCreationTimestamp="2026-03-13 10:27:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:27:26.253115419 +0000 UTC m=+1420.275645552" watchObservedRunningTime="2026-03-13 10:27:26.277934496 +0000 UTC m=+1420.300464649" Mar 13 10:27:27 crc kubenswrapper[4632]: W0313 10:27:27.168031 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod536490c7_c218_43ca_b601_84fdf0721b13.slice/crio-f402196243381c3faf3165d4fe49b7c43a1af16813bae58fca9b53eb4badf807 WatchSource:0}: Error finding container f402196243381c3faf3165d4fe49b7c43a1af16813bae58fca9b53eb4badf807: Status 404 returned error can't find the container with id f402196243381c3faf3165d4fe49b7c43a1af16813bae58fca9b53eb4badf807 Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.196609 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-fshjb" event={"ID":"09bd98be-9d10-4a53-8ef6-c4718b05c3f6","Type":"ContainerDied","Data":"c4ef1230411d68688aa0ca250739c759e42cb6d89e416542cd5bc528c1419eff"} Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.196660 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4ef1230411d68688aa0ca250739c759e42cb6d89e416542cd5bc528c1419eff" Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.222003 4632 generic.go:334] "Generic (PLEG): container finished" podID="bbaf5a79-1c34-4518-afb9-19703fe6c45b" containerID="d9f2ab5e1a5be1d4939b9fe05ba3a5cdbc725953ea1e78a027cf1f61d4444ba0" exitCode=0 Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.222082 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" event={"ID":"bbaf5a79-1c34-4518-afb9-19703fe6c45b","Type":"ContainerDied","Data":"d9f2ab5e1a5be1d4939b9fe05ba3a5cdbc725953ea1e78a027cf1f61d4444ba0"} Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.245549 4632 generic.go:334] "Generic (PLEG): container finished" podID="5abe7bf3-d44d-4ee5-b568-2d497868f1e5" containerID="2cfe7ebd70fe3427d7ef352e87ea88bca1736af36e0c260541ced9066c436503" exitCode=0 Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.245877 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7dd5c7bdcd-4969b" event={"ID":"5abe7bf3-d44d-4ee5-b568-2d497868f1e5","Type":"ContainerDied","Data":"2cfe7ebd70fe3427d7ef352e87ea88bca1736af36e0c260541ced9066c436503"} Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.258122 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-kswhw" event={"ID":"8e0fb1fc-c94a-44f0-a269-e7211c6fcfba","Type":"ContainerDied","Data":"d2670f3af135aaf60a4a9f708985b74e740f5e7bc5471b2c38d01bfe606d1cfa"} Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.258168 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2670f3af135aaf60a4a9f708985b74e740f5e7bc5471b2c38d01bfe606d1cfa" Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.390806 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fshjb" Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.400889 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kswhw" Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.482609 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdg5s\" (UniqueName: \"kubernetes.io/projected/8e0fb1fc-c94a-44f0-a269-e7211c6fcfba-kube-api-access-jdg5s\") pod \"8e0fb1fc-c94a-44f0-a269-e7211c6fcfba\" (UID: \"8e0fb1fc-c94a-44f0-a269-e7211c6fcfba\") " Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.482812 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09bd98be-9d10-4a53-8ef6-c4718b05c3f6-operator-scripts\") pod \"09bd98be-9d10-4a53-8ef6-c4718b05c3f6\" (UID: \"09bd98be-9d10-4a53-8ef6-c4718b05c3f6\") " Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.482956 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e0fb1fc-c94a-44f0-a269-e7211c6fcfba-operator-scripts\") pod \"8e0fb1fc-c94a-44f0-a269-e7211c6fcfba\" (UID: \"8e0fb1fc-c94a-44f0-a269-e7211c6fcfba\") " Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.482989 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cl442\" (UniqueName: \"kubernetes.io/projected/09bd98be-9d10-4a53-8ef6-c4718b05c3f6-kube-api-access-cl442\") pod \"09bd98be-9d10-4a53-8ef6-c4718b05c3f6\" (UID: \"09bd98be-9d10-4a53-8ef6-c4718b05c3f6\") " Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.483603 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09bd98be-9d10-4a53-8ef6-c4718b05c3f6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "09bd98be-9d10-4a53-8ef6-c4718b05c3f6" (UID: "09bd98be-9d10-4a53-8ef6-c4718b05c3f6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.485748 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e0fb1fc-c94a-44f0-a269-e7211c6fcfba-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8e0fb1fc-c94a-44f0-a269-e7211c6fcfba" (UID: "8e0fb1fc-c94a-44f0-a269-e7211c6fcfba"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.507131 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e0fb1fc-c94a-44f0-a269-e7211c6fcfba-kube-api-access-jdg5s" (OuterVolumeSpecName: "kube-api-access-jdg5s") pod "8e0fb1fc-c94a-44f0-a269-e7211c6fcfba" (UID: "8e0fb1fc-c94a-44f0-a269-e7211c6fcfba"). InnerVolumeSpecName "kube-api-access-jdg5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.508727 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09bd98be-9d10-4a53-8ef6-c4718b05c3f6-kube-api-access-cl442" (OuterVolumeSpecName: "kube-api-access-cl442") pod "09bd98be-9d10-4a53-8ef6-c4718b05c3f6" (UID: "09bd98be-9d10-4a53-8ef6-c4718b05c3f6"). InnerVolumeSpecName "kube-api-access-cl442". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.585648 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cl442\" (UniqueName: \"kubernetes.io/projected/09bd98be-9d10-4a53-8ef6-c4718b05c3f6-kube-api-access-cl442\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.585690 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdg5s\" (UniqueName: \"kubernetes.io/projected/8e0fb1fc-c94a-44f0-a269-e7211c6fcfba-kube-api-access-jdg5s\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.585700 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09bd98be-9d10-4a53-8ef6-c4718b05c3f6-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:27 crc kubenswrapper[4632]: I0313 10:27:27.585709 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e0fb1fc-c94a-44f0-a269-e7211c6fcfba-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:28 crc kubenswrapper[4632]: I0313 10:27:28.277074 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"536490c7-c218-43ca-b601-84fdf0721b13","Type":"ContainerStarted","Data":"f402196243381c3faf3165d4fe49b7c43a1af16813bae58fca9b53eb4badf807"} Mar 13 10:27:28 crc kubenswrapper[4632]: I0313 10:27:28.277425 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-kswhw" Mar 13 10:27:28 crc kubenswrapper[4632]: I0313 10:27:28.278252 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-fshjb" Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.686643 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f3f1-account-create-update-29g8s" Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.694134 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.736314 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wgv42" Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.803395 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.830739 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-logs\") pod \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.830827 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-internal-tls-certs\") pod \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.830893 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-config-data\") pod \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.830918 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xk8t2\" (UniqueName: \"kubernetes.io/projected/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-kube-api-access-xk8t2\") pod \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.830969 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-combined-ca-bundle\") pod \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.831003 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/234a900d-887b-448c-8336-010107726c1e-operator-scripts\") pod \"234a900d-887b-448c-8336-010107726c1e\" (UID: \"234a900d-887b-448c-8336-010107726c1e\") " Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.831096 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-scripts\") pod \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.831291 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-public-tls-certs\") pod \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\" (UID: \"5abe7bf3-d44d-4ee5-b568-2d497868f1e5\") " Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.831331 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qgng\" (UniqueName: \"kubernetes.io/projected/234a900d-887b-448c-8336-010107726c1e-kube-api-access-9qgng\") pod \"234a900d-887b-448c-8336-010107726c1e\" (UID: \"234a900d-887b-448c-8336-010107726c1e\") " Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.838141 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-logs" (OuterVolumeSpecName: "logs") pod "5abe7bf3-d44d-4ee5-b568-2d497868f1e5" (UID: "5abe7bf3-d44d-4ee5-b568-2d497868f1e5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.838285 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/234a900d-887b-448c-8336-010107726c1e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "234a900d-887b-448c-8336-010107726c1e" (UID: "234a900d-887b-448c-8336-010107726c1e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.880868 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-kube-api-access-xk8t2" (OuterVolumeSpecName: "kube-api-access-xk8t2") pod "5abe7bf3-d44d-4ee5-b568-2d497868f1e5" (UID: "5abe7bf3-d44d-4ee5-b568-2d497868f1e5"). InnerVolumeSpecName "kube-api-access-xk8t2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.881126 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/234a900d-887b-448c-8336-010107726c1e-kube-api-access-9qgng" (OuterVolumeSpecName: "kube-api-access-9qgng") pod "234a900d-887b-448c-8336-010107726c1e" (UID: "234a900d-887b-448c-8336-010107726c1e"). InnerVolumeSpecName "kube-api-access-9qgng". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.893273 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.902251 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-scripts" (OuterVolumeSpecName: "scripts") pod "5abe7bf3-d44d-4ee5-b568-2d497868f1e5" (UID: "5abe7bf3-d44d-4ee5-b568-2d497868f1e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.996477 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5fx7d\" (UniqueName: \"kubernetes.io/projected/bbaf5a79-1c34-4518-afb9-19703fe6c45b-kube-api-access-5fx7d\") pod \"bbaf5a79-1c34-4518-afb9-19703fe6c45b\" (UID: \"bbaf5a79-1c34-4518-afb9-19703fe6c45b\") " Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.996535 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rrgfq\" (UniqueName: \"kubernetes.io/projected/8462be25-a577-476d-b54a-73790a8aa189-kube-api-access-rrgfq\") pod \"8462be25-a577-476d-b54a-73790a8aa189\" (UID: \"8462be25-a577-476d-b54a-73790a8aa189\") " Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.996775 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0c32ed5-c3b0-45ea-99de-87c45cb1ba77-operator-scripts\") pod \"f0c32ed5-c3b0-45ea-99de-87c45cb1ba77\" (UID: \"f0c32ed5-c3b0-45ea-99de-87c45cb1ba77\") " Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.996822 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbaf5a79-1c34-4518-afb9-19703fe6c45b-operator-scripts\") pod \"bbaf5a79-1c34-4518-afb9-19703fe6c45b\" (UID: \"bbaf5a79-1c34-4518-afb9-19703fe6c45b\") " Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.996881 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8462be25-a577-476d-b54a-73790a8aa189-operator-scripts\") pod \"8462be25-a577-476d-b54a-73790a8aa189\" (UID: \"8462be25-a577-476d-b54a-73790a8aa189\") " Mar 13 10:27:29 crc kubenswrapper[4632]: I0313 10:27:29.997011 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nsxhh\" (UniqueName: \"kubernetes.io/projected/f0c32ed5-c3b0-45ea-99de-87c45cb1ba77-kube-api-access-nsxhh\") pod \"f0c32ed5-c3b0-45ea-99de-87c45cb1ba77\" (UID: \"f0c32ed5-c3b0-45ea-99de-87c45cb1ba77\") " Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:29.997628 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bbaf5a79-1c34-4518-afb9-19703fe6c45b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "bbaf5a79-1c34-4518-afb9-19703fe6c45b" (UID: "bbaf5a79-1c34-4518-afb9-19703fe6c45b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:29.997955 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0c32ed5-c3b0-45ea-99de-87c45cb1ba77-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f0c32ed5-c3b0-45ea-99de-87c45cb1ba77" (UID: "f0c32ed5-c3b0-45ea-99de-87c45cb1ba77"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:29.999009 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8462be25-a577-476d-b54a-73790a8aa189-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8462be25-a577-476d-b54a-73790a8aa189" (UID: "8462be25-a577-476d-b54a-73790a8aa189"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.010462 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qgng\" (UniqueName: \"kubernetes.io/projected/234a900d-887b-448c-8336-010107726c1e-kube-api-access-9qgng\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.010500 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f0c32ed5-c3b0-45ea-99de-87c45cb1ba77-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.010512 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/bbaf5a79-1c34-4518-afb9-19703fe6c45b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.010538 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.010553 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8462be25-a577-476d-b54a-73790a8aa189-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.010566 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xk8t2\" (UniqueName: \"kubernetes.io/projected/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-kube-api-access-xk8t2\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.010576 4632 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/234a900d-887b-448c-8336-010107726c1e-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.010585 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.073033 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbaf5a79-1c34-4518-afb9-19703fe6c45b-kube-api-access-5fx7d" (OuterVolumeSpecName: "kube-api-access-5fx7d") pod "bbaf5a79-1c34-4518-afb9-19703fe6c45b" (UID: "bbaf5a79-1c34-4518-afb9-19703fe6c45b"). InnerVolumeSpecName "kube-api-access-5fx7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.077658 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0c32ed5-c3b0-45ea-99de-87c45cb1ba77-kube-api-access-nsxhh" (OuterVolumeSpecName: "kube-api-access-nsxhh") pod "f0c32ed5-c3b0-45ea-99de-87c45cb1ba77" (UID: "f0c32ed5-c3b0-45ea-99de-87c45cb1ba77"). InnerVolumeSpecName "kube-api-access-nsxhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.085702 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8462be25-a577-476d-b54a-73790a8aa189-kube-api-access-rrgfq" (OuterVolumeSpecName: "kube-api-access-rrgfq") pod "8462be25-a577-476d-b54a-73790a8aa189" (UID: "8462be25-a577-476d-b54a-73790a8aa189"). InnerVolumeSpecName "kube-api-access-rrgfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.126424 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5fx7d\" (UniqueName: \"kubernetes.io/projected/bbaf5a79-1c34-4518-afb9-19703fe6c45b-kube-api-access-5fx7d\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.126453 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rrgfq\" (UniqueName: \"kubernetes.io/projected/8462be25-a577-476d-b54a-73790a8aa189-kube-api-access-rrgfq\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.126464 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nsxhh\" (UniqueName: \"kubernetes.io/projected/f0c32ed5-c3b0-45ea-99de-87c45cb1ba77-kube-api-access-nsxhh\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.328403 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-66b64f87f7-6z95j"] Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.352118 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" event={"ID":"bbaf5a79-1c34-4518-afb9-19703fe6c45b","Type":"ContainerDied","Data":"32dece687daf55d82de95e6edff961fa23c0b30410ca7620ec1c15a1d72b8f64"} Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.352160 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32dece687daf55d82de95e6edff961fa23c0b30410ca7620ec1c15a1d72b8f64" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.352228 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-86d4-account-create-update-5c7rj" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.357575 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-7dd5c7bdcd-4969b" event={"ID":"5abe7bf3-d44d-4ee5-b568-2d497868f1e5","Type":"ContainerDied","Data":"604b160eb4cd534ac8def868fbcdab1d748e8bc2952c85fe7198dc4a2b05d7f7"} Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.357627 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-7dd5c7bdcd-4969b" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.357633 4632 scope.go:117] "RemoveContainer" containerID="2cfe7ebd70fe3427d7ef352e87ea88bca1736af36e0c260541ced9066c436503" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.420246 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-wgv42" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.420650 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-wgv42" event={"ID":"f0c32ed5-c3b0-45ea-99de-87c45cb1ba77","Type":"ContainerDied","Data":"645f6056b5af662b78f01666e55121dd00fa1cf0b8aa9bf79ae1ebdf5a74d21d"} Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.420701 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="645f6056b5af662b78f01666e55121dd00fa1cf0b8aa9bf79ae1ebdf5a74d21d" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.461389 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" event={"ID":"8462be25-a577-476d-b54a-73790a8aa189","Type":"ContainerDied","Data":"d5791dd2ca6757eefbd1007667971c0817daf4611f261b3e01b5a58673b3e353"} Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.461435 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d5791dd2ca6757eefbd1007667971c0817daf4611f261b3e01b5a58673b3e353" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.461562 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-2f8c-account-create-update-g4b8g" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.498068 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-f3f1-account-create-update-29g8s" event={"ID":"234a900d-887b-448c-8336-010107726c1e","Type":"ContainerDied","Data":"6eb1c0223253d25c6c8ddc47c33c06c64e6f5b3a0035afe9508e0276ff9d5aaf"} Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.498143 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6eb1c0223253d25c6c8ddc47c33c06c64e6f5b3a0035afe9508e0276ff9d5aaf" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.498237 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-f3f1-account-create-update-29g8s" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.558577 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-c959f64fb-hx4t8"] Mar 13 10:27:30 crc kubenswrapper[4632]: W0313 10:27:30.599101 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod53145947_4584_4cef_b085_a0e0f550dde9.slice/crio-4b36cced26c065af3bdca097c2ba77c4f4683335facc1febb5ca7b6e7e41d41b WatchSource:0}: Error finding container 4b36cced26c065af3bdca097c2ba77c4f4683335facc1febb5ca7b6e7e41d41b: Status 404 returned error can't find the container with id 4b36cced26c065af3bdca097c2ba77c4f4683335facc1febb5ca7b6e7e41d41b Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.644008 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-6f597ccc7c-zgmpr"] Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.650549 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.670906 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.704723 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5abe7bf3-d44d-4ee5-b568-2d497868f1e5" (UID: "5abe7bf3-d44d-4ee5-b568-2d497868f1e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.799143 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.908994 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5abe7bf3-d44d-4ee5-b568-2d497868f1e5" (UID: "5abe7bf3-d44d-4ee5-b568-2d497868f1e5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.918775 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-config-data" (OuterVolumeSpecName: "config-data") pod "5abe7bf3-d44d-4ee5-b568-2d497868f1e5" (UID: "5abe7bf3-d44d-4ee5-b568-2d497868f1e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.942404 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5abe7bf3-d44d-4ee5-b568-2d497868f1e5" (UID: "5abe7bf3-d44d-4ee5-b568-2d497868f1e5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:30 crc kubenswrapper[4632]: I0313 10:27:30.968071 4632 scope.go:117] "RemoveContainer" containerID="0b15584f3607b654abe16b00ac290d1bc5ee6f763bd08234d8697e7f5b5b20bb" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.006868 4632 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.006916 4632 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.006933 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5abe7bf3-d44d-4ee5-b568-2d497868f1e5-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.060596 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-7dd5c7bdcd-4969b"] Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.090286 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-7dd5c7bdcd-4969b"] Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.118329 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-b547848c4-bn5vs"] Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.157340 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-d856c56c-cmd2q"] Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.177620 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-86bb565f45-ntq5k"] Mar 13 10:27:31 crc kubenswrapper[4632]: E0313 10:27:31.182189 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5abe7bf3-d44d-4ee5-b568-2d497868f1e5" containerName="placement-log" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182237 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="5abe7bf3-d44d-4ee5-b568-2d497868f1e5" containerName="placement-log" Mar 13 10:27:31 crc kubenswrapper[4632]: E0313 10:27:31.182267 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5abe7bf3-d44d-4ee5-b568-2d497868f1e5" containerName="placement-api" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182289 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="5abe7bf3-d44d-4ee5-b568-2d497868f1e5" containerName="placement-api" Mar 13 10:27:31 crc kubenswrapper[4632]: E0313 10:27:31.182303 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="234a900d-887b-448c-8336-010107726c1e" containerName="mariadb-account-create-update" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182310 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="234a900d-887b-448c-8336-010107726c1e" containerName="mariadb-account-create-update" Mar 13 10:27:31 crc kubenswrapper[4632]: E0313 10:27:31.182326 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbaf5a79-1c34-4518-afb9-19703fe6c45b" containerName="mariadb-account-create-update" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182332 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbaf5a79-1c34-4518-afb9-19703fe6c45b" containerName="mariadb-account-create-update" Mar 13 10:27:31 crc kubenswrapper[4632]: E0313 10:27:31.182347 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09bd98be-9d10-4a53-8ef6-c4718b05c3f6" containerName="mariadb-database-create" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182369 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="09bd98be-9d10-4a53-8ef6-c4718b05c3f6" containerName="mariadb-database-create" Mar 13 10:27:31 crc kubenswrapper[4632]: E0313 10:27:31.182382 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0c32ed5-c3b0-45ea-99de-87c45cb1ba77" containerName="mariadb-database-create" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182388 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0c32ed5-c3b0-45ea-99de-87c45cb1ba77" containerName="mariadb-database-create" Mar 13 10:27:31 crc kubenswrapper[4632]: E0313 10:27:31.182398 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e0fb1fc-c94a-44f0-a269-e7211c6fcfba" containerName="mariadb-database-create" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182404 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e0fb1fc-c94a-44f0-a269-e7211c6fcfba" containerName="mariadb-database-create" Mar 13 10:27:31 crc kubenswrapper[4632]: E0313 10:27:31.182411 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8462be25-a577-476d-b54a-73790a8aa189" containerName="mariadb-account-create-update" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182418 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8462be25-a577-476d-b54a-73790a8aa189" containerName="mariadb-account-create-update" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182753 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="5abe7bf3-d44d-4ee5-b568-2d497868f1e5" containerName="placement-log" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182769 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8462be25-a577-476d-b54a-73790a8aa189" containerName="mariadb-account-create-update" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182780 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0c32ed5-c3b0-45ea-99de-87c45cb1ba77" containerName="mariadb-database-create" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182786 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbaf5a79-1c34-4518-afb9-19703fe6c45b" containerName="mariadb-account-create-update" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182796 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="09bd98be-9d10-4a53-8ef6-c4718b05c3f6" containerName="mariadb-database-create" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182807 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="5abe7bf3-d44d-4ee5-b568-2d497868f1e5" containerName="placement-api" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182832 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="234a900d-887b-448c-8336-010107726c1e" containerName="mariadb-account-create-update" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.182843 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e0fb1fc-c94a-44f0-a269-e7211c6fcfba" containerName="mariadb-database-create" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.183725 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.193300 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.193493 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.206232 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7fcc47f8dc-lhqhx"] Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.207454 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.209529 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.209964 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.249006 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-86bb565f45-ntq5k"] Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.276341 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7fcc47f8dc-lhqhx"] Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.333753 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-config-data-custom\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.333808 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-combined-ca-bundle\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.333838 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-config-data\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.333865 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-config-data-custom\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.333892 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-config-data\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.333961 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-public-tls-certs\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.333986 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-public-tls-certs\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.334011 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9j4c4\" (UniqueName: \"kubernetes.io/projected/de2e3cc7-c5cb-449a-a19c-2d671f08c656-kube-api-access-9j4c4\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.334032 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd7g7\" (UniqueName: \"kubernetes.io/projected/00b138c6-9e7c-4782-8454-1a4c035b1fbc-kube-api-access-nd7g7\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.334054 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-internal-tls-certs\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.334070 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-internal-tls-certs\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.334087 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-combined-ca-bundle\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.436229 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-config-data\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.436295 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-config-data-custom\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.436358 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-config-data\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.436408 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-public-tls-certs\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.436445 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-public-tls-certs\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.436470 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9j4c4\" (UniqueName: \"kubernetes.io/projected/de2e3cc7-c5cb-449a-a19c-2d671f08c656-kube-api-access-9j4c4\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.436489 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd7g7\" (UniqueName: \"kubernetes.io/projected/00b138c6-9e7c-4782-8454-1a4c035b1fbc-kube-api-access-nd7g7\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.436530 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-internal-tls-certs\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.436548 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-internal-tls-certs\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.436564 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-combined-ca-bundle\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.436640 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-config-data-custom\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.436684 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-combined-ca-bundle\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.459095 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-combined-ca-bundle\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.462418 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9j4c4\" (UniqueName: \"kubernetes.io/projected/de2e3cc7-c5cb-449a-a19c-2d671f08c656-kube-api-access-9j4c4\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.464104 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-config-data\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.465713 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-config-data\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.470145 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-config-data-custom\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.470665 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-config-data-custom\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.471288 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-internal-tls-certs\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.473637 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-internal-tls-certs\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.474467 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-combined-ca-bundle\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.477096 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/de2e3cc7-c5cb-449a-a19c-2d671f08c656-public-tls-certs\") pod \"heat-cfnapi-86bb565f45-ntq5k\" (UID: \"de2e3cc7-c5cb-449a-a19c-2d671f08c656\") " pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.481681 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd7g7\" (UniqueName: \"kubernetes.io/projected/00b138c6-9e7c-4782-8454-1a4c035b1fbc-kube-api-access-nd7g7\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.481876 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/00b138c6-9e7c-4782-8454-1a4c035b1fbc-public-tls-certs\") pod \"heat-api-7fcc47f8dc-lhqhx\" (UID: \"00b138c6-9e7c-4782-8454-1a4c035b1fbc\") " pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.527989 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-c959f64fb-hx4t8" event={"ID":"53145947-4584-4cef-b085-a0e0f550dde9","Type":"ContainerStarted","Data":"4b36cced26c065af3bdca097c2ba77c4f4683335facc1febb5ca7b6e7e41d41b"} Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.530561 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-d856c56c-cmd2q" event={"ID":"5d10747e-ba77-4986-9d4b-636fcbf823ab","Type":"ContainerStarted","Data":"35f6f30aa35f7a79445d6acba6d7d99ce02bc8679e546b9d8ecccf0df51e3ce6"} Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.530736 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-d856c56c-cmd2q" podUID="5d10747e-ba77-4986-9d4b-636fcbf823ab" containerName="heat-cfnapi" containerID="cri-o://35f6f30aa35f7a79445d6acba6d7d99ce02bc8679e546b9d8ecccf0df51e3ce6" gracePeriod=60 Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.531115 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.537454 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-b547848c4-bn5vs" event={"ID":"07914020-653d-4509-9f60-22726224c7c6","Type":"ContainerStarted","Data":"b0d25fd2c9604e3f96cafaee37ffd660ff8a4f27903a0e0bf9e82ba66554c126"} Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.537625 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-b547848c4-bn5vs" podUID="07914020-653d-4509-9f60-22726224c7c6" containerName="heat-api" containerID="cri-o://b0d25fd2c9604e3f96cafaee37ffd660ff8a4f27903a0e0bf9e82ba66554c126" gracePeriod=60 Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.537702 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.541645 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" event={"ID":"8bca285e-17f7-4505-8a25-21f5ee739584","Type":"ContainerStarted","Data":"1293c4aa6c50a69d6aecb56e9f4df43ee392e8db0df66dd79160ca393da72310"} Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.541721 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" event={"ID":"8bca285e-17f7-4505-8a25-21f5ee739584","Type":"ContainerStarted","Data":"17501791bdc7f7056cbbb54c8ba1821e2768aef3ea1d8c030f27232cf3c7d16a"} Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.542821 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.545735 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6f597ccc7c-zgmpr" event={"ID":"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb","Type":"ContainerStarted","Data":"307b8fb6df45672fd631f8323c011769dd77981dd1c4da51a966ca64e7bdf956"} Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.550045 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" event={"ID":"904f04cd-8110-4637-8bb4-67c4b83e189b","Type":"ContainerStarted","Data":"4cc9fd73a35e44ae17915d74f83df931e877bf9d4b7384d1b90a6239d1a72628"} Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.550977 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.554541 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"536490c7-c218-43ca-b601-84fdf0721b13","Type":"ContainerStarted","Data":"0e185e3360ac7d555a53f4a6a5858f9b0a423c695033ffc9d71eb6f71e6ca6e1"} Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.563778 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-d856c56c-cmd2q" podStartSLOduration=11.095062347 podStartE2EDuration="18.563756691s" podCreationTimestamp="2026-03-13 10:27:13 +0000 UTC" firstStartedPulling="2026-03-13 10:27:22.002496455 +0000 UTC m=+1416.025026588" lastFinishedPulling="2026-03-13 10:27:29.471190799 +0000 UTC m=+1423.493720932" observedRunningTime="2026-03-13 10:27:31.551318988 +0000 UTC m=+1425.573849121" watchObservedRunningTime="2026-03-13 10:27:31.563756691 +0000 UTC m=+1425.586286824" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.594357 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" podStartSLOduration=18.594337035 podStartE2EDuration="18.594337035s" podCreationTimestamp="2026-03-13 10:27:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:27:31.587369756 +0000 UTC m=+1425.609899889" watchObservedRunningTime="2026-03-13 10:27:31.594337035 +0000 UTC m=+1425.616867168" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.611237 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-b547848c4-bn5vs" podStartSLOduration=12.22209661 podStartE2EDuration="18.611219476s" podCreationTimestamp="2026-03-13 10:27:13 +0000 UTC" firstStartedPulling="2026-03-13 10:27:23.082037383 +0000 UTC m=+1417.104567516" lastFinishedPulling="2026-03-13 10:27:29.471160249 +0000 UTC m=+1423.493690382" observedRunningTime="2026-03-13 10:27:31.60644589 +0000 UTC m=+1425.628976023" watchObservedRunningTime="2026-03-13 10:27:31.611219476 +0000 UTC m=+1425.633749609" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.824169 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.857329 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" podStartSLOduration=7.857304752 podStartE2EDuration="7.857304752s" podCreationTimestamp="2026-03-13 10:27:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:27:31.743863922 +0000 UTC m=+1425.766394075" watchObservedRunningTime="2026-03-13 10:27:31.857304752 +0000 UTC m=+1425.879834885" Mar 13 10:27:31 crc kubenswrapper[4632]: I0313 10:27:31.879254 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:32 crc kubenswrapper[4632]: I0313 10:27:32.062754 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5abe7bf3-d44d-4ee5-b568-2d497868f1e5" path="/var/lib/kubelet/pods/5abe7bf3-d44d-4ee5-b568-2d497868f1e5/volumes" Mar 13 10:27:32 crc kubenswrapper[4632]: I0313 10:27:32.446075 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-86bb565f45-ntq5k"] Mar 13 10:27:32 crc kubenswrapper[4632]: I0313 10:27:32.554862 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7fcc47f8dc-lhqhx"] Mar 13 10:27:32 crc kubenswrapper[4632]: I0313 10:27:32.604313 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-86bb565f45-ntq5k" event={"ID":"de2e3cc7-c5cb-449a-a19c-2d671f08c656","Type":"ContainerStarted","Data":"7d396e7b6be3f9a3a20f53a205998858ab9fe01adc0c6f2295e912f92fb9fcc5"} Mar 13 10:27:32 crc kubenswrapper[4632]: I0313 10:27:32.632150 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6f597ccc7c-zgmpr" event={"ID":"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb","Type":"ContainerStarted","Data":"2b98409c671c4984548f0269cc4d51793d9e2a7f34b6828d8febd63ffbc09eee"} Mar 13 10:27:32 crc kubenswrapper[4632]: I0313 10:27:32.634015 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:32 crc kubenswrapper[4632]: I0313 10:27:32.670866 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6785ba8c-a47b-4851-945e-c07ccecb9911","Type":"ContainerStarted","Data":"341d5aa1c19947aafc036067af21bc4eee52624758dd74337e755e99f3f5eb7b"} Mar 13 10:27:32 crc kubenswrapper[4632]: I0313 10:27:32.671563 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 13 10:27:32 crc kubenswrapper[4632]: I0313 10:27:32.701119 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"536490c7-c218-43ca-b601-84fdf0721b13","Type":"ContainerStarted","Data":"45f5d86800aa7ead2bd3ca8e9cc3cc79ae2d441610ccb1ee742ca8de3f0990d9"} Mar 13 10:27:32 crc kubenswrapper[4632]: I0313 10:27:32.702609 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-6f597ccc7c-zgmpr" podStartSLOduration=8.702579492 podStartE2EDuration="8.702579492s" podCreationTimestamp="2026-03-13 10:27:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:27:32.673927165 +0000 UTC m=+1426.696457308" watchObservedRunningTime="2026-03-13 10:27:32.702579492 +0000 UTC m=+1426.725109625" Mar 13 10:27:32 crc kubenswrapper[4632]: I0313 10:27:32.722211 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=11.722191559 podStartE2EDuration="11.722191559s" podCreationTimestamp="2026-03-13 10:27:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:27:32.715858825 +0000 UTC m=+1426.738388958" watchObservedRunningTime="2026-03-13 10:27:32.722191559 +0000 UTC m=+1426.744721692" Mar 13 10:27:32 crc kubenswrapper[4632]: I0313 10:27:32.737786 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-c959f64fb-hx4t8" event={"ID":"53145947-4584-4cef-b085-a0e0f550dde9","Type":"ContainerStarted","Data":"4b6b5b347edd33f31b21125267ebc982899aa7660b332796152bdc7475805cd9"} Mar 13 10:27:32 crc kubenswrapper[4632]: I0313 10:27:32.738200 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:32 crc kubenswrapper[4632]: I0313 10:27:32.778232 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-c959f64fb-hx4t8" podStartSLOduration=8.778212602 podStartE2EDuration="8.778212602s" podCreationTimestamp="2026-03-13 10:27:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:27:32.764450107 +0000 UTC m=+1426.786980240" watchObservedRunningTime="2026-03-13 10:27:32.778212602 +0000 UTC m=+1426.800742735" Mar 13 10:27:33 crc kubenswrapper[4632]: I0313 10:27:33.766385 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-86bb565f45-ntq5k" event={"ID":"de2e3cc7-c5cb-449a-a19c-2d671f08c656","Type":"ContainerStarted","Data":"24a4b974cc05b9cc4e7554437e75e3cf003b41e725d29de74b6b204902d42317"} Mar 13 10:27:33 crc kubenswrapper[4632]: I0313 10:27:33.768302 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:33 crc kubenswrapper[4632]: I0313 10:27:33.783565 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7fcc47f8dc-lhqhx" event={"ID":"00b138c6-9e7c-4782-8454-1a4c035b1fbc","Type":"ContainerStarted","Data":"b13fdea311401efe2917beec715abcb0d7c10958cf6ad0d72c84fbe190534cae"} Mar 13 10:27:33 crc kubenswrapper[4632]: I0313 10:27:33.783831 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7fcc47f8dc-lhqhx" event={"ID":"00b138c6-9e7c-4782-8454-1a4c035b1fbc","Type":"ContainerStarted","Data":"4b879070327e10dfbbf89f023602552a89f6b74e9ae41bf4fd6224b99c3ddb61"} Mar 13 10:27:33 crc kubenswrapper[4632]: I0313 10:27:33.784826 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:33 crc kubenswrapper[4632]: I0313 10:27:33.793961 4632 generic.go:334] "Generic (PLEG): container finished" podID="8bca285e-17f7-4505-8a25-21f5ee739584" containerID="1293c4aa6c50a69d6aecb56e9f4df43ee392e8db0df66dd79160ca393da72310" exitCode=1 Mar 13 10:27:33 crc kubenswrapper[4632]: I0313 10:27:33.794071 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" event={"ID":"8bca285e-17f7-4505-8a25-21f5ee739584","Type":"ContainerDied","Data":"1293c4aa6c50a69d6aecb56e9f4df43ee392e8db0df66dd79160ca393da72310"} Mar 13 10:27:33 crc kubenswrapper[4632]: I0313 10:27:33.794894 4632 scope.go:117] "RemoveContainer" containerID="1293c4aa6c50a69d6aecb56e9f4df43ee392e8db0df66dd79160ca393da72310" Mar 13 10:27:33 crc kubenswrapper[4632]: I0313 10:27:33.807311 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-86bb565f45-ntq5k" podStartSLOduration=2.807285153 podStartE2EDuration="2.807285153s" podCreationTimestamp="2026-03-13 10:27:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:27:33.79648559 +0000 UTC m=+1427.819015723" watchObservedRunningTime="2026-03-13 10:27:33.807285153 +0000 UTC m=+1427.829815286" Mar 13 10:27:33 crc kubenswrapper[4632]: I0313 10:27:33.821323 4632 generic.go:334] "Generic (PLEG): container finished" podID="a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" containerID="2b98409c671c4984548f0269cc4d51793d9e2a7f34b6828d8febd63ffbc09eee" exitCode=1 Mar 13 10:27:33 crc kubenswrapper[4632]: I0313 10:27:33.829953 4632 scope.go:117] "RemoveContainer" containerID="2b98409c671c4984548f0269cc4d51793d9e2a7f34b6828d8febd63ffbc09eee" Mar 13 10:27:33 crc kubenswrapper[4632]: I0313 10:27:33.830344 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6f597ccc7c-zgmpr" event={"ID":"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb","Type":"ContainerDied","Data":"2b98409c671c4984548f0269cc4d51793d9e2a7f34b6828d8febd63ffbc09eee"} Mar 13 10:27:33 crc kubenswrapper[4632]: E0313 10:27:33.870865 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda721ddbf_6e3d_4c04_9fd5_52a29a4926cb.slice/crio-conmon-2b98409c671c4984548f0269cc4d51793d9e2a7f34b6828d8febd63ffbc09eee.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda721ddbf_6e3d_4c04_9fd5_52a29a4926cb.slice/crio-2b98409c671c4984548f0269cc4d51793d9e2a7f34b6828d8febd63ffbc09eee.scope\": RecentStats: unable to find data in memory cache]" Mar 13 10:27:33 crc kubenswrapper[4632]: I0313 10:27:33.893006 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-7fcc47f8dc-lhqhx" podStartSLOduration=2.864273088 podStartE2EDuration="2.864273088s" podCreationTimestamp="2026-03-13 10:27:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:27:33.862514856 +0000 UTC m=+1427.885044989" watchObservedRunningTime="2026-03-13 10:27:33.864273088 +0000 UTC m=+1427.886803211" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.573806 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5mlm2"] Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.589748 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.603533 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.603767 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-h4qk2" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.603910 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.631656 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5mlm2"] Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.711189 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-scripts\") pod \"nova-cell0-conductor-db-sync-5mlm2\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.711292 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-config-data\") pod \"nova-cell0-conductor-db-sync-5mlm2\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.711328 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5mlm2\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.711620 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg9bb\" (UniqueName: \"kubernetes.io/projected/5de81924-9bfc-484e-8276-0216f0bbf72c-kube-api-access-bg9bb\") pod \"nova-cell0-conductor-db-sync-5mlm2\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.817379 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg9bb\" (UniqueName: \"kubernetes.io/projected/5de81924-9bfc-484e-8276-0216f0bbf72c-kube-api-access-bg9bb\") pod \"nova-cell0-conductor-db-sync-5mlm2\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.817444 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-scripts\") pod \"nova-cell0-conductor-db-sync-5mlm2\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.817491 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-config-data\") pod \"nova-cell0-conductor-db-sync-5mlm2\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.817516 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5mlm2\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.858313 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-5mlm2\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.859582 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-config-data\") pod \"nova-cell0-conductor-db-sync-5mlm2\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.861399 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg9bb\" (UniqueName: \"kubernetes.io/projected/5de81924-9bfc-484e-8276-0216f0bbf72c-kube-api-access-bg9bb\") pod \"nova-cell0-conductor-db-sync-5mlm2\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.865299 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-scripts\") pod \"nova-cell0-conductor-db-sync-5mlm2\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.879657 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" event={"ID":"8bca285e-17f7-4505-8a25-21f5ee739584","Type":"ContainerStarted","Data":"e13e391c062e12374c23aa9fcea624691b178835192b17850a99a310e07ef571"} Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.881108 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.892087 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6f597ccc7c-zgmpr" event={"ID":"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb","Type":"ContainerStarted","Data":"5149ba55c6bee8420be5c57ba0eca5a40286dd8e8d58bfbcc05f3ba9f2ac9d98"} Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.893047 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.912437 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"536490c7-c218-43ca-b601-84fdf0721b13","Type":"ContainerStarted","Data":"1fedf410b0ec76e58a0488f6518b4a44b2c019d46fe77aa230c5ac262bb32cff"} Mar 13 10:27:34 crc kubenswrapper[4632]: I0313 10:27:34.979569 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:27:35 crc kubenswrapper[4632]: I0313 10:27:35.395218 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 10:27:35 crc kubenswrapper[4632]: I0313 10:27:35.395763 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:27:35 crc kubenswrapper[4632]: I0313 10:27:35.397487 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"c9dfdd84c36e6ac95b45a488b62e176636bdecfbe3a88d3f5d2058d92ebbacdd"} pod="openstack/horizon-7bdb5f7878-ng2k2" containerMessage="Container horizon failed startup probe, will be restarted" Mar 13 10:27:35 crc kubenswrapper[4632]: I0313 10:27:35.397551 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" containerID="cri-o://c9dfdd84c36e6ac95b45a488b62e176636bdecfbe3a88d3f5d2058d92ebbacdd" gracePeriod=30 Mar 13 10:27:35 crc kubenswrapper[4632]: I0313 10:27:35.683803 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5mlm2"] Mar 13 10:27:35 crc kubenswrapper[4632]: I0313 10:27:35.868280 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-689764498d-rg7vt" podUID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 13 10:27:35 crc kubenswrapper[4632]: I0313 10:27:35.944236 4632 generic.go:334] "Generic (PLEG): container finished" podID="8bca285e-17f7-4505-8a25-21f5ee739584" containerID="e13e391c062e12374c23aa9fcea624691b178835192b17850a99a310e07ef571" exitCode=1 Mar 13 10:27:35 crc kubenswrapper[4632]: I0313 10:27:35.944368 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" event={"ID":"8bca285e-17f7-4505-8a25-21f5ee739584","Type":"ContainerDied","Data":"e13e391c062e12374c23aa9fcea624691b178835192b17850a99a310e07ef571"} Mar 13 10:27:35 crc kubenswrapper[4632]: I0313 10:27:35.944413 4632 scope.go:117] "RemoveContainer" containerID="1293c4aa6c50a69d6aecb56e9f4df43ee392e8db0df66dd79160ca393da72310" Mar 13 10:27:35 crc kubenswrapper[4632]: I0313 10:27:35.945116 4632 scope.go:117] "RemoveContainer" containerID="e13e391c062e12374c23aa9fcea624691b178835192b17850a99a310e07ef571" Mar 13 10:27:35 crc kubenswrapper[4632]: E0313 10:27:35.945798 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-66b64f87f7-6z95j_openstack(8bca285e-17f7-4505-8a25-21f5ee739584)\"" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" podUID="8bca285e-17f7-4505-8a25-21f5ee739584" Mar 13 10:27:35 crc kubenswrapper[4632]: I0313 10:27:35.956683 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5mlm2" event={"ID":"5de81924-9bfc-484e-8276-0216f0bbf72c","Type":"ContainerStarted","Data":"951453764e4945cebf39ce493b8227004659c89412dc1b5f0146d76b115b3607"} Mar 13 10:27:35 crc kubenswrapper[4632]: I0313 10:27:35.978141 4632 generic.go:334] "Generic (PLEG): container finished" podID="a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" containerID="5149ba55c6bee8420be5c57ba0eca5a40286dd8e8d58bfbcc05f3ba9f2ac9d98" exitCode=1 Mar 13 10:27:35 crc kubenswrapper[4632]: I0313 10:27:35.978425 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6f597ccc7c-zgmpr" event={"ID":"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb","Type":"ContainerDied","Data":"5149ba55c6bee8420be5c57ba0eca5a40286dd8e8d58bfbcc05f3ba9f2ac9d98"} Mar 13 10:27:35 crc kubenswrapper[4632]: I0313 10:27:35.980888 4632 scope.go:117] "RemoveContainer" containerID="5149ba55c6bee8420be5c57ba0eca5a40286dd8e8d58bfbcc05f3ba9f2ac9d98" Mar 13 10:27:35 crc kubenswrapper[4632]: E0313 10:27:35.981268 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-6f597ccc7c-zgmpr_openstack(a721ddbf-6e3d-4c04-9fd5-52a29a4926cb)\"" pod="openstack/heat-api-6f597ccc7c-zgmpr" podUID="a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.015537 4632 generic.go:334] "Generic (PLEG): container finished" podID="6d73a499-d334-4a7a-9783-640b98760672" containerID="8c839401b1db62da93454588496b8ab534c9e6313aa3bcb0003cb9137b63b2ca" exitCode=0 Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.016677 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c86b4b888-l9574" event={"ID":"6d73a499-d334-4a7a-9783-640b98760672","Type":"ContainerDied","Data":"8c839401b1db62da93454588496b8ab534c9e6313aa3bcb0003cb9137b63b2ca"} Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.128592 4632 scope.go:117] "RemoveContainer" containerID="2b98409c671c4984548f0269cc4d51793d9e2a7f34b6828d8febd63ffbc09eee" Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.619192 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.778535 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-config\") pod \"6d73a499-d334-4a7a-9783-640b98760672\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.778737 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-ovndb-tls-certs\") pod \"6d73a499-d334-4a7a-9783-640b98760672\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.778828 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-httpd-config\") pod \"6d73a499-d334-4a7a-9783-640b98760672\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.778847 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lsp7\" (UniqueName: \"kubernetes.io/projected/6d73a499-d334-4a7a-9783-640b98760672-kube-api-access-6lsp7\") pod \"6d73a499-d334-4a7a-9783-640b98760672\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.778866 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-combined-ca-bundle\") pod \"6d73a499-d334-4a7a-9783-640b98760672\" (UID: \"6d73a499-d334-4a7a-9783-640b98760672\") " Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.811048 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "6d73a499-d334-4a7a-9783-640b98760672" (UID: "6d73a499-d334-4a7a-9783-640b98760672"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.811129 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d73a499-d334-4a7a-9783-640b98760672-kube-api-access-6lsp7" (OuterVolumeSpecName: "kube-api-access-6lsp7") pod "6d73a499-d334-4a7a-9783-640b98760672" (UID: "6d73a499-d334-4a7a-9783-640b98760672"). InnerVolumeSpecName "kube-api-access-6lsp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.861037 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6d73a499-d334-4a7a-9783-640b98760672" (UID: "6d73a499-d334-4a7a-9783-640b98760672"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.883969 4632 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.884009 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lsp7\" (UniqueName: \"kubernetes.io/projected/6d73a499-d334-4a7a-9783-640b98760672-kube-api-access-6lsp7\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.884025 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.903606 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-config" (OuterVolumeSpecName: "config") pod "6d73a499-d334-4a7a-9783-640b98760672" (UID: "6d73a499-d334-4a7a-9783-640b98760672"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.961375 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "6d73a499-d334-4a7a-9783-640b98760672" (UID: "6d73a499-d334-4a7a-9783-640b98760672"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.985552 4632 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:36 crc kubenswrapper[4632]: I0313 10:27:36.985591 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/6d73a499-d334-4a7a-9783-640b98760672-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:37 crc kubenswrapper[4632]: I0313 10:27:37.027897 4632 scope.go:117] "RemoveContainer" containerID="e13e391c062e12374c23aa9fcea624691b178835192b17850a99a310e07ef571" Mar 13 10:27:37 crc kubenswrapper[4632]: E0313 10:27:37.028346 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-66b64f87f7-6z95j_openstack(8bca285e-17f7-4505-8a25-21f5ee739584)\"" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" podUID="8bca285e-17f7-4505-8a25-21f5ee739584" Mar 13 10:27:37 crc kubenswrapper[4632]: I0313 10:27:37.065689 4632 scope.go:117] "RemoveContainer" containerID="5149ba55c6bee8420be5c57ba0eca5a40286dd8e8d58bfbcc05f3ba9f2ac9d98" Mar 13 10:27:37 crc kubenswrapper[4632]: E0313 10:27:37.066161 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-6f597ccc7c-zgmpr_openstack(a721ddbf-6e3d-4c04-9fd5-52a29a4926cb)\"" pod="openstack/heat-api-6f597ccc7c-zgmpr" podUID="a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" Mar 13 10:27:37 crc kubenswrapper[4632]: I0313 10:27:37.073998 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5c86b4b888-l9574" event={"ID":"6d73a499-d334-4a7a-9783-640b98760672","Type":"ContainerDied","Data":"c201c6ed0f734df3747387db31697b083007f33831a2be5b5b4d93d97a61d2c9"} Mar 13 10:27:37 crc kubenswrapper[4632]: I0313 10:27:37.074281 4632 scope.go:117] "RemoveContainer" containerID="e005b4f09b297f1fe00efd39c9534b7382173cd69b88dca5466ba89c0f3c0de7" Mar 13 10:27:37 crc kubenswrapper[4632]: I0313 10:27:37.075561 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5c86b4b888-l9574" Mar 13 10:27:37 crc kubenswrapper[4632]: I0313 10:27:37.112241 4632 scope.go:117] "RemoveContainer" containerID="8c839401b1db62da93454588496b8ab534c9e6313aa3bcb0003cb9137b63b2ca" Mar 13 10:27:37 crc kubenswrapper[4632]: I0313 10:27:37.113463 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"536490c7-c218-43ca-b601-84fdf0721b13","Type":"ContainerStarted","Data":"62a2bf3fb649bf768c02b7ee6d2e17db6f7164a75bcbf4a85047247a011f076e"} Mar 13 10:27:37 crc kubenswrapper[4632]: I0313 10:27:37.114340 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 10:27:37 crc kubenswrapper[4632]: I0313 10:27:37.130638 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"aef9680f-df77-4e2e-ac53-9d7530c2270c","Type":"ContainerStarted","Data":"cc5e594c8a0fc7e3fadb59d676c0ee796f4179df2acec563525b4171416f0e00"} Mar 13 10:27:37 crc kubenswrapper[4632]: I0313 10:27:37.178994 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5c86b4b888-l9574"] Mar 13 10:27:37 crc kubenswrapper[4632]: I0313 10:27:37.190022 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5c86b4b888-l9574"] Mar 13 10:27:37 crc kubenswrapper[4632]: I0313 10:27:37.219015 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=7.904201063 podStartE2EDuration="14.218995108s" podCreationTimestamp="2026-03-13 10:27:23 +0000 UTC" firstStartedPulling="2026-03-13 10:27:29.226365712 +0000 UTC m=+1423.248895845" lastFinishedPulling="2026-03-13 10:27:35.541159757 +0000 UTC m=+1429.563689890" observedRunningTime="2026-03-13 10:27:37.187749958 +0000 UTC m=+1431.210280111" watchObservedRunningTime="2026-03-13 10:27:37.218995108 +0000 UTC m=+1431.241525241" Mar 13 10:27:37 crc kubenswrapper[4632]: I0313 10:27:37.230560 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.865906468 podStartE2EDuration="37.230530689s" podCreationTimestamp="2026-03-13 10:27:00 +0000 UTC" firstStartedPulling="2026-03-13 10:27:01.98182196 +0000 UTC m=+1396.004352093" lastFinishedPulling="2026-03-13 10:27:35.346446191 +0000 UTC m=+1429.368976314" observedRunningTime="2026-03-13 10:27:37.214097879 +0000 UTC m=+1431.236628002" watchObservedRunningTime="2026-03-13 10:27:37.230530689 +0000 UTC m=+1431.253060822" Mar 13 10:27:38 crc kubenswrapper[4632]: I0313 10:27:38.066566 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6d73a499-d334-4a7a-9783-640b98760672" path="/var/lib/kubelet/pods/6d73a499-d334-4a7a-9783-640b98760672/volumes" Mar 13 10:27:38 crc kubenswrapper[4632]: I0313 10:27:38.970130 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:27:39 crc kubenswrapper[4632]: I0313 10:27:39.048658 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f6d96bd7f-txx79"] Mar 13 10:27:39 crc kubenswrapper[4632]: I0313 10:27:39.048901 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" podUID="55712a50-9dcf-44ce-8bac-9aa3ecf65db4" containerName="dnsmasq-dns" containerID="cri-o://fd4487114042316df9fc87c4e68537674ac28dafb4672e7c807d655817ad05cf" gracePeriod=10 Mar 13 10:27:39 crc kubenswrapper[4632]: I0313 10:27:39.213897 4632 generic.go:334] "Generic (PLEG): container finished" podID="55712a50-9dcf-44ce-8bac-9aa3ecf65db4" containerID="fd4487114042316df9fc87c4e68537674ac28dafb4672e7c807d655817ad05cf" exitCode=0 Mar 13 10:27:39 crc kubenswrapper[4632]: I0313 10:27:39.213985 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" event={"ID":"55712a50-9dcf-44ce-8bac-9aa3ecf65db4","Type":"ContainerDied","Data":"fd4487114042316df9fc87c4e68537674ac28dafb4672e7c807d655817ad05cf"} Mar 13 10:27:39 crc kubenswrapper[4632]: I0313 10:27:39.343431 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" podUID="55712a50-9dcf-44ce-8bac-9aa3ecf65db4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.171:5353: connect: connection refused" Mar 13 10:27:39 crc kubenswrapper[4632]: I0313 10:27:39.865754 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:27:39 crc kubenswrapper[4632]: I0313 10:27:39.990403 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-config\") pod \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " Mar 13 10:27:39 crc kubenswrapper[4632]: I0313 10:27:39.990716 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-ovsdbserver-nb\") pod \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " Mar 13 10:27:39 crc kubenswrapper[4632]: I0313 10:27:39.991060 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-dns-svc\") pod \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " Mar 13 10:27:39 crc kubenswrapper[4632]: I0313 10:27:39.991221 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-ovsdbserver-sb\") pod \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " Mar 13 10:27:39 crc kubenswrapper[4632]: I0313 10:27:39.991377 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7sjz\" (UniqueName: \"kubernetes.io/projected/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-kube-api-access-g7sjz\") pod \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " Mar 13 10:27:39 crc kubenswrapper[4632]: I0313 10:27:39.991535 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-dns-swift-storage-0\") pod \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\" (UID: \"55712a50-9dcf-44ce-8bac-9aa3ecf65db4\") " Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.020315 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-kube-api-access-g7sjz" (OuterVolumeSpecName: "kube-api-access-g7sjz") pod "55712a50-9dcf-44ce-8bac-9aa3ecf65db4" (UID: "55712a50-9dcf-44ce-8bac-9aa3ecf65db4"). InnerVolumeSpecName "kube-api-access-g7sjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.113061 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7sjz\" (UniqueName: \"kubernetes.io/projected/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-kube-api-access-g7sjz\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.149487 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-config" (OuterVolumeSpecName: "config") pod "55712a50-9dcf-44ce-8bac-9aa3ecf65db4" (UID: "55712a50-9dcf-44ce-8bac-9aa3ecf65db4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.150024 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "55712a50-9dcf-44ce-8bac-9aa3ecf65db4" (UID: "55712a50-9dcf-44ce-8bac-9aa3ecf65db4"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.188921 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "55712a50-9dcf-44ce-8bac-9aa3ecf65db4" (UID: "55712a50-9dcf-44ce-8bac-9aa3ecf65db4"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.191911 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "55712a50-9dcf-44ce-8bac-9aa3ecf65db4" (UID: "55712a50-9dcf-44ce-8bac-9aa3ecf65db4"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.216869 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.216918 4632 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.216963 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.216981 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.273020 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "55712a50-9dcf-44ce-8bac-9aa3ecf65db4" (UID: "55712a50-9dcf-44ce-8bac-9aa3ecf65db4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.273235 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.321438 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/55712a50-9dcf-44ce-8bac-9aa3ecf65db4-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.339028 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.339066 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.339081 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f6d96bd7f-txx79" event={"ID":"55712a50-9dcf-44ce-8bac-9aa3ecf65db4","Type":"ContainerDied","Data":"749192ea37afcdb5bad8f984bb1339eb6de202d1531a18803ce98189920ca65c"} Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.339119 4632 scope.go:117] "RemoveContainer" containerID="fd4487114042316df9fc87c4e68537674ac28dafb4672e7c807d655817ad05cf" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.340269 4632 scope.go:117] "RemoveContainer" containerID="5149ba55c6bee8420be5c57ba0eca5a40286dd8e8d58bfbcc05f3ba9f2ac9d98" Mar 13 10:27:40 crc kubenswrapper[4632]: E0313 10:27:40.340734 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-6f597ccc7c-zgmpr_openstack(a721ddbf-6e3d-4c04-9fd5-52a29a4926cb)\"" pod="openstack/heat-api-6f597ccc7c-zgmpr" podUID="a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.341188 4632 scope.go:117] "RemoveContainer" containerID="e13e391c062e12374c23aa9fcea624691b178835192b17850a99a310e07ef571" Mar 13 10:27:40 crc kubenswrapper[4632]: E0313 10:27:40.341481 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-66b64f87f7-6z95j_openstack(8bca285e-17f7-4505-8a25-21f5ee739584)\"" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" podUID="8bca285e-17f7-4505-8a25-21f5ee739584" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.388035 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f6d96bd7f-txx79"] Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.388727 4632 scope.go:117] "RemoveContainer" containerID="af3fa8988b343c225a97a2143774e273237597ed7c92bf90057d129267e74a5e" Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.407430 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f6d96bd7f-txx79"] Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.464357 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:27:40 crc kubenswrapper[4632]: I0313 10:27:40.464414 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:27:42 crc kubenswrapper[4632]: I0313 10:27:42.058563 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55712a50-9dcf-44ce-8bac-9aa3ecf65db4" path="/var/lib/kubelet/pods/55712a50-9dcf-44ce-8bac-9aa3ecf65db4/volumes" Mar 13 10:27:42 crc kubenswrapper[4632]: I0313 10:27:42.410232 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="6785ba8c-a47b-4851-945e-c07ccecb9911" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.190:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:27:42 crc kubenswrapper[4632]: I0313 10:27:42.410250 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="6785ba8c-a47b-4851-945e-c07ccecb9911" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.190:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:27:43 crc kubenswrapper[4632]: I0313 10:27:43.917087 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:27:44 crc kubenswrapper[4632]: I0313 10:27:44.108930 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:27:45 crc kubenswrapper[4632]: I0313 10:27:45.028634 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-c959f64fb-hx4t8" Mar 13 10:27:45 crc kubenswrapper[4632]: I0313 10:27:45.097650 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7f9df5b5b5-q6dp2"] Mar 13 10:27:45 crc kubenswrapper[4632]: I0313 10:27:45.097882 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-7f9df5b5b5-q6dp2" podUID="757b852e-068c-4885-99b8-af2e6f23e445" containerName="heat-engine" containerID="cri-o://c974019fb18638d059aa9b080871c3232dbdb322c997ebb8d28de7a80fef50a0" gracePeriod=60 Mar 13 10:27:45 crc kubenswrapper[4632]: I0313 10:27:45.128594 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:27:45 crc kubenswrapper[4632]: I0313 10:27:45.359312 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-7fcc47f8dc-lhqhx" Mar 13 10:27:45 crc kubenswrapper[4632]: I0313 10:27:45.485335 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6f597ccc7c-zgmpr"] Mar 13 10:27:45 crc kubenswrapper[4632]: I0313 10:27:45.866890 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-689764498d-rg7vt" podUID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 13 10:27:45 crc kubenswrapper[4632]: I0313 10:27:45.866994 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:27:45 crc kubenswrapper[4632]: I0313 10:27:45.867966 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"433c9aa5a02161c4bc7228b52cc460020479cbbb899bc6549755a59b8ad796f4"} pod="openstack/horizon-689764498d-rg7vt" containerMessage="Container horizon failed startup probe, will be restarted" Mar 13 10:27:45 crc kubenswrapper[4632]: I0313 10:27:45.868015 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-689764498d-rg7vt" podUID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerName="horizon" containerID="cri-o://433c9aa5a02161c4bc7228b52cc460020479cbbb899bc6549755a59b8ad796f4" gracePeriod=30 Mar 13 10:27:47 crc kubenswrapper[4632]: I0313 10:27:47.418123 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="6785ba8c-a47b-4851-945e-c07ccecb9911" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.190:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:27:47 crc kubenswrapper[4632]: I0313 10:27:47.418258 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="6785ba8c-a47b-4851-945e-c07ccecb9911" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.190:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:27:47 crc kubenswrapper[4632]: I0313 10:27:47.546466 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 13 10:27:47 crc kubenswrapper[4632]: I0313 10:27:47.955477 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-86bb565f45-ntq5k" Mar 13 10:27:48 crc kubenswrapper[4632]: I0313 10:27:48.119479 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-66b64f87f7-6z95j"] Mar 13 10:27:49 crc kubenswrapper[4632]: I0313 10:27:49.508500 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:27:49 crc kubenswrapper[4632]: I0313 10:27:49.509025 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="ceilometer-central-agent" containerID="cri-o://0e185e3360ac7d555a53f4a6a5858f9b0a423c695033ffc9d71eb6f71e6ca6e1" gracePeriod=30 Mar 13 10:27:49 crc kubenswrapper[4632]: I0313 10:27:49.511288 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="sg-core" containerID="cri-o://1fedf410b0ec76e58a0488f6518b4a44b2c019d46fe77aa230c5ac262bb32cff" gracePeriod=30 Mar 13 10:27:49 crc kubenswrapper[4632]: I0313 10:27:49.511411 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="proxy-httpd" containerID="cri-o://62a2bf3fb649bf768c02b7ee6d2e17db6f7164a75bcbf4a85047247a011f076e" gracePeriod=30 Mar 13 10:27:49 crc kubenswrapper[4632]: I0313 10:27:49.511452 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="ceilometer-notification-agent" containerID="cri-o://45f5d86800aa7ead2bd3ca8e9cc3cc79ae2d441610ccb1ee742ca8de3f0990d9" gracePeriod=30 Mar 13 10:27:49 crc kubenswrapper[4632]: I0313 10:27:49.567604 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.191:3000/\": EOF" Mar 13 10:27:50 crc kubenswrapper[4632]: I0313 10:27:50.580421 4632 generic.go:334] "Generic (PLEG): container finished" podID="536490c7-c218-43ca-b601-84fdf0721b13" containerID="62a2bf3fb649bf768c02b7ee6d2e17db6f7164a75bcbf4a85047247a011f076e" exitCode=0 Mar 13 10:27:50 crc kubenswrapper[4632]: I0313 10:27:50.580482 4632 generic.go:334] "Generic (PLEG): container finished" podID="536490c7-c218-43ca-b601-84fdf0721b13" containerID="1fedf410b0ec76e58a0488f6518b4a44b2c019d46fe77aa230c5ac262bb32cff" exitCode=2 Mar 13 10:27:50 crc kubenswrapper[4632]: I0313 10:27:50.580495 4632 generic.go:334] "Generic (PLEG): container finished" podID="536490c7-c218-43ca-b601-84fdf0721b13" containerID="45f5d86800aa7ead2bd3ca8e9cc3cc79ae2d441610ccb1ee742ca8de3f0990d9" exitCode=0 Mar 13 10:27:50 crc kubenswrapper[4632]: I0313 10:27:50.580538 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"536490c7-c218-43ca-b601-84fdf0721b13","Type":"ContainerDied","Data":"62a2bf3fb649bf768c02b7ee6d2e17db6f7164a75bcbf4a85047247a011f076e"} Mar 13 10:27:50 crc kubenswrapper[4632]: I0313 10:27:50.580569 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"536490c7-c218-43ca-b601-84fdf0721b13","Type":"ContainerDied","Data":"1fedf410b0ec76e58a0488f6518b4a44b2c019d46fe77aa230c5ac262bb32cff"} Mar 13 10:27:50 crc kubenswrapper[4632]: I0313 10:27:50.580583 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"536490c7-c218-43ca-b601-84fdf0721b13","Type":"ContainerDied","Data":"45f5d86800aa7ead2bd3ca8e9cc3cc79ae2d441610ccb1ee742ca8de3f0990d9"} Mar 13 10:27:51 crc kubenswrapper[4632]: I0313 10:27:51.594599 4632 generic.go:334] "Generic (PLEG): container finished" podID="536490c7-c218-43ca-b601-84fdf0721b13" containerID="0e185e3360ac7d555a53f4a6a5858f9b0a423c695033ffc9d71eb6f71e6ca6e1" exitCode=0 Mar 13 10:27:51 crc kubenswrapper[4632]: I0313 10:27:51.594824 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"536490c7-c218-43ca-b601-84fdf0721b13","Type":"ContainerDied","Data":"0e185e3360ac7d555a53f4a6a5858f9b0a423c695033ffc9d71eb6f71e6ca6e1"} Mar 13 10:27:52 crc kubenswrapper[4632]: I0313 10:27:52.619481 4632 generic.go:334] "Generic (PLEG): container finished" podID="757b852e-068c-4885-99b8-af2e6f23e445" containerID="c974019fb18638d059aa9b080871c3232dbdb322c997ebb8d28de7a80fef50a0" exitCode=0 Mar 13 10:27:52 crc kubenswrapper[4632]: I0313 10:27:52.619521 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7f9df5b5b5-q6dp2" event={"ID":"757b852e-068c-4885-99b8-af2e6f23e445","Type":"ContainerDied","Data":"c974019fb18638d059aa9b080871c3232dbdb322c997ebb8d28de7a80fef50a0"} Mar 13 10:27:53 crc kubenswrapper[4632]: E0313 10:27:53.836929 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c974019fb18638d059aa9b080871c3232dbdb322c997ebb8d28de7a80fef50a0 is running failed: container process not found" containerID="c974019fb18638d059aa9b080871c3232dbdb322c997ebb8d28de7a80fef50a0" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 10:27:53 crc kubenswrapper[4632]: E0313 10:27:53.868694 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c974019fb18638d059aa9b080871c3232dbdb322c997ebb8d28de7a80fef50a0 is running failed: container process not found" containerID="c974019fb18638d059aa9b080871c3232dbdb322c997ebb8d28de7a80fef50a0" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 10:27:53 crc kubenswrapper[4632]: E0313 10:27:53.869441 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c974019fb18638d059aa9b080871c3232dbdb322c997ebb8d28de7a80fef50a0 is running failed: container process not found" containerID="c974019fb18638d059aa9b080871c3232dbdb322c997ebb8d28de7a80fef50a0" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 10:27:53 crc kubenswrapper[4632]: E0313 10:27:53.869493 4632 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of c974019fb18638d059aa9b080871c3232dbdb322c997ebb8d28de7a80fef50a0 is running failed: container process not found" probeType="Readiness" pod="openstack/heat-engine-7f9df5b5b5-q6dp2" podUID="757b852e-068c-4885-99b8-af2e6f23e445" containerName="heat-engine" Mar 13 10:27:54 crc kubenswrapper[4632]: I0313 10:27:54.314867 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.191:3000/\": dial tcp 10.217.0.191:3000: connect: connection refused" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.360617 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.375885 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.445750 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-combined-ca-bundle\") pod \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.446362 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-config-data-custom\") pod \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.446430 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5qrk\" (UniqueName: \"kubernetes.io/projected/8bca285e-17f7-4505-8a25-21f5ee739584-kube-api-access-w5qrk\") pod \"8bca285e-17f7-4505-8a25-21f5ee739584\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.446607 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vct9r\" (UniqueName: \"kubernetes.io/projected/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-kube-api-access-vct9r\") pod \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.446760 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-combined-ca-bundle\") pod \"8bca285e-17f7-4505-8a25-21f5ee739584\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.446853 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-config-data-custom\") pod \"8bca285e-17f7-4505-8a25-21f5ee739584\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.447018 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-config-data\") pod \"8bca285e-17f7-4505-8a25-21f5ee739584\" (UID: \"8bca285e-17f7-4505-8a25-21f5ee739584\") " Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.447108 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-config-data\") pod \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\" (UID: \"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb\") " Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.488106 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8bca285e-17f7-4505-8a25-21f5ee739584" (UID: "8bca285e-17f7-4505-8a25-21f5ee739584"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.538339 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" (UID: "a721ddbf-6e3d-4c04-9fd5-52a29a4926cb"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.564246 4632 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.564290 4632 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.570999 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-kube-api-access-vct9r" (OuterVolumeSpecName: "kube-api-access-vct9r") pod "a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" (UID: "a721ddbf-6e3d-4c04-9fd5-52a29a4926cb"). InnerVolumeSpecName "kube-api-access-vct9r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.593593 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bca285e-17f7-4505-8a25-21f5ee739584-kube-api-access-w5qrk" (OuterVolumeSpecName: "kube-api-access-w5qrk") pod "8bca285e-17f7-4505-8a25-21f5ee739584" (UID: "8bca285e-17f7-4505-8a25-21f5ee739584"). InnerVolumeSpecName "kube-api-access-w5qrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.708978 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5qrk\" (UniqueName: \"kubernetes.io/projected/8bca285e-17f7-4505-8a25-21f5ee739584-kube-api-access-w5qrk\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.710141 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vct9r\" (UniqueName: \"kubernetes.io/projected/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-kube-api-access-vct9r\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.710508 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" (UID: "a721ddbf-6e3d-4c04-9fd5-52a29a4926cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.725555 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bca285e-17f7-4505-8a25-21f5ee739584" (UID: "8bca285e-17f7-4505-8a25-21f5ee739584"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.740913 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-6f597ccc7c-zgmpr" event={"ID":"a721ddbf-6e3d-4c04-9fd5-52a29a4926cb","Type":"ContainerDied","Data":"307b8fb6df45672fd631f8323c011769dd77981dd1c4da51a966ca64e7bdf956"} Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.741345 4632 scope.go:117] "RemoveContainer" containerID="5149ba55c6bee8420be5c57ba0eca5a40286dd8e8d58bfbcc05f3ba9f2ac9d98" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.741381 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-6f597ccc7c-zgmpr" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.760165 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" event={"ID":"8bca285e-17f7-4505-8a25-21f5ee739584","Type":"ContainerDied","Data":"17501791bdc7f7056cbbb54c8ba1821e2768aef3ea1d8c030f27232cf3c7d16a"} Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.760304 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66b64f87f7-6z95j" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.766483 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-config-data" (OuterVolumeSpecName: "config-data") pod "8bca285e-17f7-4505-8a25-21f5ee739584" (UID: "8bca285e-17f7-4505-8a25-21f5ee739584"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.783628 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-config-data" (OuterVolumeSpecName: "config-data") pod "a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" (UID: "a721ddbf-6e3d-4c04-9fd5-52a29a4926cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.812822 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.813478 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.813601 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bca285e-17f7-4505-8a25-21f5ee739584-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:58 crc kubenswrapper[4632]: I0313 10:27:58.813703 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:27:59 crc kubenswrapper[4632]: I0313 10:27:59.176592 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-66b64f87f7-6z95j"] Mar 13 10:27:59 crc kubenswrapper[4632]: I0313 10:27:59.219210 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-66b64f87f7-6z95j"] Mar 13 10:27:59 crc kubenswrapper[4632]: I0313 10:27:59.243019 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-6f597ccc7c-zgmpr"] Mar 13 10:27:59 crc kubenswrapper[4632]: I0313 10:27:59.256438 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-6f597ccc7c-zgmpr"] Mar 13 10:27:59 crc kubenswrapper[4632]: E0313 10:27:59.805080 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-nova-conductor:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:27:59 crc kubenswrapper[4632]: E0313 10:27:59.805149 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-antelope-centos9/openstack-nova-conductor:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:27:59 crc kubenswrapper[4632]: E0313 10:27:59.805296 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:nova-cell0-conductor-db-sync,Image:quay.rdoproject.org/podified-antelope-centos9/openstack-nova-conductor:e43235cb19da04699a53f42b6a75afe9,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CELL_NAME,Value:cell0,ValueFrom:nil,},EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:false,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/kolla/config_files/config.json,SubPath:nova-conductor-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bg9bb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42436,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-cell0-conductor-db-sync-5mlm2_openstack(5de81924-9bfc-484e-8276-0216f0bbf72c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:27:59 crc kubenswrapper[4632]: E0313 10:27:59.809761 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/nova-cell0-conductor-db-sync-5mlm2" podUID="5de81924-9bfc-484e-8276-0216f0bbf72c" Mar 13 10:27:59 crc kubenswrapper[4632]: I0313 10:27:59.830718 4632 scope.go:117] "RemoveContainer" containerID="e13e391c062e12374c23aa9fcea624691b178835192b17850a99a310e07ef571" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.055442 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bca285e-17f7-4505-8a25-21f5ee739584" path="/var/lib/kubelet/pods/8bca285e-17f7-4505-8a25-21f5ee739584/volumes" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.056609 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" path="/var/lib/kubelet/pods/a721ddbf-6e3d-4c04-9fd5-52a29a4926cb/volumes" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.274608 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556628-479rr"] Mar 13 10:28:00 crc kubenswrapper[4632]: E0313 10:28:00.275471 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55712a50-9dcf-44ce-8bac-9aa3ecf65db4" containerName="init" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.275489 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="55712a50-9dcf-44ce-8bac-9aa3ecf65db4" containerName="init" Mar 13 10:28:00 crc kubenswrapper[4632]: E0313 10:28:00.275506 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55712a50-9dcf-44ce-8bac-9aa3ecf65db4" containerName="dnsmasq-dns" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.275516 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="55712a50-9dcf-44ce-8bac-9aa3ecf65db4" containerName="dnsmasq-dns" Mar 13 10:28:00 crc kubenswrapper[4632]: E0313 10:28:00.275535 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bca285e-17f7-4505-8a25-21f5ee739584" containerName="heat-cfnapi" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.275544 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bca285e-17f7-4505-8a25-21f5ee739584" containerName="heat-cfnapi" Mar 13 10:28:00 crc kubenswrapper[4632]: E0313 10:28:00.275554 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d73a499-d334-4a7a-9783-640b98760672" containerName="neutron-httpd" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.275561 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d73a499-d334-4a7a-9783-640b98760672" containerName="neutron-httpd" Mar 13 10:28:00 crc kubenswrapper[4632]: E0313 10:28:00.275584 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" containerName="heat-api" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.275591 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" containerName="heat-api" Mar 13 10:28:00 crc kubenswrapper[4632]: E0313 10:28:00.275614 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d73a499-d334-4a7a-9783-640b98760672" containerName="neutron-api" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.275624 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d73a499-d334-4a7a-9783-640b98760672" containerName="neutron-api" Mar 13 10:28:00 crc kubenswrapper[4632]: E0313 10:28:00.275633 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" containerName="heat-api" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.275640 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" containerName="heat-api" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.275874 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d73a499-d334-4a7a-9783-640b98760672" containerName="neutron-api" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.275886 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" containerName="heat-api" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.275896 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="a721ddbf-6e3d-4c04-9fd5-52a29a4926cb" containerName="heat-api" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.275914 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bca285e-17f7-4505-8a25-21f5ee739584" containerName="heat-cfnapi" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.275927 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bca285e-17f7-4505-8a25-21f5ee739584" containerName="heat-cfnapi" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.275985 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="55712a50-9dcf-44ce-8bac-9aa3ecf65db4" containerName="dnsmasq-dns" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.276006 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d73a499-d334-4a7a-9783-640b98760672" containerName="neutron-httpd" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.277056 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556628-479rr" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.283875 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.284113 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.284326 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.379139 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvxks\" (UniqueName: \"kubernetes.io/projected/658f9ba3-69b7-4d2d-8258-bb7bdf272398-kube-api-access-nvxks\") pod \"auto-csr-approver-29556628-479rr\" (UID: \"658f9ba3-69b7-4d2d-8258-bb7bdf272398\") " pod="openshift-infra/auto-csr-approver-29556628-479rr" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.392350 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556628-479rr"] Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.482105 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvxks\" (UniqueName: \"kubernetes.io/projected/658f9ba3-69b7-4d2d-8258-bb7bdf272398-kube-api-access-nvxks\") pod \"auto-csr-approver-29556628-479rr\" (UID: \"658f9ba3-69b7-4d2d-8258-bb7bdf272398\") " pod="openshift-infra/auto-csr-approver-29556628-479rr" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.561634 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvxks\" (UniqueName: \"kubernetes.io/projected/658f9ba3-69b7-4d2d-8258-bb7bdf272398-kube-api-access-nvxks\") pod \"auto-csr-approver-29556628-479rr\" (UID: \"658f9ba3-69b7-4d2d-8258-bb7bdf272398\") " pod="openshift-infra/auto-csr-approver-29556628-479rr" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.613916 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556628-479rr" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.789486 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.792084 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.888875 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7f9df5b5b5-q6dp2" event={"ID":"757b852e-068c-4885-99b8-af2e6f23e445","Type":"ContainerDied","Data":"db0a88b20ef1358b7cfb558aebb52cdeba5b5f143eee06ddc98fa0acfb3ab01b"} Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.888968 4632 scope.go:117] "RemoveContainer" containerID="c974019fb18638d059aa9b080871c3232dbdb322c997ebb8d28de7a80fef50a0" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.889106 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7f9df5b5b5-q6dp2" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.894123 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-combined-ca-bundle\") pod \"536490c7-c218-43ca-b601-84fdf0721b13\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.894174 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-config-data-custom\") pod \"757b852e-068c-4885-99b8-af2e6f23e445\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.894239 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/536490c7-c218-43ca-b601-84fdf0721b13-log-httpd\") pod \"536490c7-c218-43ca-b601-84fdf0721b13\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.894267 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/536490c7-c218-43ca-b601-84fdf0721b13-run-httpd\") pod \"536490c7-c218-43ca-b601-84fdf0721b13\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.894374 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-crtzk\" (UniqueName: \"kubernetes.io/projected/536490c7-c218-43ca-b601-84fdf0721b13-kube-api-access-crtzk\") pod \"536490c7-c218-43ca-b601-84fdf0721b13\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.894430 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-combined-ca-bundle\") pod \"757b852e-068c-4885-99b8-af2e6f23e445\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.894478 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-scripts\") pod \"536490c7-c218-43ca-b601-84fdf0721b13\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.894513 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-config-data\") pod \"536490c7-c218-43ca-b601-84fdf0721b13\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.894528 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5895\" (UniqueName: \"kubernetes.io/projected/757b852e-068c-4885-99b8-af2e6f23e445-kube-api-access-d5895\") pod \"757b852e-068c-4885-99b8-af2e6f23e445\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.894577 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-config-data\") pod \"757b852e-068c-4885-99b8-af2e6f23e445\" (UID: \"757b852e-068c-4885-99b8-af2e6f23e445\") " Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.894605 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-sg-core-conf-yaml\") pod \"536490c7-c218-43ca-b601-84fdf0721b13\" (UID: \"536490c7-c218-43ca-b601-84fdf0721b13\") " Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.919786 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/757b852e-068c-4885-99b8-af2e6f23e445-kube-api-access-d5895" (OuterVolumeSpecName: "kube-api-access-d5895") pod "757b852e-068c-4885-99b8-af2e6f23e445" (UID: "757b852e-068c-4885-99b8-af2e6f23e445"). InnerVolumeSpecName "kube-api-access-d5895". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.921316 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/536490c7-c218-43ca-b601-84fdf0721b13-kube-api-access-crtzk" (OuterVolumeSpecName: "kube-api-access-crtzk") pod "536490c7-c218-43ca-b601-84fdf0721b13" (UID: "536490c7-c218-43ca-b601-84fdf0721b13"). InnerVolumeSpecName "kube-api-access-crtzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.921758 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/536490c7-c218-43ca-b601-84fdf0721b13-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "536490c7-c218-43ca-b601-84fdf0721b13" (UID: "536490c7-c218-43ca-b601-84fdf0721b13"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.935264 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/536490c7-c218-43ca-b601-84fdf0721b13-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "536490c7-c218-43ca-b601-84fdf0721b13" (UID: "536490c7-c218-43ca-b601-84fdf0721b13"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.955468 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"536490c7-c218-43ca-b601-84fdf0721b13","Type":"ContainerDied","Data":"f402196243381c3faf3165d4fe49b7c43a1af16813bae58fca9b53eb4badf807"} Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.955911 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.961593 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "757b852e-068c-4885-99b8-af2e6f23e445" (UID: "757b852e-068c-4885-99b8-af2e6f23e445"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.963385 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-scripts" (OuterVolumeSpecName: "scripts") pod "536490c7-c218-43ca-b601-84fdf0721b13" (UID: "536490c7-c218-43ca-b601-84fdf0721b13"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.964351 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "757b852e-068c-4885-99b8-af2e6f23e445" (UID: "757b852e-068c-4885-99b8-af2e6f23e445"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:00 crc kubenswrapper[4632]: I0313 10:28:00.987335 4632 scope.go:117] "RemoveContainer" containerID="62a2bf3fb649bf768c02b7ee6d2e17db6f7164a75bcbf4a85047247a011f076e" Mar 13 10:28:00 crc kubenswrapper[4632]: E0313 10:28:00.987353 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"nova-cell0-conductor-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-antelope-centos9/openstack-nova-conductor:e43235cb19da04699a53f42b6a75afe9\\\"\"" pod="openstack/nova-cell0-conductor-db-sync-5mlm2" podUID="5de81924-9bfc-484e-8276-0216f0bbf72c" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:00.998822 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-crtzk\" (UniqueName: \"kubernetes.io/projected/536490c7-c218-43ca-b601-84fdf0721b13-kube-api-access-crtzk\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:00.998856 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:00.998869 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:00.998883 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5895\" (UniqueName: \"kubernetes.io/projected/757b852e-068c-4885-99b8-af2e6f23e445-kube-api-access-d5895\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:00.998896 4632 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:00.998908 4632 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/536490c7-c218-43ca-b601-84fdf0721b13-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:00.998920 4632 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/536490c7-c218-43ca-b601-84fdf0721b13-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.020509 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "536490c7-c218-43ca-b601-84fdf0721b13" (UID: "536490c7-c218-43ca-b601-84fdf0721b13"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.082101 4632 scope.go:117] "RemoveContainer" containerID="1fedf410b0ec76e58a0488f6518b4a44b2c019d46fe77aa230c5ac262bb32cff" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.084777 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-config-data" (OuterVolumeSpecName: "config-data") pod "757b852e-068c-4885-99b8-af2e6f23e445" (UID: "757b852e-068c-4885-99b8-af2e6f23e445"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.101372 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/757b852e-068c-4885-99b8-af2e6f23e445-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.101417 4632 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.129970 4632 scope.go:117] "RemoveContainer" containerID="45f5d86800aa7ead2bd3ca8e9cc3cc79ae2d441610ccb1ee742ca8de3f0990d9" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.185699 4632 scope.go:117] "RemoveContainer" containerID="0e185e3360ac7d555a53f4a6a5858f9b0a423c695033ffc9d71eb6f71e6ca6e1" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.268354 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7f9df5b5b5-q6dp2"] Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.278473 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-7f9df5b5b5-q6dp2"] Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.287698 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-config-data" (OuterVolumeSpecName: "config-data") pod "536490c7-c218-43ca-b601-84fdf0721b13" (UID: "536490c7-c218-43ca-b601-84fdf0721b13"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.304263 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "536490c7-c218-43ca-b601-84fdf0721b13" (UID: "536490c7-c218-43ca-b601-84fdf0721b13"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.318368 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.318412 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/536490c7-c218-43ca-b601-84fdf0721b13-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.431097 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556628-479rr"] Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.600529 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.610875 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.648577 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:01 crc kubenswrapper[4632]: E0313 10:28:01.649265 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="proxy-httpd" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.649292 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="proxy-httpd" Mar 13 10:28:01 crc kubenswrapper[4632]: E0313 10:28:01.649310 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="sg-core" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.649318 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="sg-core" Mar 13 10:28:01 crc kubenswrapper[4632]: E0313 10:28:01.649331 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bca285e-17f7-4505-8a25-21f5ee739584" containerName="heat-cfnapi" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.649339 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bca285e-17f7-4505-8a25-21f5ee739584" containerName="heat-cfnapi" Mar 13 10:28:01 crc kubenswrapper[4632]: E0313 10:28:01.649359 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="ceilometer-notification-agent" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.649365 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="ceilometer-notification-agent" Mar 13 10:28:01 crc kubenswrapper[4632]: E0313 10:28:01.649403 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="ceilometer-central-agent" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.649414 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="ceilometer-central-agent" Mar 13 10:28:01 crc kubenswrapper[4632]: E0313 10:28:01.649427 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="757b852e-068c-4885-99b8-af2e6f23e445" containerName="heat-engine" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.649435 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="757b852e-068c-4885-99b8-af2e6f23e445" containerName="heat-engine" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.649686 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="ceilometer-central-agent" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.649714 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="ceilometer-notification-agent" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.649725 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="sg-core" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.649737 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="757b852e-068c-4885-99b8-af2e6f23e445" containerName="heat-engine" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.649747 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="536490c7-c218-43ca-b601-84fdf0721b13" containerName="proxy-httpd" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.652101 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.660149 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.660591 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.677134 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.770010 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:01 crc kubenswrapper[4632]: E0313 10:28:01.770968 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-wsdt8 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="ae5601e7-98bb-4c10-bd02-b365269a60e5" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.833269 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae5601e7-98bb-4c10-bd02-b365269a60e5-log-httpd\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.833335 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsdt8\" (UniqueName: \"kubernetes.io/projected/ae5601e7-98bb-4c10-bd02-b365269a60e5-kube-api-access-wsdt8\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.833428 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-config-data\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.833475 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae5601e7-98bb-4c10-bd02-b365269a60e5-run-httpd\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.833495 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-scripts\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.833654 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.833710 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.935992 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-config-data\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.936120 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae5601e7-98bb-4c10-bd02-b365269a60e5-run-httpd\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.936149 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-scripts\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.936231 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.936282 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.936334 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae5601e7-98bb-4c10-bd02-b365269a60e5-log-httpd\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.936369 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsdt8\" (UniqueName: \"kubernetes.io/projected/ae5601e7-98bb-4c10-bd02-b365269a60e5-kube-api-access-wsdt8\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.937312 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae5601e7-98bb-4c10-bd02-b365269a60e5-run-httpd\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.937619 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae5601e7-98bb-4c10-bd02-b365269a60e5-log-httpd\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.949898 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-config-data\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.951782 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-scripts\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.951779 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.951954 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.963230 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsdt8\" (UniqueName: \"kubernetes.io/projected/ae5601e7-98bb-4c10-bd02-b365269a60e5-kube-api-access-wsdt8\") pod \"ceilometer-0\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " pod="openstack/ceilometer-0" Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.977833 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556628-479rr" event={"ID":"658f9ba3-69b7-4d2d-8258-bb7bdf272398","Type":"ContainerStarted","Data":"bd4fb0467d5e125fd924fc111e74635f851cb7c2c04439ff732217c17ddbb8dc"} Mar 13 10:28:01 crc kubenswrapper[4632]: I0313 10:28:01.979207 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.061162 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="536490c7-c218-43ca-b601-84fdf0721b13" path="/var/lib/kubelet/pods/536490c7-c218-43ca-b601-84fdf0721b13/volumes" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.064169 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="757b852e-068c-4885-99b8-af2e6f23e445" path="/var/lib/kubelet/pods/757b852e-068c-4885-99b8-af2e6f23e445/volumes" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.072049 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.241961 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-config-data\") pod \"ae5601e7-98bb-4c10-bd02-b365269a60e5\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.242064 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-sg-core-conf-yaml\") pod \"ae5601e7-98bb-4c10-bd02-b365269a60e5\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.242112 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae5601e7-98bb-4c10-bd02-b365269a60e5-run-httpd\") pod \"ae5601e7-98bb-4c10-bd02-b365269a60e5\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.242164 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-combined-ca-bundle\") pod \"ae5601e7-98bb-4c10-bd02-b365269a60e5\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.242198 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsdt8\" (UniqueName: \"kubernetes.io/projected/ae5601e7-98bb-4c10-bd02-b365269a60e5-kube-api-access-wsdt8\") pod \"ae5601e7-98bb-4c10-bd02-b365269a60e5\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.242240 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-scripts\") pod \"ae5601e7-98bb-4c10-bd02-b365269a60e5\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.242265 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae5601e7-98bb-4c10-bd02-b365269a60e5-log-httpd\") pod \"ae5601e7-98bb-4c10-bd02-b365269a60e5\" (UID: \"ae5601e7-98bb-4c10-bd02-b365269a60e5\") " Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.243035 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae5601e7-98bb-4c10-bd02-b365269a60e5-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ae5601e7-98bb-4c10-bd02-b365269a60e5" (UID: "ae5601e7-98bb-4c10-bd02-b365269a60e5"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.243055 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae5601e7-98bb-4c10-bd02-b365269a60e5-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ae5601e7-98bb-4c10-bd02-b365269a60e5" (UID: "ae5601e7-98bb-4c10-bd02-b365269a60e5"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.245328 4632 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae5601e7-98bb-4c10-bd02-b365269a60e5-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.245365 4632 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae5601e7-98bb-4c10-bd02-b365269a60e5-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.249064 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-config-data" (OuterVolumeSpecName: "config-data") pod "ae5601e7-98bb-4c10-bd02-b365269a60e5" (UID: "ae5601e7-98bb-4c10-bd02-b365269a60e5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.249602 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae5601e7-98bb-4c10-bd02-b365269a60e5" (UID: "ae5601e7-98bb-4c10-bd02-b365269a60e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.253212 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ae5601e7-98bb-4c10-bd02-b365269a60e5" (UID: "ae5601e7-98bb-4c10-bd02-b365269a60e5"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.255410 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae5601e7-98bb-4c10-bd02-b365269a60e5-kube-api-access-wsdt8" (OuterVolumeSpecName: "kube-api-access-wsdt8") pod "ae5601e7-98bb-4c10-bd02-b365269a60e5" (UID: "ae5601e7-98bb-4c10-bd02-b365269a60e5"). InnerVolumeSpecName "kube-api-access-wsdt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.269111 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-scripts" (OuterVolumeSpecName: "scripts") pod "ae5601e7-98bb-4c10-bd02-b365269a60e5" (UID: "ae5601e7-98bb-4c10-bd02-b365269a60e5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.347146 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.347191 4632 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.347210 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.347224 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsdt8\" (UniqueName: \"kubernetes.io/projected/ae5601e7-98bb-4c10-bd02-b365269a60e5-kube-api-access-wsdt8\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.347237 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae5601e7-98bb-4c10-bd02-b365269a60e5-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:02 crc kubenswrapper[4632]: I0313 10:28:02.991252 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.077296 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.100590 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.112684 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.116976 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.128340 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.133744 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.133987 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.264012 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-scripts\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.264099 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.264141 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.264262 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a89045f-ad86-47f4-9837-ccae12089508-log-httpd\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.264369 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a89045f-ad86-47f4-9837-ccae12089508-run-httpd\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.264404 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vmhd\" (UniqueName: \"kubernetes.io/projected/8a89045f-ad86-47f4-9837-ccae12089508-kube-api-access-8vmhd\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.265485 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-config-data\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.367404 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-config-data\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.367494 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-scripts\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.367530 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.367555 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.367572 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a89045f-ad86-47f4-9837-ccae12089508-log-httpd\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.367606 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a89045f-ad86-47f4-9837-ccae12089508-run-httpd\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.367621 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vmhd\" (UniqueName: \"kubernetes.io/projected/8a89045f-ad86-47f4-9837-ccae12089508-kube-api-access-8vmhd\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.372365 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a89045f-ad86-47f4-9837-ccae12089508-log-httpd\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.372610 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a89045f-ad86-47f4-9837-ccae12089508-run-httpd\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.375053 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-scripts\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.375304 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.375333 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.389492 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-config-data\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.390491 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vmhd\" (UniqueName: \"kubernetes.io/projected/8a89045f-ad86-47f4-9837-ccae12089508-kube-api-access-8vmhd\") pod \"ceilometer-0\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.461397 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:03 crc kubenswrapper[4632]: I0313 10:28:03.983560 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:04 crc kubenswrapper[4632]: I0313 10:28:04.008469 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556628-479rr" event={"ID":"658f9ba3-69b7-4d2d-8258-bb7bdf272398","Type":"ContainerStarted","Data":"dccd7606dfc8be32af7f5d6d0a4bf2a63f79937bfd68d93b573f727a7eb9e402"} Mar 13 10:28:04 crc kubenswrapper[4632]: I0313 10:28:04.035479 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556628-479rr" podStartSLOduration=2.777339145 podStartE2EDuration="4.035457336s" podCreationTimestamp="2026-03-13 10:28:00 +0000 UTC" firstStartedPulling="2026-03-13 10:28:01.39823517 +0000 UTC m=+1455.420765303" lastFinishedPulling="2026-03-13 10:28:02.656353361 +0000 UTC m=+1456.678883494" observedRunningTime="2026-03-13 10:28:04.0281987 +0000 UTC m=+1458.050728833" watchObservedRunningTime="2026-03-13 10:28:04.035457336 +0000 UTC m=+1458.057987469" Mar 13 10:28:04 crc kubenswrapper[4632]: I0313 10:28:04.068118 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae5601e7-98bb-4c10-bd02-b365269a60e5" path="/var/lib/kubelet/pods/ae5601e7-98bb-4c10-bd02-b365269a60e5/volumes" Mar 13 10:28:05 crc kubenswrapper[4632]: I0313 10:28:05.112046 4632 generic.go:334] "Generic (PLEG): container finished" podID="658f9ba3-69b7-4d2d-8258-bb7bdf272398" containerID="dccd7606dfc8be32af7f5d6d0a4bf2a63f79937bfd68d93b573f727a7eb9e402" exitCode=0 Mar 13 10:28:05 crc kubenswrapper[4632]: I0313 10:28:05.112416 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556628-479rr" event={"ID":"658f9ba3-69b7-4d2d-8258-bb7bdf272398","Type":"ContainerDied","Data":"dccd7606dfc8be32af7f5d6d0a4bf2a63f79937bfd68d93b573f727a7eb9e402"} Mar 13 10:28:05 crc kubenswrapper[4632]: I0313 10:28:05.134824 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a89045f-ad86-47f4-9837-ccae12089508","Type":"ContainerStarted","Data":"20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874"} Mar 13 10:28:05 crc kubenswrapper[4632]: I0313 10:28:05.134875 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a89045f-ad86-47f4-9837-ccae12089508","Type":"ContainerStarted","Data":"9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0"} Mar 13 10:28:05 crc kubenswrapper[4632]: I0313 10:28:05.134886 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a89045f-ad86-47f4-9837-ccae12089508","Type":"ContainerStarted","Data":"f18f1919b2952189a9f9202d1451d5a08fb72b179ba9686af3513e79b4500446"} Mar 13 10:28:06 crc kubenswrapper[4632]: I0313 10:28:06.149428 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a89045f-ad86-47f4-9837-ccae12089508","Type":"ContainerStarted","Data":"a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688"} Mar 13 10:28:06 crc kubenswrapper[4632]: I0313 10:28:06.157261 4632 generic.go:334] "Generic (PLEG): container finished" podID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerID="c9dfdd84c36e6ac95b45a488b62e176636bdecfbe3a88d3f5d2058d92ebbacdd" exitCode=137 Mar 13 10:28:06 crc kubenswrapper[4632]: I0313 10:28:06.157341 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb5f7878-ng2k2" event={"ID":"3e4afb91-ce26-4325-89c9-2542da2ec48a","Type":"ContainerDied","Data":"c9dfdd84c36e6ac95b45a488b62e176636bdecfbe3a88d3f5d2058d92ebbacdd"} Mar 13 10:28:06 crc kubenswrapper[4632]: I0313 10:28:06.157419 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb5f7878-ng2k2" event={"ID":"3e4afb91-ce26-4325-89c9-2542da2ec48a","Type":"ContainerStarted","Data":"2dbb3ede37abc9f5b483ae48b13ac3ed8913ac4529c34c39494b1541e21ce00b"} Mar 13 10:28:06 crc kubenswrapper[4632]: I0313 10:28:06.157445 4632 scope.go:117] "RemoveContainer" containerID="dc4a058f6feb7822333693352f32f5677ff03988b7b5b71005c85c4bf733b402" Mar 13 10:28:06 crc kubenswrapper[4632]: I0313 10:28:06.611093 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556628-479rr" Mar 13 10:28:06 crc kubenswrapper[4632]: I0313 10:28:06.763050 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvxks\" (UniqueName: \"kubernetes.io/projected/658f9ba3-69b7-4d2d-8258-bb7bdf272398-kube-api-access-nvxks\") pod \"658f9ba3-69b7-4d2d-8258-bb7bdf272398\" (UID: \"658f9ba3-69b7-4d2d-8258-bb7bdf272398\") " Mar 13 10:28:06 crc kubenswrapper[4632]: I0313 10:28:06.771179 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/658f9ba3-69b7-4d2d-8258-bb7bdf272398-kube-api-access-nvxks" (OuterVolumeSpecName: "kube-api-access-nvxks") pod "658f9ba3-69b7-4d2d-8258-bb7bdf272398" (UID: "658f9ba3-69b7-4d2d-8258-bb7bdf272398"). InnerVolumeSpecName "kube-api-access-nvxks". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:28:06 crc kubenswrapper[4632]: I0313 10:28:06.866127 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvxks\" (UniqueName: \"kubernetes.io/projected/658f9ba3-69b7-4d2d-8258-bb7bdf272398-kube-api-access-nvxks\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:07 crc kubenswrapper[4632]: I0313 10:28:07.168565 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556622-7428t"] Mar 13 10:28:07 crc kubenswrapper[4632]: I0313 10:28:07.179439 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556628-479rr" event={"ID":"658f9ba3-69b7-4d2d-8258-bb7bdf272398","Type":"ContainerDied","Data":"bd4fb0467d5e125fd924fc111e74635f851cb7c2c04439ff732217c17ddbb8dc"} Mar 13 10:28:07 crc kubenswrapper[4632]: I0313 10:28:07.179485 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bd4fb0467d5e125fd924fc111e74635f851cb7c2c04439ff732217c17ddbb8dc" Mar 13 10:28:07 crc kubenswrapper[4632]: I0313 10:28:07.179664 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556628-479rr" Mar 13 10:28:07 crc kubenswrapper[4632]: I0313 10:28:07.224579 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556622-7428t"] Mar 13 10:28:08 crc kubenswrapper[4632]: I0313 10:28:08.069069 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bedc1d17-f5c4-4a62-ab0c-f20a002e859b" path="/var/lib/kubelet/pods/bedc1d17-f5c4-4a62-ab0c-f20a002e859b/volumes" Mar 13 10:28:08 crc kubenswrapper[4632]: I0313 10:28:08.192835 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a89045f-ad86-47f4-9837-ccae12089508","Type":"ContainerStarted","Data":"33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf"} Mar 13 10:28:08 crc kubenswrapper[4632]: I0313 10:28:08.194147 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 10:28:08 crc kubenswrapper[4632]: I0313 10:28:08.219968 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.6769290909999999 podStartE2EDuration="5.219931259s" podCreationTimestamp="2026-03-13 10:28:03 +0000 UTC" firstStartedPulling="2026-03-13 10:28:04.007353453 +0000 UTC m=+1458.029883586" lastFinishedPulling="2026-03-13 10:28:07.550355621 +0000 UTC m=+1461.572885754" observedRunningTime="2026-03-13 10:28:08.21341748 +0000 UTC m=+1462.235947613" watchObservedRunningTime="2026-03-13 10:28:08.219931259 +0000 UTC m=+1462.242461412" Mar 13 10:28:10 crc kubenswrapper[4632]: I0313 10:28:10.460553 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:28:10 crc kubenswrapper[4632]: I0313 10:28:10.461187 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:28:10 crc kubenswrapper[4632]: I0313 10:28:10.462150 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:28:10 crc kubenswrapper[4632]: I0313 10:28:10.463051 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a148dfa9ef48de458189e9fda19ce88937bedd25c3ec76e22d14f43a4745805f"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:28:10 crc kubenswrapper[4632]: I0313 10:28:10.463129 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://a148dfa9ef48de458189e9fda19ce88937bedd25c3ec76e22d14f43a4745805f" gracePeriod=600 Mar 13 10:28:11 crc kubenswrapper[4632]: I0313 10:28:11.224601 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="a148dfa9ef48de458189e9fda19ce88937bedd25c3ec76e22d14f43a4745805f" exitCode=0 Mar 13 10:28:11 crc kubenswrapper[4632]: I0313 10:28:11.224690 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"a148dfa9ef48de458189e9fda19ce88937bedd25c3ec76e22d14f43a4745805f"} Mar 13 10:28:11 crc kubenswrapper[4632]: I0313 10:28:11.225424 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f"} Mar 13 10:28:11 crc kubenswrapper[4632]: I0313 10:28:11.225484 4632 scope.go:117] "RemoveContainer" containerID="e9a22f93dffae95945f5e47a3d15b0ebe11dc6b72712dcbe34fa0191ff687b27" Mar 13 10:28:14 crc kubenswrapper[4632]: I0313 10:28:14.835555 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:14 crc kubenswrapper[4632]: I0313 10:28:14.836212 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="ceilometer-central-agent" containerID="cri-o://9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0" gracePeriod=30 Mar 13 10:28:14 crc kubenswrapper[4632]: I0313 10:28:14.836360 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="ceilometer-notification-agent" containerID="cri-o://20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874" gracePeriod=30 Mar 13 10:28:14 crc kubenswrapper[4632]: I0313 10:28:14.836432 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="sg-core" containerID="cri-o://a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688" gracePeriod=30 Mar 13 10:28:14 crc kubenswrapper[4632]: I0313 10:28:14.836542 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="proxy-httpd" containerID="cri-o://33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf" gracePeriod=30 Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.294160 4632 generic.go:334] "Generic (PLEG): container finished" podID="8a89045f-ad86-47f4-9837-ccae12089508" containerID="33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf" exitCode=0 Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.294527 4632 generic.go:334] "Generic (PLEG): container finished" podID="8a89045f-ad86-47f4-9837-ccae12089508" containerID="a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688" exitCode=2 Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.294550 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a89045f-ad86-47f4-9837-ccae12089508","Type":"ContainerDied","Data":"33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf"} Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.294594 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a89045f-ad86-47f4-9837-ccae12089508","Type":"ContainerDied","Data":"a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688"} Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.395130 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.395208 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.395866 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 10:28:15 crc kubenswrapper[4632]: E0313 10:28:15.711067 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8a89045f_ad86_47f4_9837_ccae12089508.slice/crio-20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874.scope\": RecentStats: unable to find data in memory cache]" Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.824913 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.953553 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a89045f-ad86-47f4-9837-ccae12089508-run-httpd\") pod \"8a89045f-ad86-47f4-9837-ccae12089508\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.953976 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-config-data\") pod \"8a89045f-ad86-47f4-9837-ccae12089508\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.953997 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-sg-core-conf-yaml\") pod \"8a89045f-ad86-47f4-9837-ccae12089508\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.954017 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-scripts\") pod \"8a89045f-ad86-47f4-9837-ccae12089508\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.954055 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-combined-ca-bundle\") pod \"8a89045f-ad86-47f4-9837-ccae12089508\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.954203 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vmhd\" (UniqueName: \"kubernetes.io/projected/8a89045f-ad86-47f4-9837-ccae12089508-kube-api-access-8vmhd\") pod \"8a89045f-ad86-47f4-9837-ccae12089508\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.954221 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a89045f-ad86-47f4-9837-ccae12089508-log-httpd\") pod \"8a89045f-ad86-47f4-9837-ccae12089508\" (UID: \"8a89045f-ad86-47f4-9837-ccae12089508\") " Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.954370 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a89045f-ad86-47f4-9837-ccae12089508-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8a89045f-ad86-47f4-9837-ccae12089508" (UID: "8a89045f-ad86-47f4-9837-ccae12089508"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.954808 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a89045f-ad86-47f4-9837-ccae12089508-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8a89045f-ad86-47f4-9837-ccae12089508" (UID: "8a89045f-ad86-47f4-9837-ccae12089508"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.954914 4632 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a89045f-ad86-47f4-9837-ccae12089508-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.954927 4632 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8a89045f-ad86-47f4-9837-ccae12089508-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.965116 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-scripts" (OuterVolumeSpecName: "scripts") pod "8a89045f-ad86-47f4-9837-ccae12089508" (UID: "8a89045f-ad86-47f4-9837-ccae12089508"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:15 crc kubenswrapper[4632]: I0313 10:28:15.965223 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a89045f-ad86-47f4-9837-ccae12089508-kube-api-access-8vmhd" (OuterVolumeSpecName: "kube-api-access-8vmhd") pod "8a89045f-ad86-47f4-9837-ccae12089508" (UID: "8a89045f-ad86-47f4-9837-ccae12089508"). InnerVolumeSpecName "kube-api-access-8vmhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.048261 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8a89045f-ad86-47f4-9837-ccae12089508" (UID: "8a89045f-ad86-47f4-9837-ccae12089508"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.058181 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vmhd\" (UniqueName: \"kubernetes.io/projected/8a89045f-ad86-47f4-9837-ccae12089508-kube-api-access-8vmhd\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.058211 4632 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.058221 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.128871 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a89045f-ad86-47f4-9837-ccae12089508" (UID: "8a89045f-ad86-47f4-9837-ccae12089508"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.160027 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.207076 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-config-data" (OuterVolumeSpecName: "config-data") pod "8a89045f-ad86-47f4-9837-ccae12089508" (UID: "8a89045f-ad86-47f4-9837-ccae12089508"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.261747 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a89045f-ad86-47f4-9837-ccae12089508-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.306721 4632 generic.go:334] "Generic (PLEG): container finished" podID="8a89045f-ad86-47f4-9837-ccae12089508" containerID="20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874" exitCode=0 Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.306751 4632 generic.go:334] "Generic (PLEG): container finished" podID="8a89045f-ad86-47f4-9837-ccae12089508" containerID="9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0" exitCode=0 Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.306835 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.306825 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a89045f-ad86-47f4-9837-ccae12089508","Type":"ContainerDied","Data":"20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874"} Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.306994 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a89045f-ad86-47f4-9837-ccae12089508","Type":"ContainerDied","Data":"9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0"} Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.307018 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8a89045f-ad86-47f4-9837-ccae12089508","Type":"ContainerDied","Data":"f18f1919b2952189a9f9202d1451d5a08fb72b179ba9686af3513e79b4500446"} Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.307042 4632 scope.go:117] "RemoveContainer" containerID="33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.315120 4632 generic.go:334] "Generic (PLEG): container finished" podID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerID="433c9aa5a02161c4bc7228b52cc460020479cbbb899bc6549755a59b8ad796f4" exitCode=137 Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.315187 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689764498d-rg7vt" event={"ID":"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c","Type":"ContainerDied","Data":"433c9aa5a02161c4bc7228b52cc460020479cbbb899bc6549755a59b8ad796f4"} Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.672794 4632 scope.go:117] "RemoveContainer" containerID="a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.701135 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.707151 4632 scope.go:117] "RemoveContainer" containerID="20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.721462 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.734119 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:16 crc kubenswrapper[4632]: E0313 10:28:16.734661 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="sg-core" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.734700 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="sg-core" Mar 13 10:28:16 crc kubenswrapper[4632]: E0313 10:28:16.734753 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="ceilometer-central-agent" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.734767 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="ceilometer-central-agent" Mar 13 10:28:16 crc kubenswrapper[4632]: E0313 10:28:16.734792 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="ceilometer-notification-agent" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.734810 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="ceilometer-notification-agent" Mar 13 10:28:16 crc kubenswrapper[4632]: E0313 10:28:16.734868 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="658f9ba3-69b7-4d2d-8258-bb7bdf272398" containerName="oc" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.734881 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="658f9ba3-69b7-4d2d-8258-bb7bdf272398" containerName="oc" Mar 13 10:28:16 crc kubenswrapper[4632]: E0313 10:28:16.734910 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="proxy-httpd" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.734922 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="proxy-httpd" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.735233 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="sg-core" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.735272 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="ceilometer-notification-agent" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.735295 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="658f9ba3-69b7-4d2d-8258-bb7bdf272398" containerName="oc" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.735313 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="proxy-httpd" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.735334 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a89045f-ad86-47f4-9837-ccae12089508" containerName="ceilometer-central-agent" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.737311 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.741028 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.741845 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.743674 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.762449 4632 scope.go:117] "RemoveContainer" containerID="9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.796782 4632 scope.go:117] "RemoveContainer" containerID="33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf" Mar 13 10:28:16 crc kubenswrapper[4632]: E0313 10:28:16.798122 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf\": container with ID starting with 33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf not found: ID does not exist" containerID="33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.798155 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf"} err="failed to get container status \"33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf\": rpc error: code = NotFound desc = could not find container \"33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf\": container with ID starting with 33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf not found: ID does not exist" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.798175 4632 scope.go:117] "RemoveContainer" containerID="a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688" Mar 13 10:28:16 crc kubenswrapper[4632]: E0313 10:28:16.803063 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688\": container with ID starting with a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688 not found: ID does not exist" containerID="a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.803242 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688"} err="failed to get container status \"a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688\": rpc error: code = NotFound desc = could not find container \"a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688\": container with ID starting with a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688 not found: ID does not exist" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.803332 4632 scope.go:117] "RemoveContainer" containerID="20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874" Mar 13 10:28:16 crc kubenswrapper[4632]: E0313 10:28:16.803824 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874\": container with ID starting with 20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874 not found: ID does not exist" containerID="20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.803879 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874"} err="failed to get container status \"20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874\": rpc error: code = NotFound desc = could not find container \"20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874\": container with ID starting with 20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874 not found: ID does not exist" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.803907 4632 scope.go:117] "RemoveContainer" containerID="9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0" Mar 13 10:28:16 crc kubenswrapper[4632]: E0313 10:28:16.805464 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0\": container with ID starting with 9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0 not found: ID does not exist" containerID="9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.805500 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0"} err="failed to get container status \"9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0\": rpc error: code = NotFound desc = could not find container \"9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0\": container with ID starting with 9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0 not found: ID does not exist" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.805524 4632 scope.go:117] "RemoveContainer" containerID="33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.807149 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf"} err="failed to get container status \"33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf\": rpc error: code = NotFound desc = could not find container \"33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf\": container with ID starting with 33db7b55742606799ae347e8dc24186eda0df552d34a8ff8896a3591ab4845bf not found: ID does not exist" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.807189 4632 scope.go:117] "RemoveContainer" containerID="a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.807791 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688"} err="failed to get container status \"a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688\": rpc error: code = NotFound desc = could not find container \"a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688\": container with ID starting with a98d9a2847bb67335882172b35537561c8c602078a182d937a48977eab66a688 not found: ID does not exist" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.807814 4632 scope.go:117] "RemoveContainer" containerID="20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.808636 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874"} err="failed to get container status \"20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874\": rpc error: code = NotFound desc = could not find container \"20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874\": container with ID starting with 20cdc9a0316367f683a4923e148488d3441ea045da71a429f3922a31b37d7874 not found: ID does not exist" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.808661 4632 scope.go:117] "RemoveContainer" containerID="9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.808950 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0"} err="failed to get container status \"9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0\": rpc error: code = NotFound desc = could not find container \"9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0\": container with ID starting with 9e4a51f5856378812a236a35be56fcba04f467dfec7755d8229b1d7b6a9e6bd0 not found: ID does not exist" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.808971 4632 scope.go:117] "RemoveContainer" containerID="8ce0185281fb59d0c6bda2b2c484ad3711b4bd3b729b4b8677e75ca6b8e1f739" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.875656 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.875728 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efe2db4a-a5d6-4aa3-805b-5144c66afca8-run-httpd\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.876382 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.876468 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efe2db4a-a5d6-4aa3-805b-5144c66afca8-log-httpd\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.876571 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8nnb\" (UniqueName: \"kubernetes.io/projected/efe2db4a-a5d6-4aa3-805b-5144c66afca8-kube-api-access-b8nnb\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.876661 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-config-data\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.876722 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-scripts\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.979129 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.979196 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efe2db4a-a5d6-4aa3-805b-5144c66afca8-run-httpd\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.979261 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efe2db4a-a5d6-4aa3-805b-5144c66afca8-log-httpd\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.979346 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8nnb\" (UniqueName: \"kubernetes.io/projected/efe2db4a-a5d6-4aa3-805b-5144c66afca8-kube-api-access-b8nnb\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.979399 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-config-data\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.979427 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-scripts\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.979476 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.981355 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efe2db4a-a5d6-4aa3-805b-5144c66afca8-log-httpd\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.981671 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efe2db4a-a5d6-4aa3-805b-5144c66afca8-run-httpd\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.986327 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-scripts\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.986679 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.989878 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:16 crc kubenswrapper[4632]: I0313 10:28:16.991067 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-config-data\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:17 crc kubenswrapper[4632]: I0313 10:28:17.013113 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8nnb\" (UniqueName: \"kubernetes.io/projected/efe2db4a-a5d6-4aa3-805b-5144c66afca8-kube-api-access-b8nnb\") pod \"ceilometer-0\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " pod="openstack/ceilometer-0" Mar 13 10:28:17 crc kubenswrapper[4632]: I0313 10:28:17.064892 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:17 crc kubenswrapper[4632]: I0313 10:28:17.345868 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689764498d-rg7vt" event={"ID":"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c","Type":"ContainerStarted","Data":"26a7aae686bb479cfcbc8b01e8e10e3fd467e5236d6ffb2ed638373687267401"} Mar 13 10:28:17 crc kubenswrapper[4632]: I0313 10:28:17.699795 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:17 crc kubenswrapper[4632]: I0313 10:28:17.886595 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:28:17 crc kubenswrapper[4632]: I0313 10:28:17.887123 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="62c1f3f8-e898-4481-88e0-49f0c20228a4" containerName="glance-log" containerID="cri-o://75e2995816c15269a0e0bb8513c4f7b9cace1b33dd417df2fc8f694c18b89fa0" gracePeriod=30 Mar 13 10:28:17 crc kubenswrapper[4632]: I0313 10:28:17.887516 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="62c1f3f8-e898-4481-88e0-49f0c20228a4" containerName="glance-httpd" containerID="cri-o://60e19c69317a817c5bf104bc8691bdf46121d52039ad19099e25f869718b8e19" gracePeriod=30 Mar 13 10:28:18 crc kubenswrapper[4632]: I0313 10:28:18.064495 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a89045f-ad86-47f4-9837-ccae12089508" path="/var/lib/kubelet/pods/8a89045f-ad86-47f4-9837-ccae12089508/volumes" Mar 13 10:28:18 crc kubenswrapper[4632]: I0313 10:28:18.368363 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efe2db4a-a5d6-4aa3-805b-5144c66afca8","Type":"ContainerStarted","Data":"e1cfa6629f759d5883022ca95d30f701e046142894117805d6775261138f329b"} Mar 13 10:28:18 crc kubenswrapper[4632]: I0313 10:28:18.368716 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efe2db4a-a5d6-4aa3-805b-5144c66afca8","Type":"ContainerStarted","Data":"14a7a36e446e3ed959c2f673ec8a5768ce0dbc971f3d7879cc430c0491594679"} Mar 13 10:28:18 crc kubenswrapper[4632]: I0313 10:28:18.388317 4632 generic.go:334] "Generic (PLEG): container finished" podID="62c1f3f8-e898-4481-88e0-49f0c20228a4" containerID="75e2995816c15269a0e0bb8513c4f7b9cace1b33dd417df2fc8f694c18b89fa0" exitCode=143 Mar 13 10:28:18 crc kubenswrapper[4632]: I0313 10:28:18.388398 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"62c1f3f8-e898-4481-88e0-49f0c20228a4","Type":"ContainerDied","Data":"75e2995816c15269a0e0bb8513c4f7b9cace1b33dd417df2fc8f694c18b89fa0"} Mar 13 10:28:18 crc kubenswrapper[4632]: I0313 10:28:18.401856 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5mlm2" event={"ID":"5de81924-9bfc-484e-8276-0216f0bbf72c","Type":"ContainerStarted","Data":"afb05bb00debb2ea4a81d169362ff2bd38d824053184e249dbe02cc1cb10e945"} Mar 13 10:28:18 crc kubenswrapper[4632]: I0313 10:28:18.440959 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-5mlm2" podStartSLOduration=3.499571533 podStartE2EDuration="44.440921152s" podCreationTimestamp="2026-03-13 10:27:34 +0000 UTC" firstStartedPulling="2026-03-13 10:27:35.835190159 +0000 UTC m=+1429.857720292" lastFinishedPulling="2026-03-13 10:28:16.776539778 +0000 UTC m=+1470.799069911" observedRunningTime="2026-03-13 10:28:18.429693559 +0000 UTC m=+1472.452223712" watchObservedRunningTime="2026-03-13 10:28:18.440921152 +0000 UTC m=+1472.463451285" Mar 13 10:28:19 crc kubenswrapper[4632]: I0313 10:28:19.411242 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efe2db4a-a5d6-4aa3-805b-5144c66afca8","Type":"ContainerStarted","Data":"fd5b97322f3966f77a91e26539a007fdcb71ceff62ef4db0eeb167b87655caa1"} Mar 13 10:28:19 crc kubenswrapper[4632]: I0313 10:28:19.866357 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:28:19 crc kubenswrapper[4632]: I0313 10:28:19.867269 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="050df504-63b9-4453-be2b-f3b0315fb801" containerName="glance-log" containerID="cri-o://4286cd55d064d024725ded90d153143e568de28aeedc6a6060f69501102dd4cb" gracePeriod=30 Mar 13 10:28:19 crc kubenswrapper[4632]: I0313 10:28:19.867392 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="050df504-63b9-4453-be2b-f3b0315fb801" containerName="glance-httpd" containerID="cri-o://a06e9823c7700968605c221a9839cf4f237fe6a7eee8836d69bade62686f4372" gracePeriod=30 Mar 13 10:28:20 crc kubenswrapper[4632]: I0313 10:28:20.433387 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efe2db4a-a5d6-4aa3-805b-5144c66afca8","Type":"ContainerStarted","Data":"23c0b66b1c919daecf5753fdf594be220c0bfa625f5256dff7fbdd421ccaaa54"} Mar 13 10:28:20 crc kubenswrapper[4632]: I0313 10:28:20.438628 4632 generic.go:334] "Generic (PLEG): container finished" podID="050df504-63b9-4453-be2b-f3b0315fb801" containerID="4286cd55d064d024725ded90d153143e568de28aeedc6a6060f69501102dd4cb" exitCode=143 Mar 13 10:28:20 crc kubenswrapper[4632]: I0313 10:28:20.438687 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"050df504-63b9-4453-be2b-f3b0315fb801","Type":"ContainerDied","Data":"4286cd55d064d024725ded90d153143e568de28aeedc6a6060f69501102dd4cb"} Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.465423 4632 generic.go:334] "Generic (PLEG): container finished" podID="62c1f3f8-e898-4481-88e0-49f0c20228a4" containerID="60e19c69317a817c5bf104bc8691bdf46121d52039ad19099e25f869718b8e19" exitCode=0 Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.467860 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"62c1f3f8-e898-4481-88e0-49f0c20228a4","Type":"ContainerDied","Data":"60e19c69317a817c5bf104bc8691bdf46121d52039ad19099e25f869718b8e19"} Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.496151 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efe2db4a-a5d6-4aa3-805b-5144c66afca8","Type":"ContainerStarted","Data":"3519acd0a5d01d518140b71f8cb487e29259fada712cb5fe0b79aa9039c08494"} Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.724589 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.757777 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.63353254 podStartE2EDuration="6.757754344s" podCreationTimestamp="2026-03-13 10:28:16 +0000 UTC" firstStartedPulling="2026-03-13 10:28:17.715277411 +0000 UTC m=+1471.737807544" lastFinishedPulling="2026-03-13 10:28:20.839499215 +0000 UTC m=+1474.862029348" observedRunningTime="2026-03-13 10:28:22.559067011 +0000 UTC m=+1476.581597154" watchObservedRunningTime="2026-03-13 10:28:22.757754344 +0000 UTC m=+1476.780284477" Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.820405 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbnq2\" (UniqueName: \"kubernetes.io/projected/62c1f3f8-e898-4481-88e0-49f0c20228a4-kube-api-access-gbnq2\") pod \"62c1f3f8-e898-4481-88e0-49f0c20228a4\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.820523 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-public-tls-certs\") pod \"62c1f3f8-e898-4481-88e0-49f0c20228a4\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.820560 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/62c1f3f8-e898-4481-88e0-49f0c20228a4-httpd-run\") pod \"62c1f3f8-e898-4481-88e0-49f0c20228a4\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.820656 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-scripts\") pod \"62c1f3f8-e898-4481-88e0-49f0c20228a4\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.820688 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62c1f3f8-e898-4481-88e0-49f0c20228a4-logs\") pod \"62c1f3f8-e898-4481-88e0-49f0c20228a4\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.820718 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"62c1f3f8-e898-4481-88e0-49f0c20228a4\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.820763 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-config-data\") pod \"62c1f3f8-e898-4481-88e0-49f0c20228a4\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.820847 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-combined-ca-bundle\") pod \"62c1f3f8-e898-4481-88e0-49f0c20228a4\" (UID: \"62c1f3f8-e898-4481-88e0-49f0c20228a4\") " Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.821248 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62c1f3f8-e898-4481-88e0-49f0c20228a4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "62c1f3f8-e898-4481-88e0-49f0c20228a4" (UID: "62c1f3f8-e898-4481-88e0-49f0c20228a4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.821419 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/62c1f3f8-e898-4481-88e0-49f0c20228a4-logs" (OuterVolumeSpecName: "logs") pod "62c1f3f8-e898-4481-88e0-49f0c20228a4" (UID: "62c1f3f8-e898-4481-88e0-49f0c20228a4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.823928 4632 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/62c1f3f8-e898-4481-88e0-49f0c20228a4-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.824069 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62c1f3f8-e898-4481-88e0-49f0c20228a4-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.842122 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "62c1f3f8-e898-4481-88e0-49f0c20228a4" (UID: "62c1f3f8-e898-4481-88e0-49f0c20228a4"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.842322 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62c1f3f8-e898-4481-88e0-49f0c20228a4-kube-api-access-gbnq2" (OuterVolumeSpecName: "kube-api-access-gbnq2") pod "62c1f3f8-e898-4481-88e0-49f0c20228a4" (UID: "62c1f3f8-e898-4481-88e0-49f0c20228a4"). InnerVolumeSpecName "kube-api-access-gbnq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.875806 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-scripts" (OuterVolumeSpecName: "scripts") pod "62c1f3f8-e898-4481-88e0-49f0c20228a4" (UID: "62c1f3f8-e898-4481-88e0-49f0c20228a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.893480 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "62c1f3f8-e898-4481-88e0-49f0c20228a4" (UID: "62c1f3f8-e898-4481-88e0-49f0c20228a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.925365 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbnq2\" (UniqueName: \"kubernetes.io/projected/62c1f3f8-e898-4481-88e0-49f0c20228a4-kube-api-access-gbnq2\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.925404 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.926167 4632 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Mar 13 10:28:22 crc kubenswrapper[4632]: I0313 10:28:22.926201 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.006683 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-config-data" (OuterVolumeSpecName: "config-data") pod "62c1f3f8-e898-4481-88e0-49f0c20228a4" (UID: "62c1f3f8-e898-4481-88e0-49f0c20228a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.019612 4632 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.030120 4632 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.030153 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.040869 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "62c1f3f8-e898-4481-88e0-49f0c20228a4" (UID: "62c1f3f8-e898-4481-88e0-49f0c20228a4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.131584 4632 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62c1f3f8-e898-4481-88e0-49f0c20228a4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.523613 4632 generic.go:334] "Generic (PLEG): container finished" podID="050df504-63b9-4453-be2b-f3b0315fb801" containerID="a06e9823c7700968605c221a9839cf4f237fe6a7eee8836d69bade62686f4372" exitCode=0 Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.524060 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"050df504-63b9-4453-be2b-f3b0315fb801","Type":"ContainerDied","Data":"a06e9823c7700968605c221a9839cf4f237fe6a7eee8836d69bade62686f4372"} Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.531464 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.533201 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"62c1f3f8-e898-4481-88e0-49f0c20228a4","Type":"ContainerDied","Data":"da67a58c5a020c95fb415df6f51542675c8d6697cd1fafcacdcc7d6081f0a9ff"} Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.533269 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.533580 4632 scope.go:117] "RemoveContainer" containerID="60e19c69317a817c5bf104bc8691bdf46121d52039ad19099e25f869718b8e19" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.616104 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.636277 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.656985 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:28:23 crc kubenswrapper[4632]: E0313 10:28:23.657386 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c1f3f8-e898-4481-88e0-49f0c20228a4" containerName="glance-log" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.657398 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c1f3f8-e898-4481-88e0-49f0c20228a4" containerName="glance-log" Mar 13 10:28:23 crc kubenswrapper[4632]: E0313 10:28:23.657413 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c1f3f8-e898-4481-88e0-49f0c20228a4" containerName="glance-httpd" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.657419 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c1f3f8-e898-4481-88e0-49f0c20228a4" containerName="glance-httpd" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.657589 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c1f3f8-e898-4481-88e0-49f0c20228a4" containerName="glance-log" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.657613 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c1f3f8-e898-4481-88e0-49f0c20228a4" containerName="glance-httpd" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.683773 4632 scope.go:117] "RemoveContainer" containerID="75e2995816c15269a0e0bb8513c4f7b9cace1b33dd417df2fc8f694c18b89fa0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.694923 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.706648 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.716562 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.790718 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.867110 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.867512 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2394af9-fd85-4291-8d57-c2bff02eccce-logs\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.867626 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2394af9-fd85-4291-8d57-c2bff02eccce-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.867749 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2394af9-fd85-4291-8d57-c2bff02eccce-scripts\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.867905 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2394af9-fd85-4291-8d57-c2bff02eccce-config-data\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.868042 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbr58\" (UniqueName: \"kubernetes.io/projected/a2394af9-fd85-4291-8d57-c2bff02eccce-kube-api-access-rbr58\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.868162 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2394af9-fd85-4291-8d57-c2bff02eccce-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.868297 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2394af9-fd85-4291-8d57-c2bff02eccce-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.967851 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.969741 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.969850 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2394af9-fd85-4291-8d57-c2bff02eccce-logs\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.969900 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2394af9-fd85-4291-8d57-c2bff02eccce-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.969988 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2394af9-fd85-4291-8d57-c2bff02eccce-scripts\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.970049 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2394af9-fd85-4291-8d57-c2bff02eccce-config-data\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.970085 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbr58\" (UniqueName: \"kubernetes.io/projected/a2394af9-fd85-4291-8d57-c2bff02eccce-kube-api-access-rbr58\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.970103 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2394af9-fd85-4291-8d57-c2bff02eccce-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.970148 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2394af9-fd85-4291-8d57-c2bff02eccce-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.970782 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a2394af9-fd85-4291-8d57-c2bff02eccce-logs\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.971259 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.971408 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a2394af9-fd85-4291-8d57-c2bff02eccce-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.984438 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a2394af9-fd85-4291-8d57-c2bff02eccce-config-data\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:23 crc kubenswrapper[4632]: I0313 10:28:23.986296 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a2394af9-fd85-4291-8d57-c2bff02eccce-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.004799 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a2394af9-fd85-4291-8d57-c2bff02eccce-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.020718 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a2394af9-fd85-4291-8d57-c2bff02eccce-scripts\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.034726 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbr58\" (UniqueName: \"kubernetes.io/projected/a2394af9-fd85-4291-8d57-c2bff02eccce-kube-api-access-rbr58\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.063632 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"a2394af9-fd85-4291-8d57-c2bff02eccce\") " pod="openstack/glance-default-external-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.081749 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/050df504-63b9-4453-be2b-f3b0315fb801-httpd-run\") pod \"050df504-63b9-4453-be2b-f3b0315fb801\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.081843 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-config-data\") pod \"050df504-63b9-4453-be2b-f3b0315fb801\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.081895 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-combined-ca-bundle\") pod \"050df504-63b9-4453-be2b-f3b0315fb801\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.082006 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"050df504-63b9-4453-be2b-f3b0315fb801\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.082103 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5bl8\" (UniqueName: \"kubernetes.io/projected/050df504-63b9-4453-be2b-f3b0315fb801-kube-api-access-z5bl8\") pod \"050df504-63b9-4453-be2b-f3b0315fb801\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.082134 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-scripts\") pod \"050df504-63b9-4453-be2b-f3b0315fb801\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.082214 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/050df504-63b9-4453-be2b-f3b0315fb801-logs\") pod \"050df504-63b9-4453-be2b-f3b0315fb801\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.082252 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-internal-tls-certs\") pod \"050df504-63b9-4453-be2b-f3b0315fb801\" (UID: \"050df504-63b9-4453-be2b-f3b0315fb801\") " Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.083688 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/050df504-63b9-4453-be2b-f3b0315fb801-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "050df504-63b9-4453-be2b-f3b0315fb801" (UID: "050df504-63b9-4453-be2b-f3b0315fb801"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.084138 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/050df504-63b9-4453-be2b-f3b0315fb801-logs" (OuterVolumeSpecName: "logs") pod "050df504-63b9-4453-be2b-f3b0315fb801" (UID: "050df504-63b9-4453-be2b-f3b0315fb801"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.109455 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/050df504-63b9-4453-be2b-f3b0315fb801-kube-api-access-z5bl8" (OuterVolumeSpecName: "kube-api-access-z5bl8") pod "050df504-63b9-4453-be2b-f3b0315fb801" (UID: "050df504-63b9-4453-be2b-f3b0315fb801"). InnerVolumeSpecName "kube-api-access-z5bl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.113244 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "050df504-63b9-4453-be2b-f3b0315fb801" (UID: "050df504-63b9-4453-be2b-f3b0315fb801"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.115355 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62c1f3f8-e898-4481-88e0-49f0c20228a4" path="/var/lib/kubelet/pods/62c1f3f8-e898-4481-88e0-49f0c20228a4/volumes" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.120139 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-scripts" (OuterVolumeSpecName: "scripts") pod "050df504-63b9-4453-be2b-f3b0315fb801" (UID: "050df504-63b9-4453-be2b-f3b0315fb801"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.181277 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "050df504-63b9-4453-be2b-f3b0315fb801" (UID: "050df504-63b9-4453-be2b-f3b0315fb801"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.184927 4632 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/050df504-63b9-4453-be2b-f3b0315fb801-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.185016 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.185040 4632 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.185050 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5bl8\" (UniqueName: \"kubernetes.io/projected/050df504-63b9-4453-be2b-f3b0315fb801-kube-api-access-z5bl8\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.185060 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.185069 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/050df504-63b9-4453-be2b-f3b0315fb801-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.219526 4632 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.256159 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "050df504-63b9-4453-be2b-f3b0315fb801" (UID: "050df504-63b9-4453-be2b-f3b0315fb801"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.270342 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.288490 4632 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.288516 4632 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.304359 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-config-data" (OuterVolumeSpecName: "config-data") pod "050df504-63b9-4453-be2b-f3b0315fb801" (UID: "050df504-63b9-4453-be2b-f3b0315fb801"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.391571 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/050df504-63b9-4453-be2b-f3b0315fb801-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.594604 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.596008 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"050df504-63b9-4453-be2b-f3b0315fb801","Type":"ContainerDied","Data":"639dbfcf9c85b2d6df276ce37ddb572204028d4a54aa36f7c4d3026c9ff6abfc"} Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.596111 4632 scope.go:117] "RemoveContainer" containerID="a06e9823c7700968605c221a9839cf4f237fe6a7eee8836d69bade62686f4372" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.750522 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.763671 4632 scope.go:117] "RemoveContainer" containerID="4286cd55d064d024725ded90d153143e568de28aeedc6a6060f69501102dd4cb" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.780838 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.798125 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:28:24 crc kubenswrapper[4632]: E0313 10:28:24.798481 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="050df504-63b9-4453-be2b-f3b0315fb801" containerName="glance-log" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.798493 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="050df504-63b9-4453-be2b-f3b0315fb801" containerName="glance-log" Mar 13 10:28:24 crc kubenswrapper[4632]: E0313 10:28:24.798506 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="050df504-63b9-4453-be2b-f3b0315fb801" containerName="glance-httpd" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.798513 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="050df504-63b9-4453-be2b-f3b0315fb801" containerName="glance-httpd" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.798710 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="050df504-63b9-4453-be2b-f3b0315fb801" containerName="glance-httpd" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.798726 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="050df504-63b9-4453-be2b-f3b0315fb801" containerName="glance-log" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.799650 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.804308 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.804926 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.804968 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.911959 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97cf3e4a-cbe1-441c-8652-281a30fcf432-logs\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.912303 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97cf3e4a-cbe1-441c-8652-281a30fcf432-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.912365 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97cf3e4a-cbe1-441c-8652-281a30fcf432-config-data\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.912418 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97cf3e4a-cbe1-441c-8652-281a30fcf432-scripts\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.912465 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-khmr9\" (UniqueName: \"kubernetes.io/projected/97cf3e4a-cbe1-441c-8652-281a30fcf432-kube-api-access-khmr9\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.912521 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.912673 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97cf3e4a-cbe1-441c-8652-281a30fcf432-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:24 crc kubenswrapper[4632]: I0313 10:28:24.912729 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97cf3e4a-cbe1-441c-8652-281a30fcf432-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.018502 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97cf3e4a-cbe1-441c-8652-281a30fcf432-scripts\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.018575 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-khmr9\" (UniqueName: \"kubernetes.io/projected/97cf3e4a-cbe1-441c-8652-281a30fcf432-kube-api-access-khmr9\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.018616 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.018710 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97cf3e4a-cbe1-441c-8652-281a30fcf432-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.018749 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97cf3e4a-cbe1-441c-8652-281a30fcf432-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.018873 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97cf3e4a-cbe1-441c-8652-281a30fcf432-logs\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.018903 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97cf3e4a-cbe1-441c-8652-281a30fcf432-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.018925 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97cf3e4a-cbe1-441c-8652-281a30fcf432-config-data\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.024215 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.045873 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/97cf3e4a-cbe1-441c-8652-281a30fcf432-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.072641 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/97cf3e4a-cbe1-441c-8652-281a30fcf432-logs\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.094573 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/97cf3e4a-cbe1-441c-8652-281a30fcf432-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.096379 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97cf3e4a-cbe1-441c-8652-281a30fcf432-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.096685 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-khmr9\" (UniqueName: \"kubernetes.io/projected/97cf3e4a-cbe1-441c-8652-281a30fcf432-kube-api-access-khmr9\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.098868 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97cf3e4a-cbe1-441c-8652-281a30fcf432-config-data\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.101515 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/97cf3e4a-cbe1-441c-8652-281a30fcf432-scripts\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.131979 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.140674 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-0\" (UID: \"97cf3e4a-cbe1-441c-8652-281a30fcf432\") " pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.394787 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.437870 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.612417 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a2394af9-fd85-4291-8d57-c2bff02eccce","Type":"ContainerStarted","Data":"0cf5c13e47b817041ef3b078440c25d045c2d95bbad47e2762b765047a98b062"} Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.858376 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:28:25 crc kubenswrapper[4632]: I0313 10:28:25.858497 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:28:26 crc kubenswrapper[4632]: I0313 10:28:26.082317 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="050df504-63b9-4453-be2b-f3b0315fb801" path="/var/lib/kubelet/pods/050df504-63b9-4453-be2b-f3b0315fb801/volumes" Mar 13 10:28:26 crc kubenswrapper[4632]: I0313 10:28:26.241207 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 10:28:26 crc kubenswrapper[4632]: I0313 10:28:26.624990 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97cf3e4a-cbe1-441c-8652-281a30fcf432","Type":"ContainerStarted","Data":"61a795a507a3ddb80280eaf24f52ffcec864d321ef4b538cb1098ee5adec2393"} Mar 13 10:28:27 crc kubenswrapper[4632]: I0313 10:28:27.170561 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:27 crc kubenswrapper[4632]: I0313 10:28:27.170893 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="ceilometer-central-agent" containerID="cri-o://e1cfa6629f759d5883022ca95d30f701e046142894117805d6775261138f329b" gracePeriod=30 Mar 13 10:28:27 crc kubenswrapper[4632]: I0313 10:28:27.171080 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="proxy-httpd" containerID="cri-o://3519acd0a5d01d518140b71f8cb487e29259fada712cb5fe0b79aa9039c08494" gracePeriod=30 Mar 13 10:28:27 crc kubenswrapper[4632]: I0313 10:28:27.171141 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="sg-core" containerID="cri-o://23c0b66b1c919daecf5753fdf594be220c0bfa625f5256dff7fbdd421ccaaa54" gracePeriod=30 Mar 13 10:28:27 crc kubenswrapper[4632]: I0313 10:28:27.171185 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="ceilometer-notification-agent" containerID="cri-o://fd5b97322f3966f77a91e26539a007fdcb71ceff62ef4db0eeb167b87655caa1" gracePeriod=30 Mar 13 10:28:27 crc kubenswrapper[4632]: I0313 10:28:27.698431 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a2394af9-fd85-4291-8d57-c2bff02eccce","Type":"ContainerStarted","Data":"00a40bf94c6c666706b8bac7ae5ff70cf3bc92133876a5b19b0393024843c7de"} Mar 13 10:28:27 crc kubenswrapper[4632]: I0313 10:28:27.712077 4632 generic.go:334] "Generic (PLEG): container finished" podID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerID="3519acd0a5d01d518140b71f8cb487e29259fada712cb5fe0b79aa9039c08494" exitCode=0 Mar 13 10:28:27 crc kubenswrapper[4632]: I0313 10:28:27.712125 4632 generic.go:334] "Generic (PLEG): container finished" podID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerID="23c0b66b1c919daecf5753fdf594be220c0bfa625f5256dff7fbdd421ccaaa54" exitCode=2 Mar 13 10:28:27 crc kubenswrapper[4632]: I0313 10:28:27.712164 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efe2db4a-a5d6-4aa3-805b-5144c66afca8","Type":"ContainerDied","Data":"3519acd0a5d01d518140b71f8cb487e29259fada712cb5fe0b79aa9039c08494"} Mar 13 10:28:27 crc kubenswrapper[4632]: I0313 10:28:27.712203 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efe2db4a-a5d6-4aa3-805b-5144c66afca8","Type":"ContainerDied","Data":"23c0b66b1c919daecf5753fdf594be220c0bfa625f5256dff7fbdd421ccaaa54"} Mar 13 10:28:27 crc kubenswrapper[4632]: I0313 10:28:27.734200 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97cf3e4a-cbe1-441c-8652-281a30fcf432","Type":"ContainerStarted","Data":"c9b9c1eda565b34af6716a79a4e020a175fe5e6d574afe1ce2632c16c21b446c"} Mar 13 10:28:28 crc kubenswrapper[4632]: I0313 10:28:28.744631 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a2394af9-fd85-4291-8d57-c2bff02eccce","Type":"ContainerStarted","Data":"8d71651c55d74774fc4c858767799012d9f0bd393e8530074f6103fe0f351e36"} Mar 13 10:28:28 crc kubenswrapper[4632]: I0313 10:28:28.748656 4632 generic.go:334] "Generic (PLEG): container finished" podID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerID="fd5b97322f3966f77a91e26539a007fdcb71ceff62ef4db0eeb167b87655caa1" exitCode=0 Mar 13 10:28:28 crc kubenswrapper[4632]: I0313 10:28:28.748737 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efe2db4a-a5d6-4aa3-805b-5144c66afca8","Type":"ContainerDied","Data":"fd5b97322f3966f77a91e26539a007fdcb71ceff62ef4db0eeb167b87655caa1"} Mar 13 10:28:28 crc kubenswrapper[4632]: I0313 10:28:28.752354 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"97cf3e4a-cbe1-441c-8652-281a30fcf432","Type":"ContainerStarted","Data":"b8ac69eb67c3fe143261ad7193c00cd6a79532e49d72f84ee4793d18275f9cb5"} Mar 13 10:28:28 crc kubenswrapper[4632]: I0313 10:28:28.801773 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.801756916 podStartE2EDuration="5.801756916s" podCreationTimestamp="2026-03-13 10:28:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:28:28.781917364 +0000 UTC m=+1482.804447487" watchObservedRunningTime="2026-03-13 10:28:28.801756916 +0000 UTC m=+1482.824287049" Mar 13 10:28:28 crc kubenswrapper[4632]: I0313 10:28:28.806159 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.806143894 podStartE2EDuration="4.806143894s" podCreationTimestamp="2026-03-13 10:28:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:28:28.797803581 +0000 UTC m=+1482.820333714" watchObservedRunningTime="2026-03-13 10:28:28.806143894 +0000 UTC m=+1482.828674027" Mar 13 10:28:31 crc kubenswrapper[4632]: I0313 10:28:31.785535 4632 generic.go:334] "Generic (PLEG): container finished" podID="5d10747e-ba77-4986-9d4b-636fcbf823ab" containerID="35f6f30aa35f7a79445d6acba6d7d99ce02bc8679e546b9d8ecccf0df51e3ce6" exitCode=137 Mar 13 10:28:31 crc kubenswrapper[4632]: I0313 10:28:31.785872 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-d856c56c-cmd2q" event={"ID":"5d10747e-ba77-4986-9d4b-636fcbf823ab","Type":"ContainerDied","Data":"35f6f30aa35f7a79445d6acba6d7d99ce02bc8679e546b9d8ecccf0df51e3ce6"} Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.144794 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.299790 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prc4f\" (UniqueName: \"kubernetes.io/projected/5d10747e-ba77-4986-9d4b-636fcbf823ab-kube-api-access-prc4f\") pod \"5d10747e-ba77-4986-9d4b-636fcbf823ab\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.300059 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-config-data\") pod \"5d10747e-ba77-4986-9d4b-636fcbf823ab\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.300126 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-config-data-custom\") pod \"5d10747e-ba77-4986-9d4b-636fcbf823ab\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.300181 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-combined-ca-bundle\") pod \"5d10747e-ba77-4986-9d4b-636fcbf823ab\" (UID: \"5d10747e-ba77-4986-9d4b-636fcbf823ab\") " Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.312361 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d10747e-ba77-4986-9d4b-636fcbf823ab-kube-api-access-prc4f" (OuterVolumeSpecName: "kube-api-access-prc4f") pod "5d10747e-ba77-4986-9d4b-636fcbf823ab" (UID: "5d10747e-ba77-4986-9d4b-636fcbf823ab"). InnerVolumeSpecName "kube-api-access-prc4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.317813 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5d10747e-ba77-4986-9d4b-636fcbf823ab" (UID: "5d10747e-ba77-4986-9d4b-636fcbf823ab"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.397902 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5d10747e-ba77-4986-9d4b-636fcbf823ab" (UID: "5d10747e-ba77-4986-9d4b-636fcbf823ab"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.414643 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prc4f\" (UniqueName: \"kubernetes.io/projected/5d10747e-ba77-4986-9d4b-636fcbf823ab-kube-api-access-prc4f\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.415015 4632 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.415028 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.416805 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.457344 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-config-data" (OuterVolumeSpecName: "config-data") pod "5d10747e-ba77-4986-9d4b-636fcbf823ab" (UID: "5d10747e-ba77-4986-9d4b-636fcbf823ab"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.517091 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-config-data\") pod \"07914020-653d-4509-9f60-22726224c7c6\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.517244 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nn98r\" (UniqueName: \"kubernetes.io/projected/07914020-653d-4509-9f60-22726224c7c6-kube-api-access-nn98r\") pod \"07914020-653d-4509-9f60-22726224c7c6\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.517324 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-combined-ca-bundle\") pod \"07914020-653d-4509-9f60-22726224c7c6\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.517385 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-config-data-custom\") pod \"07914020-653d-4509-9f60-22726224c7c6\" (UID: \"07914020-653d-4509-9f60-22726224c7c6\") " Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.517973 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5d10747e-ba77-4986-9d4b-636fcbf823ab-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.523922 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07914020-653d-4509-9f60-22726224c7c6-kube-api-access-nn98r" (OuterVolumeSpecName: "kube-api-access-nn98r") pod "07914020-653d-4509-9f60-22726224c7c6" (UID: "07914020-653d-4509-9f60-22726224c7c6"). InnerVolumeSpecName "kube-api-access-nn98r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.524406 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "07914020-653d-4509-9f60-22726224c7c6" (UID: "07914020-653d-4509-9f60-22726224c7c6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.555254 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07914020-653d-4509-9f60-22726224c7c6" (UID: "07914020-653d-4509-9f60-22726224c7c6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.587387 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-config-data" (OuterVolumeSpecName: "config-data") pod "07914020-653d-4509-9f60-22726224c7c6" (UID: "07914020-653d-4509-9f60-22726224c7c6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.619741 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.619798 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nn98r\" (UniqueName: \"kubernetes.io/projected/07914020-653d-4509-9f60-22726224c7c6-kube-api-access-nn98r\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.619813 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.619826 4632 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/07914020-653d-4509-9f60-22726224c7c6-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.804932 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-d856c56c-cmd2q" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.805735 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-d856c56c-cmd2q" event={"ID":"5d10747e-ba77-4986-9d4b-636fcbf823ab","Type":"ContainerDied","Data":"c4b118bba3eb9eaa2f3d30625225786b624eac290ce33f3a700f116e125abbc7"} Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.805836 4632 scope.go:117] "RemoveContainer" containerID="35f6f30aa35f7a79445d6acba6d7d99ce02bc8679e546b9d8ecccf0df51e3ce6" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.808549 4632 generic.go:334] "Generic (PLEG): container finished" podID="07914020-653d-4509-9f60-22726224c7c6" containerID="b0d25fd2c9604e3f96cafaee37ffd660ff8a4f27903a0e0bf9e82ba66554c126" exitCode=137 Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.808621 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-b547848c4-bn5vs" event={"ID":"07914020-653d-4509-9f60-22726224c7c6","Type":"ContainerDied","Data":"b0d25fd2c9604e3f96cafaee37ffd660ff8a4f27903a0e0bf9e82ba66554c126"} Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.808667 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-b547848c4-bn5vs" event={"ID":"07914020-653d-4509-9f60-22726224c7c6","Type":"ContainerDied","Data":"bb01c2352414aa3e5bdfcb4abaaae4c47a152945a1d74d64f5cf1228335558e9"} Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.808682 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-b547848c4-bn5vs" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.861342 4632 scope.go:117] "RemoveContainer" containerID="b0d25fd2c9604e3f96cafaee37ffd660ff8a4f27903a0e0bf9e82ba66554c126" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.880283 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-d856c56c-cmd2q"] Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.899265 4632 scope.go:117] "RemoveContainer" containerID="b0d25fd2c9604e3f96cafaee37ffd660ff8a4f27903a0e0bf9e82ba66554c126" Mar 13 10:28:32 crc kubenswrapper[4632]: E0313 10:28:32.900250 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0d25fd2c9604e3f96cafaee37ffd660ff8a4f27903a0e0bf9e82ba66554c126\": container with ID starting with b0d25fd2c9604e3f96cafaee37ffd660ff8a4f27903a0e0bf9e82ba66554c126 not found: ID does not exist" containerID="b0d25fd2c9604e3f96cafaee37ffd660ff8a4f27903a0e0bf9e82ba66554c126" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.900293 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0d25fd2c9604e3f96cafaee37ffd660ff8a4f27903a0e0bf9e82ba66554c126"} err="failed to get container status \"b0d25fd2c9604e3f96cafaee37ffd660ff8a4f27903a0e0bf9e82ba66554c126\": rpc error: code = NotFound desc = could not find container \"b0d25fd2c9604e3f96cafaee37ffd660ff8a4f27903a0e0bf9e82ba66554c126\": container with ID starting with b0d25fd2c9604e3f96cafaee37ffd660ff8a4f27903a0e0bf9e82ba66554c126 not found: ID does not exist" Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.900674 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-d856c56c-cmd2q"] Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.915330 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-b547848c4-bn5vs"] Mar 13 10:28:32 crc kubenswrapper[4632]: I0313 10:28:32.931798 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-b547848c4-bn5vs"] Mar 13 10:28:34 crc kubenswrapper[4632]: I0313 10:28:34.057299 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07914020-653d-4509-9f60-22726224c7c6" path="/var/lib/kubelet/pods/07914020-653d-4509-9f60-22726224c7c6/volumes" Mar 13 10:28:34 crc kubenswrapper[4632]: I0313 10:28:34.058386 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d10747e-ba77-4986-9d4b-636fcbf823ab" path="/var/lib/kubelet/pods/5d10747e-ba77-4986-9d4b-636fcbf823ab/volumes" Mar 13 10:28:34 crc kubenswrapper[4632]: I0313 10:28:34.271921 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 10:28:34 crc kubenswrapper[4632]: I0313 10:28:34.272672 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 10:28:34 crc kubenswrapper[4632]: I0313 10:28:34.310429 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 10:28:34 crc kubenswrapper[4632]: I0313 10:28:34.328959 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 10:28:34 crc kubenswrapper[4632]: I0313 10:28:34.837593 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 10:28:34 crc kubenswrapper[4632]: I0313 10:28:34.837639 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.396759 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.397409 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.401184 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"2dbb3ede37abc9f5b483ae48b13ac3ed8913ac4529c34c39494b1541e21ce00b"} pod="openstack/horizon-7bdb5f7878-ng2k2" containerMessage="Container horizon failed startup probe, will be restarted" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.401337 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" containerID="cri-o://2dbb3ede37abc9f5b483ae48b13ac3ed8913ac4529c34c39494b1541e21ce00b" gracePeriod=30 Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.439840 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.439891 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.471164 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.487077 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.506594 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.594788 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-scripts\") pod \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.594878 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-sg-core-conf-yaml\") pod \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.594917 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efe2db4a-a5d6-4aa3-805b-5144c66afca8-log-httpd\") pod \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.594951 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efe2db4a-a5d6-4aa3-805b-5144c66afca8-run-httpd\") pod \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.595184 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-config-data\") pod \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.595226 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8nnb\" (UniqueName: \"kubernetes.io/projected/efe2db4a-a5d6-4aa3-805b-5144c66afca8-kube-api-access-b8nnb\") pod \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.595280 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-combined-ca-bundle\") pod \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\" (UID: \"efe2db4a-a5d6-4aa3-805b-5144c66afca8\") " Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.596720 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efe2db4a-a5d6-4aa3-805b-5144c66afca8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "efe2db4a-a5d6-4aa3-805b-5144c66afca8" (UID: "efe2db4a-a5d6-4aa3-805b-5144c66afca8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.597660 4632 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efe2db4a-a5d6-4aa3-805b-5144c66afca8-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.597880 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/efe2db4a-a5d6-4aa3-805b-5144c66afca8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "efe2db4a-a5d6-4aa3-805b-5144c66afca8" (UID: "efe2db4a-a5d6-4aa3-805b-5144c66afca8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.625104 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-scripts" (OuterVolumeSpecName: "scripts") pod "efe2db4a-a5d6-4aa3-805b-5144c66afca8" (UID: "efe2db4a-a5d6-4aa3-805b-5144c66afca8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.629805 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efe2db4a-a5d6-4aa3-805b-5144c66afca8-kube-api-access-b8nnb" (OuterVolumeSpecName: "kube-api-access-b8nnb") pod "efe2db4a-a5d6-4aa3-805b-5144c66afca8" (UID: "efe2db4a-a5d6-4aa3-805b-5144c66afca8"). InnerVolumeSpecName "kube-api-access-b8nnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.668109 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "efe2db4a-a5d6-4aa3-805b-5144c66afca8" (UID: "efe2db4a-a5d6-4aa3-805b-5144c66afca8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.699192 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8nnb\" (UniqueName: \"kubernetes.io/projected/efe2db4a-a5d6-4aa3-805b-5144c66afca8-kube-api-access-b8nnb\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.699223 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.699233 4632 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.699241 4632 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/efe2db4a-a5d6-4aa3-805b-5144c66afca8-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.751233 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-config-data" (OuterVolumeSpecName: "config-data") pod "efe2db4a-a5d6-4aa3-805b-5144c66afca8" (UID: "efe2db4a-a5d6-4aa3-805b-5144c66afca8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.767896 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "efe2db4a-a5d6-4aa3-805b-5144c66afca8" (UID: "efe2db4a-a5d6-4aa3-805b-5144c66afca8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.801381 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.801425 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/efe2db4a-a5d6-4aa3-805b-5144c66afca8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.847847 4632 generic.go:334] "Generic (PLEG): container finished" podID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerID="e1cfa6629f759d5883022ca95d30f701e046142894117805d6775261138f329b" exitCode=0 Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.849088 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.851250 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efe2db4a-a5d6-4aa3-805b-5144c66afca8","Type":"ContainerDied","Data":"e1cfa6629f759d5883022ca95d30f701e046142894117805d6775261138f329b"} Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.851304 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"efe2db4a-a5d6-4aa3-805b-5144c66afca8","Type":"ContainerDied","Data":"14a7a36e446e3ed959c2f673ec8a5768ce0dbc971f3d7879cc430c0491594679"} Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.851327 4632 scope.go:117] "RemoveContainer" containerID="3519acd0a5d01d518140b71f8cb487e29259fada712cb5fe0b79aa9039c08494" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.852619 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.852665 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.858098 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-689764498d-rg7vt" podUID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.882718 4632 scope.go:117] "RemoveContainer" containerID="23c0b66b1c919daecf5753fdf594be220c0bfa625f5256dff7fbdd421ccaaa54" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.913410 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.923829 4632 scope.go:117] "RemoveContainer" containerID="fd5b97322f3966f77a91e26539a007fdcb71ceff62ef4db0eeb167b87655caa1" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.942433 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.969564 4632 scope.go:117] "RemoveContainer" containerID="e1cfa6629f759d5883022ca95d30f701e046142894117805d6775261138f329b" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.971477 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:35 crc kubenswrapper[4632]: E0313 10:28:35.972786 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07914020-653d-4509-9f60-22726224c7c6" containerName="heat-api" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.972813 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="07914020-653d-4509-9f60-22726224c7c6" containerName="heat-api" Mar 13 10:28:35 crc kubenswrapper[4632]: E0313 10:28:35.972856 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d10747e-ba77-4986-9d4b-636fcbf823ab" containerName="heat-cfnapi" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.972868 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d10747e-ba77-4986-9d4b-636fcbf823ab" containerName="heat-cfnapi" Mar 13 10:28:35 crc kubenswrapper[4632]: E0313 10:28:35.972895 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="ceilometer-notification-agent" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.972904 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="ceilometer-notification-agent" Mar 13 10:28:35 crc kubenswrapper[4632]: E0313 10:28:35.972928 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="proxy-httpd" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.972939 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="proxy-httpd" Mar 13 10:28:35 crc kubenswrapper[4632]: E0313 10:28:35.972970 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="ceilometer-central-agent" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.972979 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="ceilometer-central-agent" Mar 13 10:28:35 crc kubenswrapper[4632]: E0313 10:28:35.973009 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="sg-core" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.973019 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="sg-core" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.973317 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="ceilometer-notification-agent" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.973341 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d10747e-ba77-4986-9d4b-636fcbf823ab" containerName="heat-cfnapi" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.973356 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="sg-core" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.978176 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="07914020-653d-4509-9f60-22726224c7c6" containerName="heat-api" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.978255 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="proxy-httpd" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.978271 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" containerName="ceilometer-central-agent" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.981012 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.988787 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 10:28:35 crc kubenswrapper[4632]: I0313 10:28:35.991581 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.009764 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.088832 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efe2db4a-a5d6-4aa3-805b-5144c66afca8" path="/var/lib/kubelet/pods/efe2db4a-a5d6-4aa3-805b-5144c66afca8/volumes" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.109855 4632 scope.go:117] "RemoveContainer" containerID="3519acd0a5d01d518140b71f8cb487e29259fada712cb5fe0b79aa9039c08494" Mar 13 10:28:36 crc kubenswrapper[4632]: E0313 10:28:36.111575 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3519acd0a5d01d518140b71f8cb487e29259fada712cb5fe0b79aa9039c08494\": container with ID starting with 3519acd0a5d01d518140b71f8cb487e29259fada712cb5fe0b79aa9039c08494 not found: ID does not exist" containerID="3519acd0a5d01d518140b71f8cb487e29259fada712cb5fe0b79aa9039c08494" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.111727 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3519acd0a5d01d518140b71f8cb487e29259fada712cb5fe0b79aa9039c08494"} err="failed to get container status \"3519acd0a5d01d518140b71f8cb487e29259fada712cb5fe0b79aa9039c08494\": rpc error: code = NotFound desc = could not find container \"3519acd0a5d01d518140b71f8cb487e29259fada712cb5fe0b79aa9039c08494\": container with ID starting with 3519acd0a5d01d518140b71f8cb487e29259fada712cb5fe0b79aa9039c08494 not found: ID does not exist" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.111771 4632 scope.go:117] "RemoveContainer" containerID="23c0b66b1c919daecf5753fdf594be220c0bfa625f5256dff7fbdd421ccaaa54" Mar 13 10:28:36 crc kubenswrapper[4632]: E0313 10:28:36.112408 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23c0b66b1c919daecf5753fdf594be220c0bfa625f5256dff7fbdd421ccaaa54\": container with ID starting with 23c0b66b1c919daecf5753fdf594be220c0bfa625f5256dff7fbdd421ccaaa54 not found: ID does not exist" containerID="23c0b66b1c919daecf5753fdf594be220c0bfa625f5256dff7fbdd421ccaaa54" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.112527 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23c0b66b1c919daecf5753fdf594be220c0bfa625f5256dff7fbdd421ccaaa54"} err="failed to get container status \"23c0b66b1c919daecf5753fdf594be220c0bfa625f5256dff7fbdd421ccaaa54\": rpc error: code = NotFound desc = could not find container \"23c0b66b1c919daecf5753fdf594be220c0bfa625f5256dff7fbdd421ccaaa54\": container with ID starting with 23c0b66b1c919daecf5753fdf594be220c0bfa625f5256dff7fbdd421ccaaa54 not found: ID does not exist" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.112651 4632 scope.go:117] "RemoveContainer" containerID="fd5b97322f3966f77a91e26539a007fdcb71ceff62ef4db0eeb167b87655caa1" Mar 13 10:28:36 crc kubenswrapper[4632]: E0313 10:28:36.113231 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd5b97322f3966f77a91e26539a007fdcb71ceff62ef4db0eeb167b87655caa1\": container with ID starting with fd5b97322f3966f77a91e26539a007fdcb71ceff62ef4db0eeb167b87655caa1 not found: ID does not exist" containerID="fd5b97322f3966f77a91e26539a007fdcb71ceff62ef4db0eeb167b87655caa1" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.113318 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd5b97322f3966f77a91e26539a007fdcb71ceff62ef4db0eeb167b87655caa1"} err="failed to get container status \"fd5b97322f3966f77a91e26539a007fdcb71ceff62ef4db0eeb167b87655caa1\": rpc error: code = NotFound desc = could not find container \"fd5b97322f3966f77a91e26539a007fdcb71ceff62ef4db0eeb167b87655caa1\": container with ID starting with fd5b97322f3966f77a91e26539a007fdcb71ceff62ef4db0eeb167b87655caa1 not found: ID does not exist" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.113384 4632 scope.go:117] "RemoveContainer" containerID="e1cfa6629f759d5883022ca95d30f701e046142894117805d6775261138f329b" Mar 13 10:28:36 crc kubenswrapper[4632]: E0313 10:28:36.113741 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1cfa6629f759d5883022ca95d30f701e046142894117805d6775261138f329b\": container with ID starting with e1cfa6629f759d5883022ca95d30f701e046142894117805d6775261138f329b not found: ID does not exist" containerID="e1cfa6629f759d5883022ca95d30f701e046142894117805d6775261138f329b" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.113839 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1cfa6629f759d5883022ca95d30f701e046142894117805d6775261138f329b"} err="failed to get container status \"e1cfa6629f759d5883022ca95d30f701e046142894117805d6775261138f329b\": rpc error: code = NotFound desc = could not find container \"e1cfa6629f759d5883022ca95d30f701e046142894117805d6775261138f329b\": container with ID starting with e1cfa6629f759d5883022ca95d30f701e046142894117805d6775261138f329b not found: ID does not exist" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.122592 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.123185 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d6a890-da6f-4a62-a73d-ad22f8b97586-log-httpd\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.123490 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnfcl\" (UniqueName: \"kubernetes.io/projected/92d6a890-da6f-4a62-a73d-ad22f8b97586-kube-api-access-dnfcl\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.123609 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d6a890-da6f-4a62-a73d-ad22f8b97586-run-httpd\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.123652 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-config-data\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.123689 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.123716 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-scripts\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.224953 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d6a890-da6f-4a62-a73d-ad22f8b97586-log-httpd\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.225044 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnfcl\" (UniqueName: \"kubernetes.io/projected/92d6a890-da6f-4a62-a73d-ad22f8b97586-kube-api-access-dnfcl\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.225086 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d6a890-da6f-4a62-a73d-ad22f8b97586-run-httpd\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.225112 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-config-data\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.225133 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.225155 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-scripts\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.225179 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.226806 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d6a890-da6f-4a62-a73d-ad22f8b97586-run-httpd\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.227252 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d6a890-da6f-4a62-a73d-ad22f8b97586-log-httpd\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.238380 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-scripts\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.238840 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-config-data\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.248910 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.252434 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.259046 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnfcl\" (UniqueName: \"kubernetes.io/projected/92d6a890-da6f-4a62-a73d-ad22f8b97586-kube-api-access-dnfcl\") pod \"ceilometer-0\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.408580 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.872319 4632 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 10:28:36 crc kubenswrapper[4632]: I0313 10:28:36.873279 4632 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 10:28:37 crc kubenswrapper[4632]: I0313 10:28:37.089712 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:37 crc kubenswrapper[4632]: W0313 10:28:37.096782 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92d6a890_da6f_4a62_a73d_ad22f8b97586.slice/crio-61eb61c712456be9b4257c5e2ea6a70dfbfca01a50f8412f9ea8b2cdb5c8b498 WatchSource:0}: Error finding container 61eb61c712456be9b4257c5e2ea6a70dfbfca01a50f8412f9ea8b2cdb5c8b498: Status 404 returned error can't find the container with id 61eb61c712456be9b4257c5e2ea6a70dfbfca01a50f8412f9ea8b2cdb5c8b498 Mar 13 10:28:37 crc kubenswrapper[4632]: I0313 10:28:37.887167 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d6a890-da6f-4a62-a73d-ad22f8b97586","Type":"ContainerStarted","Data":"13beea25c7ec581a71ff8aed4dcb89b5326c0045c02a48578bb1e384a8c92d16"} Mar 13 10:28:37 crc kubenswrapper[4632]: I0313 10:28:37.887554 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d6a890-da6f-4a62-a73d-ad22f8b97586","Type":"ContainerStarted","Data":"61eb61c712456be9b4257c5e2ea6a70dfbfca01a50f8412f9ea8b2cdb5c8b498"} Mar 13 10:28:38 crc kubenswrapper[4632]: I0313 10:28:38.903347 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d6a890-da6f-4a62-a73d-ad22f8b97586","Type":"ContainerStarted","Data":"a78858d52f4edb9f1b215cb0b5d9d5d059b8c3bfd31b64cf5e0deaf6ab27d4b4"} Mar 13 10:28:39 crc kubenswrapper[4632]: I0313 10:28:39.553845 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 10:28:39 crc kubenswrapper[4632]: I0313 10:28:39.553956 4632 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 10:28:39 crc kubenswrapper[4632]: I0313 10:28:39.807234 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 10:28:39 crc kubenswrapper[4632]: I0313 10:28:39.807661 4632 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 10:28:39 crc kubenswrapper[4632]: I0313 10:28:39.914499 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d6a890-da6f-4a62-a73d-ad22f8b97586","Type":"ContainerStarted","Data":"c2c74a9428ab3dfaa995d259e75dccffb44018988a076ef192a947b75ff6a7f1"} Mar 13 10:28:39 crc kubenswrapper[4632]: I0313 10:28:39.916855 4632 generic.go:334] "Generic (PLEG): container finished" podID="5de81924-9bfc-484e-8276-0216f0bbf72c" containerID="afb05bb00debb2ea4a81d169362ff2bd38d824053184e249dbe02cc1cb10e945" exitCode=0 Mar 13 10:28:39 crc kubenswrapper[4632]: I0313 10:28:39.916898 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5mlm2" event={"ID":"5de81924-9bfc-484e-8276-0216f0bbf72c","Type":"ContainerDied","Data":"afb05bb00debb2ea4a81d169362ff2bd38d824053184e249dbe02cc1cb10e945"} Mar 13 10:28:40 crc kubenswrapper[4632]: I0313 10:28:40.025834 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 10:28:40 crc kubenswrapper[4632]: I0313 10:28:40.790023 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.535454 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.660488 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-config-data\") pod \"5de81924-9bfc-484e-8276-0216f0bbf72c\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.661026 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-scripts\") pod \"5de81924-9bfc-484e-8276-0216f0bbf72c\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.661103 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg9bb\" (UniqueName: \"kubernetes.io/projected/5de81924-9bfc-484e-8276-0216f0bbf72c-kube-api-access-bg9bb\") pod \"5de81924-9bfc-484e-8276-0216f0bbf72c\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.661225 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-combined-ca-bundle\") pod \"5de81924-9bfc-484e-8276-0216f0bbf72c\" (UID: \"5de81924-9bfc-484e-8276-0216f0bbf72c\") " Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.676718 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-scripts" (OuterVolumeSpecName: "scripts") pod "5de81924-9bfc-484e-8276-0216f0bbf72c" (UID: "5de81924-9bfc-484e-8276-0216f0bbf72c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.684273 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5de81924-9bfc-484e-8276-0216f0bbf72c-kube-api-access-bg9bb" (OuterVolumeSpecName: "kube-api-access-bg9bb") pod "5de81924-9bfc-484e-8276-0216f0bbf72c" (UID: "5de81924-9bfc-484e-8276-0216f0bbf72c"). InnerVolumeSpecName "kube-api-access-bg9bb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.765495 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.765537 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bg9bb\" (UniqueName: \"kubernetes.io/projected/5de81924-9bfc-484e-8276-0216f0bbf72c-kube-api-access-bg9bb\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.786238 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5de81924-9bfc-484e-8276-0216f0bbf72c" (UID: "5de81924-9bfc-484e-8276-0216f0bbf72c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.797148 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-config-data" (OuterVolumeSpecName: "config-data") pod "5de81924-9bfc-484e-8276-0216f0bbf72c" (UID: "5de81924-9bfc-484e-8276-0216f0bbf72c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.867451 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.867501 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5de81924-9bfc-484e-8276-0216f0bbf72c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.943869 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-5mlm2" event={"ID":"5de81924-9bfc-484e-8276-0216f0bbf72c","Type":"ContainerDied","Data":"951453764e4945cebf39ce493b8227004659c89412dc1b5f0146d76b115b3607"} Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.943914 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="951453764e4945cebf39ce493b8227004659c89412dc1b5f0146d76b115b3607" Mar 13 10:28:41 crc kubenswrapper[4632]: I0313 10:28:41.944004 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-5mlm2" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.126043 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 10:28:42 crc kubenswrapper[4632]: E0313 10:28:42.126527 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5de81924-9bfc-484e-8276-0216f0bbf72c" containerName="nova-cell0-conductor-db-sync" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.126555 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="5de81924-9bfc-484e-8276-0216f0bbf72c" containerName="nova-cell0-conductor-db-sync" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.126784 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="5de81924-9bfc-484e-8276-0216f0bbf72c" containerName="nova-cell0-conductor-db-sync" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.127583 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.133698 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-h4qk2" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.134420 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.154788 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.191215 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.191360 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.191409 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjsd4\" (UniqueName: \"kubernetes.io/projected/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-kube-api-access-pjsd4\") pod \"nova-cell0-conductor-0\" (UID: \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.295131 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjsd4\" (UniqueName: \"kubernetes.io/projected/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-kube-api-access-pjsd4\") pod \"nova-cell0-conductor-0\" (UID: \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.296234 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.296375 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.304236 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.304738 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.318563 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjsd4\" (UniqueName: \"kubernetes.io/projected/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-kube-api-access-pjsd4\") pod \"nova-cell0-conductor-0\" (UID: \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.463411 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.960782 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d6a890-da6f-4a62-a73d-ad22f8b97586","Type":"ContainerStarted","Data":"07674d77dbcfb4d04e610536847653ba6a156f4e167fb3e30be00823bd80251e"} Mar 13 10:28:42 crc kubenswrapper[4632]: I0313 10:28:42.961416 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 10:28:43 crc kubenswrapper[4632]: I0313 10:28:43.134304 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.5756441839999997 podStartE2EDuration="8.134280978s" podCreationTimestamp="2026-03-13 10:28:35 +0000 UTC" firstStartedPulling="2026-03-13 10:28:37.105162837 +0000 UTC m=+1491.127692970" lastFinishedPulling="2026-03-13 10:28:41.663799631 +0000 UTC m=+1495.686329764" observedRunningTime="2026-03-13 10:28:42.988585934 +0000 UTC m=+1497.011116097" watchObservedRunningTime="2026-03-13 10:28:43.134280978 +0000 UTC m=+1497.156811141" Mar 13 10:28:43 crc kubenswrapper[4632]: I0313 10:28:43.135014 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 10:28:43 crc kubenswrapper[4632]: W0313 10:28:43.135852 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89c5451e_248e_46eb_ac20_f52c3e3bcdc4.slice/crio-fea8b62da5fff833a90864e9fa4a28877f40e3642c9c75596310ee934707e980 WatchSource:0}: Error finding container fea8b62da5fff833a90864e9fa4a28877f40e3642c9c75596310ee934707e980: Status 404 returned error can't find the container with id fea8b62da5fff833a90864e9fa4a28877f40e3642c9c75596310ee934707e980 Mar 13 10:28:43 crc kubenswrapper[4632]: I0313 10:28:43.971715 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"89c5451e-248e-46eb-ac20-f52c3e3bcdc4","Type":"ContainerStarted","Data":"b6bf072e344c147c11f620ba388aa0850202403d9ea7d55387c6d06560823aa8"} Mar 13 10:28:43 crc kubenswrapper[4632]: I0313 10:28:43.972349 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"89c5451e-248e-46eb-ac20-f52c3e3bcdc4","Type":"ContainerStarted","Data":"fea8b62da5fff833a90864e9fa4a28877f40e3642c9c75596310ee934707e980"} Mar 13 10:28:43 crc kubenswrapper[4632]: I0313 10:28:43.972369 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 13 10:28:45 crc kubenswrapper[4632]: I0313 10:28:45.857960 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-689764498d-rg7vt" podUID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 13 10:28:46 crc kubenswrapper[4632]: I0313 10:28:46.912561 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=4.912536629 podStartE2EDuration="4.912536629s" podCreationTimestamp="2026-03-13 10:28:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:28:43.999388511 +0000 UTC m=+1498.021918654" watchObservedRunningTime="2026-03-13 10:28:46.912536629 +0000 UTC m=+1500.935066762" Mar 13 10:28:46 crc kubenswrapper[4632]: I0313 10:28:46.914850 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-n4z22"] Mar 13 10:28:46 crc kubenswrapper[4632]: I0313 10:28:46.916829 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:28:46 crc kubenswrapper[4632]: I0313 10:28:46.928646 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n4z22"] Mar 13 10:28:46 crc kubenswrapper[4632]: I0313 10:28:46.997612 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpwh9\" (UniqueName: \"kubernetes.io/projected/d0cabd29-ef3e-4808-8c92-3b032483789e-kube-api-access-xpwh9\") pod \"redhat-operators-n4z22\" (UID: \"d0cabd29-ef3e-4808-8c92-3b032483789e\") " pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:28:46 crc kubenswrapper[4632]: I0313 10:28:46.997784 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0cabd29-ef3e-4808-8c92-3b032483789e-utilities\") pod \"redhat-operators-n4z22\" (UID: \"d0cabd29-ef3e-4808-8c92-3b032483789e\") " pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:28:46 crc kubenswrapper[4632]: I0313 10:28:46.997853 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0cabd29-ef3e-4808-8c92-3b032483789e-catalog-content\") pod \"redhat-operators-n4z22\" (UID: \"d0cabd29-ef3e-4808-8c92-3b032483789e\") " pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:28:47 crc kubenswrapper[4632]: I0313 10:28:47.100080 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpwh9\" (UniqueName: \"kubernetes.io/projected/d0cabd29-ef3e-4808-8c92-3b032483789e-kube-api-access-xpwh9\") pod \"redhat-operators-n4z22\" (UID: \"d0cabd29-ef3e-4808-8c92-3b032483789e\") " pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:28:47 crc kubenswrapper[4632]: I0313 10:28:47.100167 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0cabd29-ef3e-4808-8c92-3b032483789e-utilities\") pod \"redhat-operators-n4z22\" (UID: \"d0cabd29-ef3e-4808-8c92-3b032483789e\") " pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:28:47 crc kubenswrapper[4632]: I0313 10:28:47.100256 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0cabd29-ef3e-4808-8c92-3b032483789e-catalog-content\") pod \"redhat-operators-n4z22\" (UID: \"d0cabd29-ef3e-4808-8c92-3b032483789e\") " pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:28:47 crc kubenswrapper[4632]: I0313 10:28:47.101642 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0cabd29-ef3e-4808-8c92-3b032483789e-catalog-content\") pod \"redhat-operators-n4z22\" (UID: \"d0cabd29-ef3e-4808-8c92-3b032483789e\") " pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:28:47 crc kubenswrapper[4632]: I0313 10:28:47.101682 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0cabd29-ef3e-4808-8c92-3b032483789e-utilities\") pod \"redhat-operators-n4z22\" (UID: \"d0cabd29-ef3e-4808-8c92-3b032483789e\") " pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:28:47 crc kubenswrapper[4632]: I0313 10:28:47.102806 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:28:47 crc kubenswrapper[4632]: I0313 10:28:47.103148 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="ceilometer-central-agent" containerID="cri-o://13beea25c7ec581a71ff8aed4dcb89b5326c0045c02a48578bb1e384a8c92d16" gracePeriod=30 Mar 13 10:28:47 crc kubenswrapper[4632]: I0313 10:28:47.103292 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="proxy-httpd" containerID="cri-o://07674d77dbcfb4d04e610536847653ba6a156f4e167fb3e30be00823bd80251e" gracePeriod=30 Mar 13 10:28:47 crc kubenswrapper[4632]: I0313 10:28:47.103342 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="sg-core" containerID="cri-o://c2c74a9428ab3dfaa995d259e75dccffb44018988a076ef192a947b75ff6a7f1" gracePeriod=30 Mar 13 10:28:47 crc kubenswrapper[4632]: I0313 10:28:47.103394 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="ceilometer-notification-agent" containerID="cri-o://a78858d52f4edb9f1b215cb0b5d9d5d059b8c3bfd31b64cf5e0deaf6ab27d4b4" gracePeriod=30 Mar 13 10:28:47 crc kubenswrapper[4632]: I0313 10:28:47.134921 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpwh9\" (UniqueName: \"kubernetes.io/projected/d0cabd29-ef3e-4808-8c92-3b032483789e-kube-api-access-xpwh9\") pod \"redhat-operators-n4z22\" (UID: \"d0cabd29-ef3e-4808-8c92-3b032483789e\") " pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:28:47 crc kubenswrapper[4632]: I0313 10:28:47.247407 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:28:47 crc kubenswrapper[4632]: W0313 10:28:47.824446 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0cabd29_ef3e_4808_8c92_3b032483789e.slice/crio-dd0fe42db5b99209dcd168810b0996ceb728a9055a395258ce5d2c5e8afe18b9 WatchSource:0}: Error finding container dd0fe42db5b99209dcd168810b0996ceb728a9055a395258ce5d2c5e8afe18b9: Status 404 returned error can't find the container with id dd0fe42db5b99209dcd168810b0996ceb728a9055a395258ce5d2c5e8afe18b9 Mar 13 10:28:47 crc kubenswrapper[4632]: I0313 10:28:47.825704 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-n4z22"] Mar 13 10:28:48 crc kubenswrapper[4632]: I0313 10:28:48.092665 4632 generic.go:334] "Generic (PLEG): container finished" podID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerID="07674d77dbcfb4d04e610536847653ba6a156f4e167fb3e30be00823bd80251e" exitCode=0 Mar 13 10:28:48 crc kubenswrapper[4632]: I0313 10:28:48.095656 4632 generic.go:334] "Generic (PLEG): container finished" podID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerID="c2c74a9428ab3dfaa995d259e75dccffb44018988a076ef192a947b75ff6a7f1" exitCode=2 Mar 13 10:28:48 crc kubenswrapper[4632]: I0313 10:28:48.095787 4632 generic.go:334] "Generic (PLEG): container finished" podID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerID="a78858d52f4edb9f1b215cb0b5d9d5d059b8c3bfd31b64cf5e0deaf6ab27d4b4" exitCode=0 Mar 13 10:28:48 crc kubenswrapper[4632]: I0313 10:28:48.093671 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d6a890-da6f-4a62-a73d-ad22f8b97586","Type":"ContainerDied","Data":"07674d77dbcfb4d04e610536847653ba6a156f4e167fb3e30be00823bd80251e"} Mar 13 10:28:48 crc kubenswrapper[4632]: I0313 10:28:48.096314 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d6a890-da6f-4a62-a73d-ad22f8b97586","Type":"ContainerDied","Data":"c2c74a9428ab3dfaa995d259e75dccffb44018988a076ef192a947b75ff6a7f1"} Mar 13 10:28:48 crc kubenswrapper[4632]: I0313 10:28:48.096421 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d6a890-da6f-4a62-a73d-ad22f8b97586","Type":"ContainerDied","Data":"a78858d52f4edb9f1b215cb0b5d9d5d059b8c3bfd31b64cf5e0deaf6ab27d4b4"} Mar 13 10:28:48 crc kubenswrapper[4632]: I0313 10:28:48.112339 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n4z22" event={"ID":"d0cabd29-ef3e-4808-8c92-3b032483789e","Type":"ContainerStarted","Data":"dd0fe42db5b99209dcd168810b0996ceb728a9055a395258ce5d2c5e8afe18b9"} Mar 13 10:28:49 crc kubenswrapper[4632]: I0313 10:28:49.127089 4632 generic.go:334] "Generic (PLEG): container finished" podID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerID="201e16e880cd2d90fce52e678a22a53fb8633269541a12fab257e648ab87ec2d" exitCode=0 Mar 13 10:28:49 crc kubenswrapper[4632]: I0313 10:28:49.127225 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n4z22" event={"ID":"d0cabd29-ef3e-4808-8c92-3b032483789e","Type":"ContainerDied","Data":"201e16e880cd2d90fce52e678a22a53fb8633269541a12fab257e648ab87ec2d"} Mar 13 10:28:50 crc kubenswrapper[4632]: I0313 10:28:50.144621 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n4z22" event={"ID":"d0cabd29-ef3e-4808-8c92-3b032483789e","Type":"ContainerStarted","Data":"71596437e6a2d325baac09e68c289bacf1de16594d0329b6a19a5f5d92099872"} Mar 13 10:28:52 crc kubenswrapper[4632]: I0313 10:28:52.506910 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.303146 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-gwj5n"] Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.304431 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.313572 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.314272 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.327402 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-gwj5n"] Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.391156 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-scripts\") pod \"nova-cell0-cell-mapping-gwj5n\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.391280 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxd74\" (UniqueName: \"kubernetes.io/projected/bcce9343-52a3-4e6d-98fd-8e66390020ac-kube-api-access-sxd74\") pod \"nova-cell0-cell-mapping-gwj5n\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.391325 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gwj5n\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.391358 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-config-data\") pod \"nova-cell0-cell-mapping-gwj5n\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.495767 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxd74\" (UniqueName: \"kubernetes.io/projected/bcce9343-52a3-4e6d-98fd-8e66390020ac-kube-api-access-sxd74\") pod \"nova-cell0-cell-mapping-gwj5n\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.495849 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gwj5n\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.495908 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-config-data\") pod \"nova-cell0-cell-mapping-gwj5n\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.496056 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-scripts\") pod \"nova-cell0-cell-mapping-gwj5n\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.510894 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.511210 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-gwj5n\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.514829 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.531454 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.541624 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-scripts\") pod \"nova-cell0-cell-mapping-gwj5n\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.562979 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.564990 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-config-data\") pod \"nova-cell0-cell-mapping-gwj5n\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.599167 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9h2v\" (UniqueName: \"kubernetes.io/projected/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-kube-api-access-g9h2v\") pod \"nova-api-0\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " pod="openstack/nova-api-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.599234 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-logs\") pod \"nova-api-0\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " pod="openstack/nova-api-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.599460 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " pod="openstack/nova-api-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.599491 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-config-data\") pod \"nova-api-0\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " pod="openstack/nova-api-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.604082 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxd74\" (UniqueName: \"kubernetes.io/projected/bcce9343-52a3-4e6d-98fd-8e66390020ac-kube-api-access-sxd74\") pod \"nova-cell0-cell-mapping-gwj5n\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.634321 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.666341 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.676698 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.705681 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9h2v\" (UniqueName: \"kubernetes.io/projected/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-kube-api-access-g9h2v\") pod \"nova-api-0\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " pod="openstack/nova-api-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.705748 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-logs\") pod \"nova-api-0\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " pod="openstack/nova-api-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.705825 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/229b0823-7d97-48cf-9b38-188a3f4ecde3-config-data\") pod \"nova-metadata-0\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " pod="openstack/nova-metadata-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.705858 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/229b0823-7d97-48cf-9b38-188a3f4ecde3-logs\") pod \"nova-metadata-0\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " pod="openstack/nova-metadata-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.705896 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7cq6\" (UniqueName: \"kubernetes.io/projected/229b0823-7d97-48cf-9b38-188a3f4ecde3-kube-api-access-c7cq6\") pod \"nova-metadata-0\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " pod="openstack/nova-metadata-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.705989 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229b0823-7d97-48cf-9b38-188a3f4ecde3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " pod="openstack/nova-metadata-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.706074 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " pod="openstack/nova-api-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.706104 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-config-data\") pod \"nova-api-0\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " pod="openstack/nova-api-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.712251 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-logs\") pod \"nova-api-0\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " pod="openstack/nova-api-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.722095 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.742062 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-config-data\") pod \"nova-api-0\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " pod="openstack/nova-api-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.743490 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " pod="openstack/nova-api-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.789391 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.810165 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229b0823-7d97-48cf-9b38-188a3f4ecde3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " pod="openstack/nova-metadata-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.810653 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/229b0823-7d97-48cf-9b38-188a3f4ecde3-config-data\") pod \"nova-metadata-0\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " pod="openstack/nova-metadata-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.811536 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/229b0823-7d97-48cf-9b38-188a3f4ecde3-logs\") pod \"nova-metadata-0\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " pod="openstack/nova-metadata-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.811703 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7cq6\" (UniqueName: \"kubernetes.io/projected/229b0823-7d97-48cf-9b38-188a3f4ecde3-kube-api-access-c7cq6\") pod \"nova-metadata-0\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " pod="openstack/nova-metadata-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.813473 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/229b0823-7d97-48cf-9b38-188a3f4ecde3-logs\") pod \"nova-metadata-0\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " pod="openstack/nova-metadata-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.830160 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/229b0823-7d97-48cf-9b38-188a3f4ecde3-config-data\") pod \"nova-metadata-0\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " pod="openstack/nova-metadata-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.854314 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229b0823-7d97-48cf-9b38-188a3f4ecde3-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " pod="openstack/nova-metadata-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.893700 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9h2v\" (UniqueName: \"kubernetes.io/projected/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-kube-api-access-g9h2v\") pod \"nova-api-0\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " pod="openstack/nova-api-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.904881 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7cq6\" (UniqueName: \"kubernetes.io/projected/229b0823-7d97-48cf-9b38-188a3f4ecde3-kube-api-access-c7cq6\") pod \"nova-metadata-0\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " pod="openstack/nova-metadata-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.914740 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.942105 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.943360 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.958651 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.958876 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 10:28:53 crc kubenswrapper[4632]: I0313 10:28:53.977064 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.104032 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3548645f-4c72-4c75-b1bd-95116d47f6e2-config-data\") pod \"nova-scheduler-0\" (UID: \"3548645f-4c72-4c75-b1bd-95116d47f6e2\") " pod="openstack/nova-scheduler-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.104469 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tntfg\" (UniqueName: \"kubernetes.io/projected/3548645f-4c72-4c75-b1bd-95116d47f6e2-kube-api-access-tntfg\") pod \"nova-scheduler-0\" (UID: \"3548645f-4c72-4c75-b1bd-95116d47f6e2\") " pod="openstack/nova-scheduler-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.104570 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3548645f-4c72-4c75-b1bd-95116d47f6e2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3548645f-4c72-4c75-b1bd-95116d47f6e2\") " pod="openstack/nova-scheduler-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.212843 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tntfg\" (UniqueName: \"kubernetes.io/projected/3548645f-4c72-4c75-b1bd-95116d47f6e2-kube-api-access-tntfg\") pod \"nova-scheduler-0\" (UID: \"3548645f-4c72-4c75-b1bd-95116d47f6e2\") " pod="openstack/nova-scheduler-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.213279 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3548645f-4c72-4c75-b1bd-95116d47f6e2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3548645f-4c72-4c75-b1bd-95116d47f6e2\") " pod="openstack/nova-scheduler-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.213464 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3548645f-4c72-4c75-b1bd-95116d47f6e2-config-data\") pod \"nova-scheduler-0\" (UID: \"3548645f-4c72-4c75-b1bd-95116d47f6e2\") " pod="openstack/nova-scheduler-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.259053 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3548645f-4c72-4c75-b1bd-95116d47f6e2-config-data\") pod \"nova-scheduler-0\" (UID: \"3548645f-4c72-4c75-b1bd-95116d47f6e2\") " pod="openstack/nova-scheduler-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.285477 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tntfg\" (UniqueName: \"kubernetes.io/projected/3548645f-4c72-4c75-b1bd-95116d47f6e2-kube-api-access-tntfg\") pod \"nova-scheduler-0\" (UID: \"3548645f-4c72-4c75-b1bd-95116d47f6e2\") " pod="openstack/nova-scheduler-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.285992 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3548645f-4c72-4c75-b1bd-95116d47f6e2-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"3548645f-4c72-4c75-b1bd-95116d47f6e2\") " pod="openstack/nova-scheduler-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.350171 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.420323 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.422859 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75788dd97c-r8qnr"] Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.424079 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.424117 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75788dd97c-r8qnr"] Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.424196 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.426471 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.434020 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.529369 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-dns-svc\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.531180 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-ovsdbserver-nb\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.531318 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-config\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.531451 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-dns-swift-storage-0\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.531603 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.532330 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlgdp\" (UniqueName: \"kubernetes.io/projected/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-kube-api-access-hlgdp\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.532455 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.532799 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-ovsdbserver-sb\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.532986 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc7jm\" (UniqueName: \"kubernetes.io/projected/e0f17959-fde8-4cf1-b255-db5fc3325b70-kube-api-access-pc7jm\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.640095 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-ovsdbserver-sb\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.640162 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc7jm\" (UniqueName: \"kubernetes.io/projected/e0f17959-fde8-4cf1-b255-db5fc3325b70-kube-api-access-pc7jm\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.640210 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-dns-svc\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.640259 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-ovsdbserver-nb\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.640299 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-config\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.640353 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-dns-swift-storage-0\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.640410 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.640440 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlgdp\" (UniqueName: \"kubernetes.io/projected/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-kube-api-access-hlgdp\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.640500 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.652433 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-ovsdbserver-sb\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.668589 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-dns-svc\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.669728 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-ovsdbserver-nb\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.670658 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-config\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.689682 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-dns-swift-storage-0\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.720451 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.747671 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.756783 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlgdp\" (UniqueName: \"kubernetes.io/projected/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-kube-api-access-hlgdp\") pod \"nova-cell1-novncproxy-0\" (UID: \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.773636 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc7jm\" (UniqueName: \"kubernetes.io/projected/e0f17959-fde8-4cf1-b255-db5fc3325b70-kube-api-access-pc7jm\") pod \"dnsmasq-dns-75788dd97c-r8qnr\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.907404 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:54 crc kubenswrapper[4632]: I0313 10:28:54.915993 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:28:55 crc kubenswrapper[4632]: I0313 10:28:55.199916 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-gwj5n"] Mar 13 10:28:55 crc kubenswrapper[4632]: I0313 10:28:55.235047 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:28:55 crc kubenswrapper[4632]: I0313 10:28:55.252425 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gwj5n" event={"ID":"bcce9343-52a3-4e6d-98fd-8e66390020ac","Type":"ContainerStarted","Data":"f930aa1f0069e1fe78556089c981d58cc3cf6a82579a76e645212ffad42b673e"} Mar 13 10:28:55 crc kubenswrapper[4632]: I0313 10:28:55.274097 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:28:55 crc kubenswrapper[4632]: I0313 10:28:55.583272 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:28:56 crc kubenswrapper[4632]: I0313 10:28:56.189042 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75788dd97c-r8qnr"] Mar 13 10:28:56 crc kubenswrapper[4632]: I0313 10:28:56.239740 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 10:28:56 crc kubenswrapper[4632]: I0313 10:28:56.309773 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" event={"ID":"e0f17959-fde8-4cf1-b255-db5fc3325b70","Type":"ContainerStarted","Data":"a2b2fdcf6ef7efc2eca17a814eb5b4394c29b09fe6419666d04ee4759d7660a8"} Mar 13 10:28:56 crc kubenswrapper[4632]: I0313 10:28:56.348305 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ac666e99-860b-4f76-8b34-0ac5d3f67e9e","Type":"ContainerStarted","Data":"21ad142ba62995b1f6903534cfcf5ce85e20fd10841d57b3365ce522bedc21e5"} Mar 13 10:28:56 crc kubenswrapper[4632]: I0313 10:28:56.374297 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3548645f-4c72-4c75-b1bd-95116d47f6e2","Type":"ContainerStarted","Data":"2e867316abee7ad3f39e6bce10fac89a2148bd2f0cb280bc0cfe2baf59687b11"} Mar 13 10:28:56 crc kubenswrapper[4632]: I0313 10:28:56.390261 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"229b0823-7d97-48cf-9b38-188a3f4ecde3","Type":"ContainerStarted","Data":"e09e0609bf87522d60a55be578e8526df6a87fdb2167dcc1dc7feca3fdedd742"} Mar 13 10:28:56 crc kubenswrapper[4632]: I0313 10:28:56.404813 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998","Type":"ContainerStarted","Data":"34daac34beb9c4f24aabab024257734fad48bd622c9a19af69564a5c1af316f2"} Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.388443 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9n7gj"] Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.390036 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.394208 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.394468 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.405882 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9n7gj"] Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.407966 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cpkl\" (UniqueName: \"kubernetes.io/projected/cf19672e-3284-49bc-a460-f2e629881d9b-kube-api-access-2cpkl\") pod \"nova-cell1-conductor-db-sync-9n7gj\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.408114 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9n7gj\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.408138 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-config-data\") pod \"nova-cell1-conductor-db-sync-9n7gj\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.408169 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-scripts\") pod \"nova-cell1-conductor-db-sync-9n7gj\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.483643 4632 generic.go:334] "Generic (PLEG): container finished" podID="e0f17959-fde8-4cf1-b255-db5fc3325b70" containerID="c6848744dc1fd449bb0df7b7ca2c04941331806f97abf20c11372e120fb30d31" exitCode=0 Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.483722 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" event={"ID":"e0f17959-fde8-4cf1-b255-db5fc3325b70","Type":"ContainerDied","Data":"c6848744dc1fd449bb0df7b7ca2c04941331806f97abf20c11372e120fb30d31"} Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.515888 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cpkl\" (UniqueName: \"kubernetes.io/projected/cf19672e-3284-49bc-a460-f2e629881d9b-kube-api-access-2cpkl\") pod \"nova-cell1-conductor-db-sync-9n7gj\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.516290 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9n7gj\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.516325 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-config-data\") pod \"nova-cell1-conductor-db-sync-9n7gj\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.516384 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-scripts\") pod \"nova-cell1-conductor-db-sync-9n7gj\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.518609 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gwj5n" event={"ID":"bcce9343-52a3-4e6d-98fd-8e66390020ac","Type":"ContainerStarted","Data":"379356ecac878a5f4776d015be267e8c7eec62c977ce924abd53ff44455ce8e4"} Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.552421 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-gwj5n" podStartSLOduration=4.552392671 podStartE2EDuration="4.552392671s" podCreationTimestamp="2026-03-13 10:28:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:28:57.536418082 +0000 UTC m=+1511.558948225" watchObservedRunningTime="2026-03-13 10:28:57.552392671 +0000 UTC m=+1511.574922804" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.566202 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-9n7gj\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.569254 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cpkl\" (UniqueName: \"kubernetes.io/projected/cf19672e-3284-49bc-a460-f2e629881d9b-kube-api-access-2cpkl\") pod \"nova-cell1-conductor-db-sync-9n7gj\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.580687 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-scripts\") pod \"nova-cell1-conductor-db-sync-9n7gj\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.581064 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-config-data\") pod \"nova-cell1-conductor-db-sync-9n7gj\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:28:57 crc kubenswrapper[4632]: I0313 10:28:57.767770 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:28:58 crc kubenswrapper[4632]: I0313 10:28:58.565966 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" event={"ID":"e0f17959-fde8-4cf1-b255-db5fc3325b70","Type":"ContainerStarted","Data":"f957b291649cd64b5f0c12f7a4a8a32abd88e0067f00c5ae80a3e106aedde5a8"} Mar 13 10:28:58 crc kubenswrapper[4632]: I0313 10:28:58.566537 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:28:58 crc kubenswrapper[4632]: I0313 10:28:58.594223 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" podStartSLOduration=5.594198242 podStartE2EDuration="5.594198242s" podCreationTimestamp="2026-03-13 10:28:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:28:58.586365671 +0000 UTC m=+1512.608895834" watchObservedRunningTime="2026-03-13 10:28:58.594198242 +0000 UTC m=+1512.616728385" Mar 13 10:28:58 crc kubenswrapper[4632]: I0313 10:28:58.733256 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9n7gj"] Mar 13 10:28:59 crc kubenswrapper[4632]: I0313 10:28:59.534169 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:28:59 crc kubenswrapper[4632]: I0313 10:28:59.548530 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 10:28:59 crc kubenswrapper[4632]: I0313 10:28:59.589442 4632 generic.go:334] "Generic (PLEG): container finished" podID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerID="71596437e6a2d325baac09e68c289bacf1de16594d0329b6a19a5f5d92099872" exitCode=0 Mar 13 10:28:59 crc kubenswrapper[4632]: I0313 10:28:59.589524 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n4z22" event={"ID":"d0cabd29-ef3e-4808-8c92-3b032483789e","Type":"ContainerDied","Data":"71596437e6a2d325baac09e68c289bacf1de16594d0329b6a19a5f5d92099872"} Mar 13 10:28:59 crc kubenswrapper[4632]: I0313 10:28:59.597426 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9n7gj" event={"ID":"cf19672e-3284-49bc-a460-f2e629881d9b","Type":"ContainerStarted","Data":"a8f98d9cfd7da7677c0fe463edd081d6aa2858ecb1027917673862b2700f1545"} Mar 13 10:28:59 crc kubenswrapper[4632]: I0313 10:28:59.597477 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9n7gj" event={"ID":"cf19672e-3284-49bc-a460-f2e629881d9b","Type":"ContainerStarted","Data":"2552baf18c3652ea7e85b1cf98a826d4538b2dbd01aa4210322514f06f9b9c99"} Mar 13 10:28:59 crc kubenswrapper[4632]: I0313 10:28:59.639748 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-9n7gj" podStartSLOduration=2.6397271719999997 podStartE2EDuration="2.639727172s" podCreationTimestamp="2026-03-13 10:28:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:28:59.636856482 +0000 UTC m=+1513.659386615" watchObservedRunningTime="2026-03-13 10:28:59.639727172 +0000 UTC m=+1513.662257315" Mar 13 10:29:00 crc kubenswrapper[4632]: I0313 10:29:00.866150 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-689764498d-rg7vt" podUID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:00 crc kubenswrapper[4632]: I0313 10:29:00.866596 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:29:00 crc kubenswrapper[4632]: I0313 10:29:00.868009 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"26a7aae686bb479cfcbc8b01e8e10e3fd467e5236d6ffb2ed638373687267401"} pod="openstack/horizon-689764498d-rg7vt" containerMessage="Container horizon failed startup probe, will be restarted" Mar 13 10:29:00 crc kubenswrapper[4632]: I0313 10:29:00.868052 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-689764498d-rg7vt" podUID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerName="horizon" containerID="cri-o://26a7aae686bb479cfcbc8b01e8e10e3fd467e5236d6ffb2ed638373687267401" gracePeriod=30 Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.671019 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"229b0823-7d97-48cf-9b38-188a3f4ecde3","Type":"ContainerStarted","Data":"c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5"} Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.671809 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"229b0823-7d97-48cf-9b38-188a3f4ecde3","Type":"ContainerStarted","Data":"9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2"} Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.671338 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="229b0823-7d97-48cf-9b38-188a3f4ecde3" containerName="nova-metadata-metadata" containerID="cri-o://c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5" gracePeriod=30 Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.671256 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="229b0823-7d97-48cf-9b38-188a3f4ecde3" containerName="nova-metadata-log" containerID="cri-o://9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2" gracePeriod=30 Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.678171 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998","Type":"ContainerStarted","Data":"29733198b90e2c4e70c2780c9fc89c7f6fe69e2a7c8898353f2770088741eed7"} Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.678219 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998","Type":"ContainerStarted","Data":"f1168e3921eed21e9e9f3fc46e51a896484996f89685c7e47f03c083dd2451a8"} Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.682161 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ac666e99-860b-4f76-8b34-0ac5d3f67e9e","Type":"ContainerStarted","Data":"a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a"} Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.682281 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="ac666e99-860b-4f76-8b34-0ac5d3f67e9e" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a" gracePeriod=30 Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.688327 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3548645f-4c72-4c75-b1bd-95116d47f6e2","Type":"ContainerStarted","Data":"cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713"} Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.691296 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n4z22" event={"ID":"d0cabd29-ef3e-4808-8c92-3b032483789e","Type":"ContainerStarted","Data":"d90ad363a11a672d65b86841626f5e66c7cb30326c75375f126c473ca51ca722"} Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.722900 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.445747854 podStartE2EDuration="10.72284758s" podCreationTimestamp="2026-03-13 10:28:53 +0000 UTC" firstStartedPulling="2026-03-13 10:28:55.258236408 +0000 UTC m=+1509.280766541" lastFinishedPulling="2026-03-13 10:29:02.535336134 +0000 UTC m=+1516.557866267" observedRunningTime="2026-03-13 10:29:03.70110244 +0000 UTC m=+1517.723632593" watchObservedRunningTime="2026-03-13 10:29:03.72284758 +0000 UTC m=+1517.745377713" Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.738340 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-n4z22" podStartSLOduration=4.326789488 podStartE2EDuration="17.738316436s" podCreationTimestamp="2026-03-13 10:28:46 +0000 UTC" firstStartedPulling="2026-03-13 10:28:49.129264399 +0000 UTC m=+1503.151794532" lastFinishedPulling="2026-03-13 10:29:02.540791347 +0000 UTC m=+1516.563321480" observedRunningTime="2026-03-13 10:29:03.726510578 +0000 UTC m=+1517.749040721" watchObservedRunningTime="2026-03-13 10:29:03.738316436 +0000 UTC m=+1517.760846569" Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.760246 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.494929709 podStartE2EDuration="10.760225008s" podCreationTimestamp="2026-03-13 10:28:53 +0000 UTC" firstStartedPulling="2026-03-13 10:28:55.275494578 +0000 UTC m=+1509.298024721" lastFinishedPulling="2026-03-13 10:29:02.540789897 +0000 UTC m=+1516.563320020" observedRunningTime="2026-03-13 10:29:03.752923681 +0000 UTC m=+1517.775453814" watchObservedRunningTime="2026-03-13 10:29:03.760225008 +0000 UTC m=+1517.782755141" Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.778810 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.888070252 podStartE2EDuration="10.778791681s" podCreationTimestamp="2026-03-13 10:28:53 +0000 UTC" firstStartedPulling="2026-03-13 10:28:55.649347151 +0000 UTC m=+1509.671877284" lastFinishedPulling="2026-03-13 10:29:02.54006858 +0000 UTC m=+1516.562598713" observedRunningTime="2026-03-13 10:29:03.778354349 +0000 UTC m=+1517.800884482" watchObservedRunningTime="2026-03-13 10:29:03.778791681 +0000 UTC m=+1517.801321814" Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.844490 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=4.575620436 podStartE2EDuration="10.844464627s" podCreationTimestamp="2026-03-13 10:28:53 +0000 UTC" firstStartedPulling="2026-03-13 10:28:56.246872106 +0000 UTC m=+1510.269402239" lastFinishedPulling="2026-03-13 10:29:02.515716297 +0000 UTC m=+1516.538246430" observedRunningTime="2026-03-13 10:29:03.826333737 +0000 UTC m=+1517.848863870" watchObservedRunningTime="2026-03-13 10:29:03.844464627 +0000 UTC m=+1517.866994760" Mar 13 10:29:03 crc kubenswrapper[4632]: W0313 10:29:03.836536 4632 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0f17959_fde8_4cf1_b255_db5fc3325b70.slice/crio-conmon-c6848744dc1fd449bb0df7b7ca2c04941331806f97abf20c11372e120fb30d31.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0f17959_fde8_4cf1_b255_db5fc3325b70.slice/crio-conmon-c6848744dc1fd449bb0df7b7ca2c04941331806f97abf20c11372e120fb30d31.scope: no such file or directory Mar 13 10:29:03 crc kubenswrapper[4632]: W0313 10:29:03.850412 4632 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0f17959_fde8_4cf1_b255_db5fc3325b70.slice/crio-c6848744dc1fd449bb0df7b7ca2c04941331806f97abf20c11372e120fb30d31.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode0f17959_fde8_4cf1_b255_db5fc3325b70.slice/crio-c6848744dc1fd449bb0df7b7ca2c04941331806f97abf20c11372e120fb30d31.scope: no such file or directory Mar 13 10:29:03 crc kubenswrapper[4632]: W0313 10:29:03.851368 4632 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod229b0823_7d97_48cf_9b38_188a3f4ecde3.slice/crio-conmon-9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod229b0823_7d97_48cf_9b38_188a3f4ecde3.slice/crio-conmon-9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2.scope: no such file or directory Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.915796 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.915839 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.917747 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.208:8774/\": dial tcp 10.217.0.208:8774: connect: connection refused" Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.917755 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.208:8774/\": dial tcp 10.217.0.208:8774: connect: connection refused" Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.961857 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 10:29:03 crc kubenswrapper[4632]: I0313 10:29:03.961907 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.010584 4632 scope.go:117] "RemoveContainer" containerID="0e23e3344de45eadba8d2e2f7dead6b7591126ab6ec56a759524e9fc0c54694e" Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.351392 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.351454 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.393363 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.465178 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.513621 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.560356 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.560721 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="89c5451e-248e-46eb-ac20-f52c3e3bcdc4" containerName="nova-cell0-conductor-conductor" containerID="cri-o://b6bf072e344c147c11f620ba388aa0850202403d9ea7d55387c6d06560823aa8" gracePeriod=30 Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.745771 4632 generic.go:334] "Generic (PLEG): container finished" podID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerID="13beea25c7ec581a71ff8aed4dcb89b5326c0045c02a48578bb1e384a8c92d16" exitCode=0 Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.746142 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d6a890-da6f-4a62-a73d-ad22f8b97586","Type":"ContainerDied","Data":"13beea25c7ec581a71ff8aed4dcb89b5326c0045c02a48578bb1e384a8c92d16"} Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.754909 4632 generic.go:334] "Generic (PLEG): container finished" podID="229b0823-7d97-48cf-9b38-188a3f4ecde3" containerID="9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2" exitCode=143 Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.755043 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"229b0823-7d97-48cf-9b38-188a3f4ecde3","Type":"ContainerDied","Data":"9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2"} Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.817008 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.914121 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:29:04 crc kubenswrapper[4632]: I0313 10:29:04.919031 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.114675 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7888df55c7-mw5p4"] Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.114976 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" podUID="904f04cd-8110-4637-8bb4-67c4b83e189b" containerName="dnsmasq-dns" containerID="cri-o://4cc9fd73a35e44ae17915d74f83df931e877bf9d4b7384d1b90a6239d1a72628" gracePeriod=10 Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.347755 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.463810 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d6a890-da6f-4a62-a73d-ad22f8b97586-log-httpd\") pod \"92d6a890-da6f-4a62-a73d-ad22f8b97586\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.464175 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-config-data\") pod \"92d6a890-da6f-4a62-a73d-ad22f8b97586\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.464373 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-combined-ca-bundle\") pod \"92d6a890-da6f-4a62-a73d-ad22f8b97586\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.464547 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92d6a890-da6f-4a62-a73d-ad22f8b97586-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "92d6a890-da6f-4a62-a73d-ad22f8b97586" (UID: "92d6a890-da6f-4a62-a73d-ad22f8b97586"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.465235 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-sg-core-conf-yaml\") pod \"92d6a890-da6f-4a62-a73d-ad22f8b97586\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.465775 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-scripts\") pod \"92d6a890-da6f-4a62-a73d-ad22f8b97586\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.466000 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dnfcl\" (UniqueName: \"kubernetes.io/projected/92d6a890-da6f-4a62-a73d-ad22f8b97586-kube-api-access-dnfcl\") pod \"92d6a890-da6f-4a62-a73d-ad22f8b97586\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.466179 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d6a890-da6f-4a62-a73d-ad22f8b97586-run-httpd\") pod \"92d6a890-da6f-4a62-a73d-ad22f8b97586\" (UID: \"92d6a890-da6f-4a62-a73d-ad22f8b97586\") " Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.466849 4632 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d6a890-da6f-4a62-a73d-ad22f8b97586-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.467340 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92d6a890-da6f-4a62-a73d-ad22f8b97586-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "92d6a890-da6f-4a62-a73d-ad22f8b97586" (UID: "92d6a890-da6f-4a62-a73d-ad22f8b97586"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.477281 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92d6a890-da6f-4a62-a73d-ad22f8b97586-kube-api-access-dnfcl" (OuterVolumeSpecName: "kube-api-access-dnfcl") pod "92d6a890-da6f-4a62-a73d-ad22f8b97586" (UID: "92d6a890-da6f-4a62-a73d-ad22f8b97586"). InnerVolumeSpecName "kube-api-access-dnfcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.505177 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-scripts" (OuterVolumeSpecName: "scripts") pod "92d6a890-da6f-4a62-a73d-ad22f8b97586" (UID: "92d6a890-da6f-4a62-a73d-ad22f8b97586"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.524171 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "92d6a890-da6f-4a62-a73d-ad22f8b97586" (UID: "92d6a890-da6f-4a62-a73d-ad22f8b97586"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.570444 4632 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.570492 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.570503 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dnfcl\" (UniqueName: \"kubernetes.io/projected/92d6a890-da6f-4a62-a73d-ad22f8b97586-kube-api-access-dnfcl\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.570514 4632 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92d6a890-da6f-4a62-a73d-ad22f8b97586-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.716422 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92d6a890-da6f-4a62-a73d-ad22f8b97586" (UID: "92d6a890-da6f-4a62-a73d-ad22f8b97586"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.780711 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.798289 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.802158 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92d6a890-da6f-4a62-a73d-ad22f8b97586","Type":"ContainerDied","Data":"61eb61c712456be9b4257c5e2ea6a70dfbfca01a50f8412f9ea8b2cdb5c8b498"} Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.802351 4632 scope.go:117] "RemoveContainer" containerID="07674d77dbcfb4d04e610536847653ba6a156f4e167fb3e30be00823bd80251e" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.824184 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" event={"ID":"904f04cd-8110-4637-8bb4-67c4b83e189b","Type":"ContainerDied","Data":"4cc9fd73a35e44ae17915d74f83df931e877bf9d4b7384d1b90a6239d1a72628"} Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.820239 4632 generic.go:334] "Generic (PLEG): container finished" podID="904f04cd-8110-4637-8bb4-67c4b83e189b" containerID="4cc9fd73a35e44ae17915d74f83df931e877bf9d4b7384d1b90a6239d1a72628" exitCode=0 Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.843169 4632 generic.go:334] "Generic (PLEG): container finished" podID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerID="2dbb3ede37abc9f5b483ae48b13ac3ed8913ac4529c34c39494b1541e21ce00b" exitCode=137 Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.843731 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" containerName="nova-api-log" containerID="cri-o://f1168e3921eed21e9e9f3fc46e51a896484996f89685c7e47f03c083dd2451a8" gracePeriod=30 Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.844337 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="3548645f-4c72-4c75-b1bd-95116d47f6e2" containerName="nova-scheduler-scheduler" containerID="cri-o://cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" gracePeriod=30 Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.845279 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" containerName="nova-api-api" containerID="cri-o://29733198b90e2c4e70c2780c9fc89c7f6fe69e2a7c8898353f2770088741eed7" gracePeriod=30 Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.845387 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb5f7878-ng2k2" event={"ID":"3e4afb91-ce26-4325-89c9-2542da2ec48a","Type":"ContainerDied","Data":"2dbb3ede37abc9f5b483ae48b13ac3ed8913ac4529c34c39494b1541e21ce00b"} Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.878766 4632 scope.go:117] "RemoveContainer" containerID="c2c74a9428ab3dfaa995d259e75dccffb44018988a076ef192a947b75ff6a7f1" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.924993 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.942099 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-config-data" (OuterVolumeSpecName: "config-data") pod "92d6a890-da6f-4a62-a73d-ad22f8b97586" (UID: "92d6a890-da6f-4a62-a73d-ad22f8b97586"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.948350 4632 scope.go:117] "RemoveContainer" containerID="a78858d52f4edb9f1b215cb0b5d9d5d059b8c3bfd31b64cf5e0deaf6ab27d4b4" Mar 13 10:29:05 crc kubenswrapper[4632]: I0313 10:29:05.987876 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92d6a890-da6f-4a62-a73d-ad22f8b97586-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.096886 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k6mzz\" (UniqueName: \"kubernetes.io/projected/904f04cd-8110-4637-8bb4-67c4b83e189b-kube-api-access-k6mzz\") pod \"904f04cd-8110-4637-8bb4-67c4b83e189b\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.163314 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-ovsdbserver-sb\") pod \"904f04cd-8110-4637-8bb4-67c4b83e189b\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.163473 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-config\") pod \"904f04cd-8110-4637-8bb4-67c4b83e189b\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.325616 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/904f04cd-8110-4637-8bb4-67c4b83e189b-kube-api-access-k6mzz" (OuterVolumeSpecName: "kube-api-access-k6mzz") pod "904f04cd-8110-4637-8bb4-67c4b83e189b" (UID: "904f04cd-8110-4637-8bb4-67c4b83e189b"). InnerVolumeSpecName "kube-api-access-k6mzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.331772 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-dns-svc\") pod \"904f04cd-8110-4637-8bb4-67c4b83e189b\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.332045 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-ovsdbserver-nb\") pod \"904f04cd-8110-4637-8bb4-67c4b83e189b\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.332155 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-dns-swift-storage-0\") pod \"904f04cd-8110-4637-8bb4-67c4b83e189b\" (UID: \"904f04cd-8110-4637-8bb4-67c4b83e189b\") " Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.333163 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k6mzz\" (UniqueName: \"kubernetes.io/projected/904f04cd-8110-4637-8bb4-67c4b83e189b-kube-api-access-k6mzz\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.337154 4632 scope.go:117] "RemoveContainer" containerID="13beea25c7ec581a71ff8aed4dcb89b5326c0045c02a48578bb1e384a8c92d16" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.455114 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-config" (OuterVolumeSpecName: "config") pod "904f04cd-8110-4637-8bb4-67c4b83e189b" (UID: "904f04cd-8110-4637-8bb4-67c4b83e189b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.505087 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "904f04cd-8110-4637-8bb4-67c4b83e189b" (UID: "904f04cd-8110-4637-8bb4-67c4b83e189b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.538901 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.538965 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.558155 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "904f04cd-8110-4637-8bb4-67c4b83e189b" (UID: "904f04cd-8110-4637-8bb4-67c4b83e189b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.563381 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "904f04cd-8110-4637-8bb4-67c4b83e189b" (UID: "904f04cd-8110-4637-8bb4-67c4b83e189b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.573514 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "904f04cd-8110-4637-8bb4-67c4b83e189b" (UID: "904f04cd-8110-4637-8bb4-67c4b83e189b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.640689 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.640728 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.640738 4632 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/904f04cd-8110-4637-8bb4-67c4b83e189b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.686209 4632 scope.go:117] "RemoveContainer" containerID="c9dfdd84c36e6ac95b45a488b62e176636bdecfbe3a88d3f5d2058d92ebbacdd" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.696445 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.710803 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.724216 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:29:06 crc kubenswrapper[4632]: E0313 10:29:06.729834 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="proxy-httpd" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.729931 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="proxy-httpd" Mar 13 10:29:06 crc kubenswrapper[4632]: E0313 10:29:06.730024 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="ceilometer-notification-agent" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.730103 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="ceilometer-notification-agent" Mar 13 10:29:06 crc kubenswrapper[4632]: E0313 10:29:06.730167 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="904f04cd-8110-4637-8bb4-67c4b83e189b" containerName="init" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.730221 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="904f04cd-8110-4637-8bb4-67c4b83e189b" containerName="init" Mar 13 10:29:06 crc kubenswrapper[4632]: E0313 10:29:06.730278 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="904f04cd-8110-4637-8bb4-67c4b83e189b" containerName="dnsmasq-dns" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.730328 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="904f04cd-8110-4637-8bb4-67c4b83e189b" containerName="dnsmasq-dns" Mar 13 10:29:06 crc kubenswrapper[4632]: E0313 10:29:06.730390 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="sg-core" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.730446 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="sg-core" Mar 13 10:29:06 crc kubenswrapper[4632]: E0313 10:29:06.730532 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="ceilometer-central-agent" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.730593 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="ceilometer-central-agent" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.730820 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="ceilometer-central-agent" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.730889 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="904f04cd-8110-4637-8bb4-67c4b83e189b" containerName="dnsmasq-dns" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.730984 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="proxy-httpd" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.731045 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="sg-core" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.731180 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" containerName="ceilometer-notification-agent" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.732898 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.737502 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.737853 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.771788 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.847677 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-config-data\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.848537 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-log-httpd\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.848743 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-run-httpd\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.848901 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.849198 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.849377 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-scripts\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.849495 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtzmr\" (UniqueName: \"kubernetes.io/projected/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-kube-api-access-vtzmr\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.857476 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" event={"ID":"904f04cd-8110-4637-8bb4-67c4b83e189b","Type":"ContainerDied","Data":"b305d4370882ddeb316b7136e1b6a31fb9b050f68adc94baa9487a0176e85bb7"} Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.857771 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7888df55c7-mw5p4" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.874666 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb5f7878-ng2k2" event={"ID":"3e4afb91-ce26-4325-89c9-2542da2ec48a","Type":"ContainerStarted","Data":"0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30"} Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.882188 4632 generic.go:334] "Generic (PLEG): container finished" podID="1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" containerID="29733198b90e2c4e70c2780c9fc89c7f6fe69e2a7c8898353f2770088741eed7" exitCode=0 Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.882232 4632 generic.go:334] "Generic (PLEG): container finished" podID="1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" containerID="f1168e3921eed21e9e9f3fc46e51a896484996f89685c7e47f03c083dd2451a8" exitCode=143 Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.882279 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998","Type":"ContainerDied","Data":"29733198b90e2c4e70c2780c9fc89c7f6fe69e2a7c8898353f2770088741eed7"} Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.882305 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998","Type":"ContainerDied","Data":"f1168e3921eed21e9e9f3fc46e51a896484996f89685c7e47f03c083dd2451a8"} Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.892394 4632 generic.go:334] "Generic (PLEG): container finished" podID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerID="26a7aae686bb479cfcbc8b01e8e10e3fd467e5236d6ffb2ed638373687267401" exitCode=0 Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.892451 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689764498d-rg7vt" event={"ID":"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c","Type":"ContainerDied","Data":"26a7aae686bb479cfcbc8b01e8e10e3fd467e5236d6ffb2ed638373687267401"} Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.896731 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7888df55c7-mw5p4"] Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.910517 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7888df55c7-mw5p4"] Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.952544 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-config-data\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.952790 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-log-httpd\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.952938 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-run-httpd\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.953142 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.953272 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.953375 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-scripts\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.953625 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtzmr\" (UniqueName: \"kubernetes.io/projected/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-kube-api-access-vtzmr\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.955331 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-log-httpd\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.956649 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-run-httpd\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.958389 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.959142 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-config-data\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.962704 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-scripts\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.963521 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:06 crc kubenswrapper[4632]: I0313 10:29:06.980438 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtzmr\" (UniqueName: \"kubernetes.io/projected/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-kube-api-access-vtzmr\") pod \"ceilometer-0\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " pod="openstack/ceilometer-0" Mar 13 10:29:07 crc kubenswrapper[4632]: I0313 10:29:07.097079 4632 scope.go:117] "RemoveContainer" containerID="4cc9fd73a35e44ae17915d74f83df931e877bf9d4b7384d1b90a6239d1a72628" Mar 13 10:29:07 crc kubenswrapper[4632]: I0313 10:29:07.100375 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:29:07 crc kubenswrapper[4632]: I0313 10:29:07.150197 4632 scope.go:117] "RemoveContainer" containerID="10ef0805fc14af19dcea5ad4d4426bd1471fa5008be0ab704ad9b901662ea060" Mar 13 10:29:07 crc kubenswrapper[4632]: I0313 10:29:07.230984 4632 scope.go:117] "RemoveContainer" containerID="433c9aa5a02161c4bc7228b52cc460020479cbbb899bc6549755a59b8ad796f4" Mar 13 10:29:07 crc kubenswrapper[4632]: I0313 10:29:07.249240 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:29:07 crc kubenswrapper[4632]: I0313 10:29:07.254070 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:29:07 crc kubenswrapper[4632]: E0313 10:29:07.470575 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b6bf072e344c147c11f620ba388aa0850202403d9ea7d55387c6d06560823aa8" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 10:29:07 crc kubenswrapper[4632]: E0313 10:29:07.474772 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b6bf072e344c147c11f620ba388aa0850202403d9ea7d55387c6d06560823aa8" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 10:29:07 crc kubenswrapper[4632]: E0313 10:29:07.483768 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b6bf072e344c147c11f620ba388aa0850202403d9ea7d55387c6d06560823aa8" cmd=["/usr/bin/pgrep","-r","DRST","nova-conductor"] Mar 13 10:29:07 crc kubenswrapper[4632]: E0313 10:29:07.483835 4632 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-cell0-conductor-0" podUID="89c5451e-248e-46eb-ac20-f52c3e3bcdc4" containerName="nova-cell0-conductor-conductor" Mar 13 10:29:07 crc kubenswrapper[4632]: I0313 10:29:07.877445 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.026160 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-689764498d-rg7vt" event={"ID":"5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c","Type":"ContainerStarted","Data":"468b0c833599c14f4c7d5ed1aa0e813466a55e5432fa312ef3fc463200e9d1b1"} Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.090733 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="904f04cd-8110-4637-8bb4-67c4b83e189b" path="/var/lib/kubelet/pods/904f04cd-8110-4637-8bb4-67c4b83e189b/volumes" Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.095606 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92d6a890-da6f-4a62-a73d-ad22f8b97586" path="/var/lib/kubelet/pods/92d6a890-da6f-4a62-a73d-ad22f8b97586/volumes" Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.279300 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.400871 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-config-data\") pod \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.401003 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-logs\") pod \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.401033 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-combined-ca-bundle\") pod \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.401083 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9h2v\" (UniqueName: \"kubernetes.io/projected/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-kube-api-access-g9h2v\") pod \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\" (UID: \"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998\") " Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.401404 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-logs" (OuterVolumeSpecName: "logs") pod "1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" (UID: "1b3f0cc6-72ae-4738-baee-ce7bc9ef2998"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.401499 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.414209 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-kube-api-access-g9h2v" (OuterVolumeSpecName: "kube-api-access-g9h2v") pod "1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" (UID: "1b3f0cc6-72ae-4738-baee-ce7bc9ef2998"). InnerVolumeSpecName "kube-api-access-g9h2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.447544 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n4z22" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="registry-server" probeResult="failure" output=< Mar 13 10:29:08 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:29:08 crc kubenswrapper[4632]: > Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.485142 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" (UID: "1b3f0cc6-72ae-4738-baee-ce7bc9ef2998"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.503405 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.503434 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9h2v\" (UniqueName: \"kubernetes.io/projected/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-kube-api-access-g9h2v\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.503506 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-config-data" (OuterVolumeSpecName: "config-data") pod "1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" (UID: "1b3f0cc6-72ae-4738-baee-ce7bc9ef2998"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:08 crc kubenswrapper[4632]: I0313 10:29:08.671230 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.078522 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae6e23c-4223-4a5c-8074-0cdfb5d99e78","Type":"ContainerStarted","Data":"fb2da7115022d9a5776bf008164ce9b7e7fedf403dd765c92f96767951f31f8f"} Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.079718 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae6e23c-4223-4a5c-8074-0cdfb5d99e78","Type":"ContainerStarted","Data":"ac0a78784a347100c3f7eb503566ca534d51cff4225e41d7e119e9bf5dd8e6ea"} Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.096528 4632 generic.go:334] "Generic (PLEG): container finished" podID="89c5451e-248e-46eb-ac20-f52c3e3bcdc4" containerID="b6bf072e344c147c11f620ba388aa0850202403d9ea7d55387c6d06560823aa8" exitCode=0 Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.096831 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"89c5451e-248e-46eb-ac20-f52c3e3bcdc4","Type":"ContainerDied","Data":"b6bf072e344c147c11f620ba388aa0850202403d9ea7d55387c6d06560823aa8"} Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.143328 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.143601 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1b3f0cc6-72ae-4738-baee-ce7bc9ef2998","Type":"ContainerDied","Data":"34daac34beb9c4f24aabab024257734fad48bd622c9a19af69564a5c1af316f2"} Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.143696 4632 scope.go:117] "RemoveContainer" containerID="29733198b90e2c4e70c2780c9fc89c7f6fe69e2a7c8898353f2770088741eed7" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.207771 4632 scope.go:117] "RemoveContainer" containerID="f1168e3921eed21e9e9f3fc46e51a896484996f89685c7e47f03c083dd2451a8" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.252991 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.283615 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.302921 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 10:29:09 crc kubenswrapper[4632]: E0313 10:29:09.303337 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" containerName="nova-api-log" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.303356 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" containerName="nova-api-log" Mar 13 10:29:09 crc kubenswrapper[4632]: E0313 10:29:09.303366 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" containerName="nova-api-api" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.303372 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" containerName="nova-api-api" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.303597 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" containerName="nova-api-log" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.303636 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" containerName="nova-api-api" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.304664 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.307248 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.326252 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:29:09 crc kubenswrapper[4632]: E0313 10:29:09.357842 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:09 crc kubenswrapper[4632]: E0313 10:29:09.359965 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:09 crc kubenswrapper[4632]: E0313 10:29:09.363505 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:09 crc kubenswrapper[4632]: E0313 10:29:09.363566 4632 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="3548645f-4c72-4c75-b1bd-95116d47f6e2" containerName="nova-scheduler-scheduler" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.490984 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-config-data\") pod \"nova-api-0\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.491118 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-logs\") pod \"nova-api-0\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.491206 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8l5th\" (UniqueName: \"kubernetes.io/projected/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-kube-api-access-8l5th\") pod \"nova-api-0\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.491254 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.593268 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.593381 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-config-data\") pod \"nova-api-0\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.593486 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-logs\") pod \"nova-api-0\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.593570 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8l5th\" (UniqueName: \"kubernetes.io/projected/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-kube-api-access-8l5th\") pod \"nova-api-0\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.600126 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-logs\") pod \"nova-api-0\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.607846 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-config-data\") pod \"nova-api-0\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.616216 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.650741 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8l5th\" (UniqueName: \"kubernetes.io/projected/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-kube-api-access-8l5th\") pod \"nova-api-0\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.695503 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.799659 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.910121 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjsd4\" (UniqueName: \"kubernetes.io/projected/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-kube-api-access-pjsd4\") pod \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\" (UID: \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\") " Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.910454 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-combined-ca-bundle\") pod \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\" (UID: \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\") " Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.910558 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-config-data\") pod \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\" (UID: \"89c5451e-248e-46eb-ac20-f52c3e3bcdc4\") " Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.926093 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-kube-api-access-pjsd4" (OuterVolumeSpecName: "kube-api-access-pjsd4") pod "89c5451e-248e-46eb-ac20-f52c3e3bcdc4" (UID: "89c5451e-248e-46eb-ac20-f52c3e3bcdc4"). InnerVolumeSpecName "kube-api-access-pjsd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:29:09 crc kubenswrapper[4632]: I0313 10:29:09.966216 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "89c5451e-248e-46eb-ac20-f52c3e3bcdc4" (UID: "89c5451e-248e-46eb-ac20-f52c3e3bcdc4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.014456 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjsd4\" (UniqueName: \"kubernetes.io/projected/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-kube-api-access-pjsd4\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.014725 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.111097 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-config-data" (OuterVolumeSpecName: "config-data") pod "89c5451e-248e-46eb-ac20-f52c3e3bcdc4" (UID: "89c5451e-248e-46eb-ac20-f52c3e3bcdc4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.113384 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b3f0cc6-72ae-4738-baee-ce7bc9ef2998" path="/var/lib/kubelet/pods/1b3f0cc6-72ae-4738-baee-ce7bc9ef2998/volumes" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.120602 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/89c5451e-248e-46eb-ac20-f52c3e3bcdc4-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.214398 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.215344 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"89c5451e-248e-46eb-ac20-f52c3e3bcdc4","Type":"ContainerDied","Data":"fea8b62da5fff833a90864e9fa4a28877f40e3642c9c75596310ee934707e980"} Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.215406 4632 scope.go:117] "RemoveContainer" containerID="b6bf072e344c147c11f620ba388aa0850202403d9ea7d55387c6d06560823aa8" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.239010 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.278469 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae6e23c-4223-4a5c-8074-0cdfb5d99e78","Type":"ContainerStarted","Data":"b21cf0ac61b24fc322cab134fff86780fa6ec89891d057825ed89069f94de13b"} Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.307045 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.340862 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.427014 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 10:29:10 crc kubenswrapper[4632]: E0313 10:29:10.427517 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="89c5451e-248e-46eb-ac20-f52c3e3bcdc4" containerName="nova-cell0-conductor-conductor" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.427534 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="89c5451e-248e-46eb-ac20-f52c3e3bcdc4" containerName="nova-cell0-conductor-conductor" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.427791 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="89c5451e-248e-46eb-ac20-f52c3e3bcdc4" containerName="nova-cell0-conductor-conductor" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.436222 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.446743 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.449965 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.540801 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe53f0a-8bf3-4572-b5c8-01d5ed72c426-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"dbe53f0a-8bf3-4572-b5c8-01d5ed72c426\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.541183 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe53f0a-8bf3-4572-b5c8-01d5ed72c426-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"dbe53f0a-8bf3-4572-b5c8-01d5ed72c426\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.541236 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j5rv\" (UniqueName: \"kubernetes.io/projected/dbe53f0a-8bf3-4572-b5c8-01d5ed72c426-kube-api-access-5j5rv\") pod \"nova-cell0-conductor-0\" (UID: \"dbe53f0a-8bf3-4572-b5c8-01d5ed72c426\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.642682 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe53f0a-8bf3-4572-b5c8-01d5ed72c426-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"dbe53f0a-8bf3-4572-b5c8-01d5ed72c426\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.643064 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j5rv\" (UniqueName: \"kubernetes.io/projected/dbe53f0a-8bf3-4572-b5c8-01d5ed72c426-kube-api-access-5j5rv\") pod \"nova-cell0-conductor-0\" (UID: \"dbe53f0a-8bf3-4572-b5c8-01d5ed72c426\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.643251 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe53f0a-8bf3-4572-b5c8-01d5ed72c426-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"dbe53f0a-8bf3-4572-b5c8-01d5ed72c426\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.646835 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.654345 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dbe53f0a-8bf3-4572-b5c8-01d5ed72c426-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"dbe53f0a-8bf3-4572-b5c8-01d5ed72c426\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.657474 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dbe53f0a-8bf3-4572-b5c8-01d5ed72c426-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"dbe53f0a-8bf3-4572-b5c8-01d5ed72c426\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.670613 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j5rv\" (UniqueName: \"kubernetes.io/projected/dbe53f0a-8bf3-4572-b5c8-01d5ed72c426-kube-api-access-5j5rv\") pod \"nova-cell0-conductor-0\" (UID: \"dbe53f0a-8bf3-4572-b5c8-01d5ed72c426\") " pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:10 crc kubenswrapper[4632]: I0313 10:29:10.793639 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:11 crc kubenswrapper[4632]: I0313 10:29:11.297119 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"71d36cc3-a7e3-47e5-b98f-599dc669ccc5","Type":"ContainerStarted","Data":"4df2156f6fe32fab45f05d256a8ec2adb23f786a2989c939b92b996a496f122f"} Mar 13 10:29:11 crc kubenswrapper[4632]: I0313 10:29:11.297450 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"71d36cc3-a7e3-47e5-b98f-599dc669ccc5","Type":"ContainerStarted","Data":"85c564a65a9f43d06ad1647efab572b9e533b3cb0feadf64be1f6226a656d6e9"} Mar 13 10:29:11 crc kubenswrapper[4632]: I0313 10:29:11.309101 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae6e23c-4223-4a5c-8074-0cdfb5d99e78","Type":"ContainerStarted","Data":"585bc9dbde7bcb94e64b2121baa3ef51bee81b44562ad73e38572b5035710787"} Mar 13 10:29:11 crc kubenswrapper[4632]: I0313 10:29:11.446953 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 10:29:12 crc kubenswrapper[4632]: I0313 10:29:12.055103 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="89c5451e-248e-46eb-ac20-f52c3e3bcdc4" path="/var/lib/kubelet/pods/89c5451e-248e-46eb-ac20-f52c3e3bcdc4/volumes" Mar 13 10:29:12 crc kubenswrapper[4632]: I0313 10:29:12.320984 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"dbe53f0a-8bf3-4572-b5c8-01d5ed72c426","Type":"ContainerStarted","Data":"37c2692b2411684aa86eeacfa77a4627a5e26055bee50129ef79a6b5d713b14b"} Mar 13 10:29:12 crc kubenswrapper[4632]: I0313 10:29:12.321033 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"dbe53f0a-8bf3-4572-b5c8-01d5ed72c426","Type":"ContainerStarted","Data":"db803b63c80478ad379bb99fe0594d8b3d74a21a8ad14d3a30fc918f8f1ba6e1"} Mar 13 10:29:12 crc kubenswrapper[4632]: I0313 10:29:12.322923 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:12 crc kubenswrapper[4632]: I0313 10:29:12.323669 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"71d36cc3-a7e3-47e5-b98f-599dc669ccc5","Type":"ContainerStarted","Data":"e604b5ae6ce92dde6f33a140a99a7c7d5949aebd7f4821ef087f38b50a0e872b"} Mar 13 10:29:12 crc kubenswrapper[4632]: I0313 10:29:12.341648 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.34161136 podStartE2EDuration="2.34161136s" podCreationTimestamp="2026-03-13 10:29:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:29:12.338865823 +0000 UTC m=+1526.361395966" watchObservedRunningTime="2026-03-13 10:29:12.34161136 +0000 UTC m=+1526.364141493" Mar 13 10:29:12 crc kubenswrapper[4632]: I0313 10:29:12.360621 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.360598262 podStartE2EDuration="3.360598262s" podCreationTimestamp="2026-03-13 10:29:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:29:12.358495111 +0000 UTC m=+1526.381025244" watchObservedRunningTime="2026-03-13 10:29:12.360598262 +0000 UTC m=+1526.383128385" Mar 13 10:29:13 crc kubenswrapper[4632]: I0313 10:29:13.336446 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae6e23c-4223-4a5c-8074-0cdfb5d99e78","Type":"ContainerStarted","Data":"604c723ae68fd3213f4723bdf3524877331344576a1c363f42185040543011d1"} Mar 13 10:29:13 crc kubenswrapper[4632]: I0313 10:29:13.337454 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="ceilometer-central-agent" containerID="cri-o://fb2da7115022d9a5776bf008164ce9b7e7fedf403dd765c92f96767951f31f8f" gracePeriod=30 Mar 13 10:29:13 crc kubenswrapper[4632]: I0313 10:29:13.337480 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 10:29:13 crc kubenswrapper[4632]: I0313 10:29:13.337479 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="proxy-httpd" containerID="cri-o://604c723ae68fd3213f4723bdf3524877331344576a1c363f42185040543011d1" gracePeriod=30 Mar 13 10:29:13 crc kubenswrapper[4632]: I0313 10:29:13.337495 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="sg-core" containerID="cri-o://585bc9dbde7bcb94e64b2121baa3ef51bee81b44562ad73e38572b5035710787" gracePeriod=30 Mar 13 10:29:13 crc kubenswrapper[4632]: I0313 10:29:13.337521 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="ceilometer-notification-agent" containerID="cri-o://b21cf0ac61b24fc322cab134fff86780fa6ec89891d057825ed89069f94de13b" gracePeriod=30 Mar 13 10:29:13 crc kubenswrapper[4632]: I0313 10:29:13.393908 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.929082165 podStartE2EDuration="7.393884886s" podCreationTimestamp="2026-03-13 10:29:06 +0000 UTC" firstStartedPulling="2026-03-13 10:29:07.972752583 +0000 UTC m=+1521.995282726" lastFinishedPulling="2026-03-13 10:29:12.437555314 +0000 UTC m=+1526.460085447" observedRunningTime="2026-03-13 10:29:13.376074042 +0000 UTC m=+1527.398604175" watchObservedRunningTime="2026-03-13 10:29:13.393884886 +0000 UTC m=+1527.416415019" Mar 13 10:29:14 crc kubenswrapper[4632]: I0313 10:29:14.349887 4632 generic.go:334] "Generic (PLEG): container finished" podID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerID="604c723ae68fd3213f4723bdf3524877331344576a1c363f42185040543011d1" exitCode=0 Mar 13 10:29:14 crc kubenswrapper[4632]: I0313 10:29:14.350208 4632 generic.go:334] "Generic (PLEG): container finished" podID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerID="585bc9dbde7bcb94e64b2121baa3ef51bee81b44562ad73e38572b5035710787" exitCode=2 Mar 13 10:29:14 crc kubenswrapper[4632]: I0313 10:29:14.349932 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae6e23c-4223-4a5c-8074-0cdfb5d99e78","Type":"ContainerDied","Data":"604c723ae68fd3213f4723bdf3524877331344576a1c363f42185040543011d1"} Mar 13 10:29:14 crc kubenswrapper[4632]: I0313 10:29:14.350257 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae6e23c-4223-4a5c-8074-0cdfb5d99e78","Type":"ContainerDied","Data":"585bc9dbde7bcb94e64b2121baa3ef51bee81b44562ad73e38572b5035710787"} Mar 13 10:29:14 crc kubenswrapper[4632]: E0313 10:29:14.353407 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:14 crc kubenswrapper[4632]: E0313 10:29:14.354776 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:14 crc kubenswrapper[4632]: E0313 10:29:14.356467 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:14 crc kubenswrapper[4632]: E0313 10:29:14.356505 4632 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="3548645f-4c72-4c75-b1bd-95116d47f6e2" containerName="nova-scheduler-scheduler" Mar 13 10:29:15 crc kubenswrapper[4632]: I0313 10:29:15.394377 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:29:15 crc kubenswrapper[4632]: I0313 10:29:15.394813 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:29:16 crc kubenswrapper[4632]: I0313 10:29:15.857046 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:29:16 crc kubenswrapper[4632]: I0313 10:29:15.858031 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:29:16 crc kubenswrapper[4632]: I0313 10:29:16.393873 4632 generic.go:334] "Generic (PLEG): container finished" podID="bcce9343-52a3-4e6d-98fd-8e66390020ac" containerID="379356ecac878a5f4776d015be267e8c7eec62c977ce924abd53ff44455ce8e4" exitCode=0 Mar 13 10:29:16 crc kubenswrapper[4632]: I0313 10:29:16.394002 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gwj5n" event={"ID":"bcce9343-52a3-4e6d-98fd-8e66390020ac","Type":"ContainerDied","Data":"379356ecac878a5f4776d015be267e8c7eec62c977ce924abd53ff44455ce8e4"} Mar 13 10:29:17 crc kubenswrapper[4632]: I0313 10:29:17.411111 4632 generic.go:334] "Generic (PLEG): container finished" podID="cf19672e-3284-49bc-a460-f2e629881d9b" containerID="a8f98d9cfd7da7677c0fe463edd081d6aa2858ecb1027917673862b2700f1545" exitCode=0 Mar 13 10:29:17 crc kubenswrapper[4632]: I0313 10:29:17.411366 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9n7gj" event={"ID":"cf19672e-3284-49bc-a460-f2e629881d9b","Type":"ContainerDied","Data":"a8f98d9cfd7da7677c0fe463edd081d6aa2858ecb1027917673862b2700f1545"} Mar 13 10:29:17 crc kubenswrapper[4632]: I0313 10:29:17.850579 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:29:17 crc kubenswrapper[4632]: I0313 10:29:17.967913 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-config-data\") pod \"bcce9343-52a3-4e6d-98fd-8e66390020ac\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " Mar 13 10:29:17 crc kubenswrapper[4632]: I0313 10:29:17.968100 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-combined-ca-bundle\") pod \"bcce9343-52a3-4e6d-98fd-8e66390020ac\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " Mar 13 10:29:17 crc kubenswrapper[4632]: I0313 10:29:17.968287 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxd74\" (UniqueName: \"kubernetes.io/projected/bcce9343-52a3-4e6d-98fd-8e66390020ac-kube-api-access-sxd74\") pod \"bcce9343-52a3-4e6d-98fd-8e66390020ac\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " Mar 13 10:29:17 crc kubenswrapper[4632]: I0313 10:29:17.968367 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-scripts\") pod \"bcce9343-52a3-4e6d-98fd-8e66390020ac\" (UID: \"bcce9343-52a3-4e6d-98fd-8e66390020ac\") " Mar 13 10:29:17 crc kubenswrapper[4632]: I0313 10:29:17.974707 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcce9343-52a3-4e6d-98fd-8e66390020ac-kube-api-access-sxd74" (OuterVolumeSpecName: "kube-api-access-sxd74") pod "bcce9343-52a3-4e6d-98fd-8e66390020ac" (UID: "bcce9343-52a3-4e6d-98fd-8e66390020ac"). InnerVolumeSpecName "kube-api-access-sxd74". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:29:17 crc kubenswrapper[4632]: I0313 10:29:17.977783 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-scripts" (OuterVolumeSpecName: "scripts") pod "bcce9343-52a3-4e6d-98fd-8e66390020ac" (UID: "bcce9343-52a3-4e6d-98fd-8e66390020ac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.007744 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-config-data" (OuterVolumeSpecName: "config-data") pod "bcce9343-52a3-4e6d-98fd-8e66390020ac" (UID: "bcce9343-52a3-4e6d-98fd-8e66390020ac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.016106 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bcce9343-52a3-4e6d-98fd-8e66390020ac" (UID: "bcce9343-52a3-4e6d-98fd-8e66390020ac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.070185 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.070217 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxd74\" (UniqueName: \"kubernetes.io/projected/bcce9343-52a3-4e6d-98fd-8e66390020ac-kube-api-access-sxd74\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.070228 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.070237 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bcce9343-52a3-4e6d-98fd-8e66390020ac-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.298036 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n4z22" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="registry-server" probeResult="failure" output=< Mar 13 10:29:18 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:29:18 crc kubenswrapper[4632]: > Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.423624 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-gwj5n" Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.424832 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-gwj5n" event={"ID":"bcce9343-52a3-4e6d-98fd-8e66390020ac","Type":"ContainerDied","Data":"f930aa1f0069e1fe78556089c981d58cc3cf6a82579a76e645212ffad42b673e"} Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.424857 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f930aa1f0069e1fe78556089c981d58cc3cf6a82579a76e645212ffad42b673e" Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.883148 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.990545 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cpkl\" (UniqueName: \"kubernetes.io/projected/cf19672e-3284-49bc-a460-f2e629881d9b-kube-api-access-2cpkl\") pod \"cf19672e-3284-49bc-a460-f2e629881d9b\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.991064 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-scripts\") pod \"cf19672e-3284-49bc-a460-f2e629881d9b\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.991159 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-combined-ca-bundle\") pod \"cf19672e-3284-49bc-a460-f2e629881d9b\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " Mar 13 10:29:18 crc kubenswrapper[4632]: I0313 10:29:18.991268 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-config-data\") pod \"cf19672e-3284-49bc-a460-f2e629881d9b\" (UID: \"cf19672e-3284-49bc-a460-f2e629881d9b\") " Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.006414 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf19672e-3284-49bc-a460-f2e629881d9b-kube-api-access-2cpkl" (OuterVolumeSpecName: "kube-api-access-2cpkl") pod "cf19672e-3284-49bc-a460-f2e629881d9b" (UID: "cf19672e-3284-49bc-a460-f2e629881d9b"). InnerVolumeSpecName "kube-api-access-2cpkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.015261 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-scripts" (OuterVolumeSpecName: "scripts") pod "cf19672e-3284-49bc-a460-f2e629881d9b" (UID: "cf19672e-3284-49bc-a460-f2e629881d9b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.034894 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cf19672e-3284-49bc-a460-f2e629881d9b" (UID: "cf19672e-3284-49bc-a460-f2e629881d9b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.035556 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-config-data" (OuterVolumeSpecName: "config-data") pod "cf19672e-3284-49bc-a460-f2e629881d9b" (UID: "cf19672e-3284-49bc-a460-f2e629881d9b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.095668 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cpkl\" (UniqueName: \"kubernetes.io/projected/cf19672e-3284-49bc-a460-f2e629881d9b-kube-api-access-2cpkl\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.095704 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.095716 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.095727 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf19672e-3284-49bc-a460-f2e629881d9b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:19 crc kubenswrapper[4632]: E0313 10:29:19.355343 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:19 crc kubenswrapper[4632]: E0313 10:29:19.356711 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:19 crc kubenswrapper[4632]: E0313 10:29:19.357814 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:19 crc kubenswrapper[4632]: E0313 10:29:19.357851 4632 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="3548645f-4c72-4c75-b1bd-95116d47f6e2" containerName="nova-scheduler-scheduler" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.435077 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-9n7gj" event={"ID":"cf19672e-3284-49bc-a460-f2e629881d9b","Type":"ContainerDied","Data":"2552baf18c3652ea7e85b1cf98a826d4538b2dbd01aa4210322514f06f9b9c99"} Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.435119 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2552baf18c3652ea7e85b1cf98a826d4538b2dbd01aa4210322514f06f9b9c99" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.435152 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-9n7gj" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.526259 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 10:29:19 crc kubenswrapper[4632]: E0313 10:29:19.527412 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcce9343-52a3-4e6d-98fd-8e66390020ac" containerName="nova-manage" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.527429 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcce9343-52a3-4e6d-98fd-8e66390020ac" containerName="nova-manage" Mar 13 10:29:19 crc kubenswrapper[4632]: E0313 10:29:19.527446 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf19672e-3284-49bc-a460-f2e629881d9b" containerName="nova-cell1-conductor-db-sync" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.527453 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf19672e-3284-49bc-a460-f2e629881d9b" containerName="nova-cell1-conductor-db-sync" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.527628 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcce9343-52a3-4e6d-98fd-8e66390020ac" containerName="nova-manage" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.527666 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf19672e-3284-49bc-a460-f2e629881d9b" containerName="nova-cell1-conductor-db-sync" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.528270 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.531363 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.588303 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.619583 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/febcbdc5-25a6-46f7-8c06-d6f45624a466-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"febcbdc5-25a6-46f7-8c06-d6f45624a466\") " pod="openstack/nova-cell1-conductor-0" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.619649 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febcbdc5-25a6-46f7-8c06-d6f45624a466-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"febcbdc5-25a6-46f7-8c06-d6f45624a466\") " pod="openstack/nova-cell1-conductor-0" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.619727 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wxm2\" (UniqueName: \"kubernetes.io/projected/febcbdc5-25a6-46f7-8c06-d6f45624a466-kube-api-access-4wxm2\") pod \"nova-cell1-conductor-0\" (UID: \"febcbdc5-25a6-46f7-8c06-d6f45624a466\") " pod="openstack/nova-cell1-conductor-0" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.697040 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.697449 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.722196 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/febcbdc5-25a6-46f7-8c06-d6f45624a466-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"febcbdc5-25a6-46f7-8c06-d6f45624a466\") " pod="openstack/nova-cell1-conductor-0" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.722267 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febcbdc5-25a6-46f7-8c06-d6f45624a466-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"febcbdc5-25a6-46f7-8c06-d6f45624a466\") " pod="openstack/nova-cell1-conductor-0" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.722334 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wxm2\" (UniqueName: \"kubernetes.io/projected/febcbdc5-25a6-46f7-8c06-d6f45624a466-kube-api-access-4wxm2\") pod \"nova-cell1-conductor-0\" (UID: \"febcbdc5-25a6-46f7-8c06-d6f45624a466\") " pod="openstack/nova-cell1-conductor-0" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.737250 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/febcbdc5-25a6-46f7-8c06-d6f45624a466-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"febcbdc5-25a6-46f7-8c06-d6f45624a466\") " pod="openstack/nova-cell1-conductor-0" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.741641 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febcbdc5-25a6-46f7-8c06-d6f45624a466-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"febcbdc5-25a6-46f7-8c06-d6f45624a466\") " pod="openstack/nova-cell1-conductor-0" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.747646 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wxm2\" (UniqueName: \"kubernetes.io/projected/febcbdc5-25a6-46f7-8c06-d6f45624a466-kube-api-access-4wxm2\") pod \"nova-cell1-conductor-0\" (UID: \"febcbdc5-25a6-46f7-8c06-d6f45624a466\") " pod="openstack/nova-cell1-conductor-0" Mar 13 10:29:19 crc kubenswrapper[4632]: I0313 10:29:19.853215 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 10:29:20 crc kubenswrapper[4632]: I0313 10:29:20.779284 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="71d36cc3-a7e3-47e5-b98f-599dc669ccc5" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.215:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:20 crc kubenswrapper[4632]: I0313 10:29:20.779307 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="71d36cc3-a7e3-47e5-b98f-599dc669ccc5" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.215:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:20 crc kubenswrapper[4632]: I0313 10:29:20.853843 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 13 10:29:21 crc kubenswrapper[4632]: I0313 10:29:21.309446 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 10:29:21 crc kubenswrapper[4632]: I0313 10:29:21.489748 4632 generic.go:334] "Generic (PLEG): container finished" podID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerID="fb2da7115022d9a5776bf008164ce9b7e7fedf403dd765c92f96767951f31f8f" exitCode=0 Mar 13 10:29:21 crc kubenswrapper[4632]: I0313 10:29:21.490126 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae6e23c-4223-4a5c-8074-0cdfb5d99e78","Type":"ContainerDied","Data":"fb2da7115022d9a5776bf008164ce9b7e7fedf403dd765c92f96767951f31f8f"} Mar 13 10:29:21 crc kubenswrapper[4632]: I0313 10:29:21.496624 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"febcbdc5-25a6-46f7-8c06-d6f45624a466","Type":"ContainerStarted","Data":"7db6edb34f93eb0d1c2adcd3461408006f8a86257dba9d9ea35c1b93c50a9f25"} Mar 13 10:29:21 crc kubenswrapper[4632]: I0313 10:29:21.603959 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:29:21 crc kubenswrapper[4632]: I0313 10:29:21.604191 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="71d36cc3-a7e3-47e5-b98f-599dc669ccc5" containerName="nova-api-log" containerID="cri-o://4df2156f6fe32fab45f05d256a8ec2adb23f786a2989c939b92b996a496f122f" gracePeriod=30 Mar 13 10:29:21 crc kubenswrapper[4632]: I0313 10:29:21.604712 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="71d36cc3-a7e3-47e5-b98f-599dc669ccc5" containerName="nova-api-api" containerID="cri-o://e604b5ae6ce92dde6f33a140a99a7c7d5949aebd7f4821ef087f38b50a0e872b" gracePeriod=30 Mar 13 10:29:22 crc kubenswrapper[4632]: I0313 10:29:22.512110 4632 generic.go:334] "Generic (PLEG): container finished" podID="71d36cc3-a7e3-47e5-b98f-599dc669ccc5" containerID="4df2156f6fe32fab45f05d256a8ec2adb23f786a2989c939b92b996a496f122f" exitCode=143 Mar 13 10:29:22 crc kubenswrapper[4632]: I0313 10:29:22.512169 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"71d36cc3-a7e3-47e5-b98f-599dc669ccc5","Type":"ContainerDied","Data":"4df2156f6fe32fab45f05d256a8ec2adb23f786a2989c939b92b996a496f122f"} Mar 13 10:29:22 crc kubenswrapper[4632]: I0313 10:29:22.515543 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"febcbdc5-25a6-46f7-8c06-d6f45624a466","Type":"ContainerStarted","Data":"dc567f2c9ac20bcf2a21914a8653c160f2de73b0622331bb81ae937c887e0ffb"} Mar 13 10:29:22 crc kubenswrapper[4632]: I0313 10:29:22.516974 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 13 10:29:22 crc kubenswrapper[4632]: I0313 10:29:22.545353 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.542492154 podStartE2EDuration="3.542492154s" podCreationTimestamp="2026-03-13 10:29:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:29:22.536373596 +0000 UTC m=+1536.558903739" watchObservedRunningTime="2026-03-13 10:29:22.542492154 +0000 UTC m=+1536.565022297" Mar 13 10:29:24 crc kubenswrapper[4632]: E0313 10:29:24.352682 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:24 crc kubenswrapper[4632]: E0313 10:29:24.354698 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:24 crc kubenswrapper[4632]: E0313 10:29:24.366409 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:24 crc kubenswrapper[4632]: E0313 10:29:24.366486 4632 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="3548645f-4c72-4c75-b1bd-95116d47f6e2" containerName="nova-scheduler-scheduler" Mar 13 10:29:25 crc kubenswrapper[4632]: I0313 10:29:25.395670 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 10:29:25 crc kubenswrapper[4632]: I0313 10:29:25.864954 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-689764498d-rg7vt" podUID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 13 10:29:28 crc kubenswrapper[4632]: I0313 10:29:28.294485 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n4z22" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="registry-server" probeResult="failure" output=< Mar 13 10:29:28 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:29:28 crc kubenswrapper[4632]: > Mar 13 10:29:29 crc kubenswrapper[4632]: E0313 10:29:29.355089 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:29 crc kubenswrapper[4632]: E0313 10:29:29.358331 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:29 crc kubenswrapper[4632]: E0313 10:29:29.372355 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:29 crc kubenswrapper[4632]: E0313 10:29:29.372488 4632 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="3548645f-4c72-4c75-b1bd-95116d47f6e2" containerName="nova-scheduler-scheduler" Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.581488 4632 generic.go:334] "Generic (PLEG): container finished" podID="71d36cc3-a7e3-47e5-b98f-599dc669ccc5" containerID="e604b5ae6ce92dde6f33a140a99a7c7d5949aebd7f4821ef087f38b50a0e872b" exitCode=0 Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.581553 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"71d36cc3-a7e3-47e5-b98f-599dc669ccc5","Type":"ContainerDied","Data":"e604b5ae6ce92dde6f33a140a99a7c7d5949aebd7f4821ef087f38b50a0e872b"} Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.581585 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"71d36cc3-a7e3-47e5-b98f-599dc669ccc5","Type":"ContainerDied","Data":"85c564a65a9f43d06ad1647efab572b9e533b3cb0feadf64be1f6226a656d6e9"} Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.581599 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85c564a65a9f43d06ad1647efab572b9e533b3cb0feadf64be1f6226a656d6e9" Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.651807 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.742337 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-config-data\") pod \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.742462 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8l5th\" (UniqueName: \"kubernetes.io/projected/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-kube-api-access-8l5th\") pod \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.742511 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-logs\") pod \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.742760 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-combined-ca-bundle\") pod \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\" (UID: \"71d36cc3-a7e3-47e5-b98f-599dc669ccc5\") " Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.743152 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-logs" (OuterVolumeSpecName: "logs") pod "71d36cc3-a7e3-47e5-b98f-599dc669ccc5" (UID: "71d36cc3-a7e3-47e5-b98f-599dc669ccc5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.764042 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-kube-api-access-8l5th" (OuterVolumeSpecName: "kube-api-access-8l5th") pod "71d36cc3-a7e3-47e5-b98f-599dc669ccc5" (UID: "71d36cc3-a7e3-47e5-b98f-599dc669ccc5"). InnerVolumeSpecName "kube-api-access-8l5th". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.773669 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71d36cc3-a7e3-47e5-b98f-599dc669ccc5" (UID: "71d36cc3-a7e3-47e5-b98f-599dc669ccc5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.776186 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-config-data" (OuterVolumeSpecName: "config-data") pod "71d36cc3-a7e3-47e5-b98f-599dc669ccc5" (UID: "71d36cc3-a7e3-47e5-b98f-599dc669ccc5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:29 crc kubenswrapper[4632]: E0313 10:29:29.782581 4632 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/364743f09c9046eb7ee9d937ed57e59f3a447250991c2f722abb845ea5ccd856/diff" to get inode usage: stat /var/lib/containers/storage/overlay/364743f09c9046eb7ee9d937ed57e59f3a447250991c2f722abb845ea5ccd856/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_dnsmasq-dns-7888df55c7-mw5p4_904f04cd-8110-4637-8bb4-67c4b83e189b/dnsmasq-dns/0.log" to get inode usage: stat /var/log/pods/openstack_dnsmasq-dns-7888df55c7-mw5p4_904f04cd-8110-4637-8bb4-67c4b83e189b/dnsmasq-dns/0.log: no such file or directory Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.844738 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8l5th\" (UniqueName: \"kubernetes.io/projected/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-kube-api-access-8l5th\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.844806 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.844824 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.844835 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71d36cc3-a7e3-47e5-b98f-599dc669ccc5-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:29 crc kubenswrapper[4632]: I0313 10:29:29.885604 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.592874 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.615163 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.632659 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.646106 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 10:29:30 crc kubenswrapper[4632]: E0313 10:29:30.646742 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71d36cc3-a7e3-47e5-b98f-599dc669ccc5" containerName="nova-api-api" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.646814 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="71d36cc3-a7e3-47e5-b98f-599dc669ccc5" containerName="nova-api-api" Mar 13 10:29:30 crc kubenswrapper[4632]: E0313 10:29:30.646881 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71d36cc3-a7e3-47e5-b98f-599dc669ccc5" containerName="nova-api-log" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.646929 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="71d36cc3-a7e3-47e5-b98f-599dc669ccc5" containerName="nova-api-log" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.647216 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="71d36cc3-a7e3-47e5-b98f-599dc669ccc5" containerName="nova-api-api" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.647284 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="71d36cc3-a7e3-47e5-b98f-599dc669ccc5" containerName="nova-api-log" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.648656 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.652302 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.673194 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.762006 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de71c6bf-377b-44e8-a5fb-e654b259404f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " pod="openstack/nova-api-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.762065 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de71c6bf-377b-44e8-a5fb-e654b259404f-logs\") pod \"nova-api-0\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " pod="openstack/nova-api-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.762160 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjszk\" (UniqueName: \"kubernetes.io/projected/de71c6bf-377b-44e8-a5fb-e654b259404f-kube-api-access-cjszk\") pod \"nova-api-0\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " pod="openstack/nova-api-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.762236 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de71c6bf-377b-44e8-a5fb-e654b259404f-config-data\") pod \"nova-api-0\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " pod="openstack/nova-api-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.863745 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de71c6bf-377b-44e8-a5fb-e654b259404f-config-data\") pod \"nova-api-0\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " pod="openstack/nova-api-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.864216 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de71c6bf-377b-44e8-a5fb-e654b259404f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " pod="openstack/nova-api-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.864337 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de71c6bf-377b-44e8-a5fb-e654b259404f-logs\") pod \"nova-api-0\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " pod="openstack/nova-api-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.864526 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjszk\" (UniqueName: \"kubernetes.io/projected/de71c6bf-377b-44e8-a5fb-e654b259404f-kube-api-access-cjszk\") pod \"nova-api-0\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " pod="openstack/nova-api-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.864883 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de71c6bf-377b-44e8-a5fb-e654b259404f-logs\") pod \"nova-api-0\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " pod="openstack/nova-api-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.871368 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de71c6bf-377b-44e8-a5fb-e654b259404f-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " pod="openstack/nova-api-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.875215 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de71c6bf-377b-44e8-a5fb-e654b259404f-config-data\") pod \"nova-api-0\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " pod="openstack/nova-api-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.880890 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjszk\" (UniqueName: \"kubernetes.io/projected/de71c6bf-377b-44e8-a5fb-e654b259404f-kube-api-access-cjszk\") pod \"nova-api-0\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " pod="openstack/nova-api-0" Mar 13 10:29:30 crc kubenswrapper[4632]: I0313 10:29:30.966850 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 10:29:31 crc kubenswrapper[4632]: I0313 10:29:31.491588 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:29:31 crc kubenswrapper[4632]: I0313 10:29:31.606416 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de71c6bf-377b-44e8-a5fb-e654b259404f","Type":"ContainerStarted","Data":"ee45ab9f33fdda93f7c890750739536f8547b9a2cf6264542af4cb74ce30fa4b"} Mar 13 10:29:32 crc kubenswrapper[4632]: I0313 10:29:32.058222 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71d36cc3-a7e3-47e5-b98f-599dc669ccc5" path="/var/lib/kubelet/pods/71d36cc3-a7e3-47e5-b98f-599dc669ccc5/volumes" Mar 13 10:29:32 crc kubenswrapper[4632]: I0313 10:29:32.615302 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de71c6bf-377b-44e8-a5fb-e654b259404f","Type":"ContainerStarted","Data":"4fa9439436746dc39d49be92d41774ffb73ecdcce50f69453fca69442efcc0cf"} Mar 13 10:29:32 crc kubenswrapper[4632]: I0313 10:29:32.615347 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de71c6bf-377b-44e8-a5fb-e654b259404f","Type":"ContainerStarted","Data":"66957b9fbbc860bd8b0b4ba61ac2afb5edc2532051b2d051493e658595c97c89"} Mar 13 10:29:32 crc kubenswrapper[4632]: I0313 10:29:32.636414 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.636393396 podStartE2EDuration="2.636393396s" podCreationTimestamp="2026-03-13 10:29:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:29:32.630915172 +0000 UTC m=+1546.653445305" watchObservedRunningTime="2026-03-13 10:29:32.636393396 +0000 UTC m=+1546.658923529" Mar 13 10:29:33 crc kubenswrapper[4632]: W0313 10:29:33.747489 4632 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod229b0823_7d97_48cf_9b38_188a3f4ecde3.slice/crio-9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod229b0823_7d97_48cf_9b38_188a3f4ecde3.slice/crio-9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2.scope: no such file or directory Mar 13 10:29:34 crc kubenswrapper[4632]: E0313 10:29:34.023485 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0cabd29_ef3e_4808_8c92_3b032483789e.slice/crio-conmon-71596437e6a2d325baac09e68c289bacf1de16594d0329b6a19a5f5d92099872.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0cabd29_ef3e_4808_8c92_3b032483789e.slice/crio-71596437e6a2d325baac09e68c289bacf1de16594d0329b6a19a5f5d92099872.scope\": RecentStats: unable to find data in memory cache]" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.345526 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.352234 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 10:29:34 crc kubenswrapper[4632]: E0313 10:29:34.357065 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:34 crc kubenswrapper[4632]: E0313 10:29:34.359382 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:34 crc kubenswrapper[4632]: E0313 10:29:34.367263 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:29:34 crc kubenswrapper[4632]: E0313 10:29:34.367337 4632 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="3548645f-4c72-4c75-b1bd-95116d47f6e2" containerName="nova-scheduler-scheduler" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.478153 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7cq6\" (UniqueName: \"kubernetes.io/projected/229b0823-7d97-48cf-9b38-188a3f4ecde3-kube-api-access-c7cq6\") pod \"229b0823-7d97-48cf-9b38-188a3f4ecde3\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.478269 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/229b0823-7d97-48cf-9b38-188a3f4ecde3-logs\") pod \"229b0823-7d97-48cf-9b38-188a3f4ecde3\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.478416 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlgdp\" (UniqueName: \"kubernetes.io/projected/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-kube-api-access-hlgdp\") pod \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\" (UID: \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\") " Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.478462 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-combined-ca-bundle\") pod \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\" (UID: \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\") " Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.478652 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229b0823-7d97-48cf-9b38-188a3f4ecde3-combined-ca-bundle\") pod \"229b0823-7d97-48cf-9b38-188a3f4ecde3\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.478697 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-config-data\") pod \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\" (UID: \"ac666e99-860b-4f76-8b34-0ac5d3f67e9e\") " Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.478743 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/229b0823-7d97-48cf-9b38-188a3f4ecde3-config-data\") pod \"229b0823-7d97-48cf-9b38-188a3f4ecde3\" (UID: \"229b0823-7d97-48cf-9b38-188a3f4ecde3\") " Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.481051 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/229b0823-7d97-48cf-9b38-188a3f4ecde3-logs" (OuterVolumeSpecName: "logs") pod "229b0823-7d97-48cf-9b38-188a3f4ecde3" (UID: "229b0823-7d97-48cf-9b38-188a3f4ecde3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.481438 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/229b0823-7d97-48cf-9b38-188a3f4ecde3-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.485028 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-kube-api-access-hlgdp" (OuterVolumeSpecName: "kube-api-access-hlgdp") pod "ac666e99-860b-4f76-8b34-0ac5d3f67e9e" (UID: "ac666e99-860b-4f76-8b34-0ac5d3f67e9e"). InnerVolumeSpecName "kube-api-access-hlgdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.486764 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/229b0823-7d97-48cf-9b38-188a3f4ecde3-kube-api-access-c7cq6" (OuterVolumeSpecName: "kube-api-access-c7cq6") pod "229b0823-7d97-48cf-9b38-188a3f4ecde3" (UID: "229b0823-7d97-48cf-9b38-188a3f4ecde3"). InnerVolumeSpecName "kube-api-access-c7cq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.520763 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/229b0823-7d97-48cf-9b38-188a3f4ecde3-config-data" (OuterVolumeSpecName: "config-data") pod "229b0823-7d97-48cf-9b38-188a3f4ecde3" (UID: "229b0823-7d97-48cf-9b38-188a3f4ecde3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.525057 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-config-data" (OuterVolumeSpecName: "config-data") pod "ac666e99-860b-4f76-8b34-0ac5d3f67e9e" (UID: "ac666e99-860b-4f76-8b34-0ac5d3f67e9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.527399 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/229b0823-7d97-48cf-9b38-188a3f4ecde3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "229b0823-7d97-48cf-9b38-188a3f4ecde3" (UID: "229b0823-7d97-48cf-9b38-188a3f4ecde3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.536895 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac666e99-860b-4f76-8b34-0ac5d3f67e9e" (UID: "ac666e99-860b-4f76-8b34-0ac5d3f67e9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.583106 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlgdp\" (UniqueName: \"kubernetes.io/projected/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-kube-api-access-hlgdp\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.583409 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.583467 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/229b0823-7d97-48cf-9b38-188a3f4ecde3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.583517 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac666e99-860b-4f76-8b34-0ac5d3f67e9e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.583583 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/229b0823-7d97-48cf-9b38-188a3f4ecde3-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.583638 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7cq6\" (UniqueName: \"kubernetes.io/projected/229b0823-7d97-48cf-9b38-188a3f4ecde3-kube-api-access-c7cq6\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.638986 4632 generic.go:334] "Generic (PLEG): container finished" podID="229b0823-7d97-48cf-9b38-188a3f4ecde3" containerID="c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5" exitCode=137 Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.639062 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"229b0823-7d97-48cf-9b38-188a3f4ecde3","Type":"ContainerDied","Data":"c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5"} Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.639096 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"229b0823-7d97-48cf-9b38-188a3f4ecde3","Type":"ContainerDied","Data":"e09e0609bf87522d60a55be578e8526df6a87fdb2167dcc1dc7feca3fdedd742"} Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.639115 4632 scope.go:117] "RemoveContainer" containerID="c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.639262 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.646517 4632 generic.go:334] "Generic (PLEG): container finished" podID="ac666e99-860b-4f76-8b34-0ac5d3f67e9e" containerID="a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a" exitCode=137 Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.646563 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ac666e99-860b-4f76-8b34-0ac5d3f67e9e","Type":"ContainerDied","Data":"a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a"} Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.646587 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ac666e99-860b-4f76-8b34-0ac5d3f67e9e","Type":"ContainerDied","Data":"21ad142ba62995b1f6903534cfcf5ce85e20fd10841d57b3365ce522bedc21e5"} Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.646697 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.683008 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.685551 4632 scope.go:117] "RemoveContainer" containerID="9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.727184 4632 scope.go:117] "RemoveContainer" containerID="c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5" Mar 13 10:29:34 crc kubenswrapper[4632]: E0313 10:29:34.728961 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5\": container with ID starting with c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5 not found: ID does not exist" containerID="c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.729004 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5"} err="failed to get container status \"c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5\": rpc error: code = NotFound desc = could not find container \"c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5\": container with ID starting with c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5 not found: ID does not exist" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.729032 4632 scope.go:117] "RemoveContainer" containerID="9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.734153 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:29:34 crc kubenswrapper[4632]: E0313 10:29:34.737287 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2\": container with ID starting with 9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2 not found: ID does not exist" containerID="9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.737341 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2"} err="failed to get container status \"9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2\": rpc error: code = NotFound desc = could not find container \"9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2\": container with ID starting with 9461ca171120686afab14291efb3881e8b77ff7ec5bc19c4bdf3f4e55da83af2 not found: ID does not exist" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.737376 4632 scope.go:117] "RemoveContainer" containerID="a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.763147 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.780055 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.793087 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:29:34 crc kubenswrapper[4632]: E0313 10:29:34.793638 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="229b0823-7d97-48cf-9b38-188a3f4ecde3" containerName="nova-metadata-metadata" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.793664 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="229b0823-7d97-48cf-9b38-188a3f4ecde3" containerName="nova-metadata-metadata" Mar 13 10:29:34 crc kubenswrapper[4632]: E0313 10:29:34.793680 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="229b0823-7d97-48cf-9b38-188a3f4ecde3" containerName="nova-metadata-log" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.793689 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="229b0823-7d97-48cf-9b38-188a3f4ecde3" containerName="nova-metadata-log" Mar 13 10:29:34 crc kubenswrapper[4632]: E0313 10:29:34.793739 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac666e99-860b-4f76-8b34-0ac5d3f67e9e" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.793751 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac666e99-860b-4f76-8b34-0ac5d3f67e9e" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.794000 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="229b0823-7d97-48cf-9b38-188a3f4ecde3" containerName="nova-metadata-log" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.794026 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="229b0823-7d97-48cf-9b38-188a3f4ecde3" containerName="nova-metadata-metadata" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.794048 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac666e99-860b-4f76-8b34-0ac5d3f67e9e" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.796579 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.802603 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.802983 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.807334 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.816403 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.822374 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.822708 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.822906 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.828186 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.866214 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.892192 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9zdc\" (UniqueName: \"kubernetes.io/projected/bf01307f-1529-4aa7-95fc-8af84b061970-kube-api-access-j9zdc\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.892294 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-logs\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.892360 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf01307f-1529-4aa7-95fc-8af84b061970-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.892469 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf01307f-1529-4aa7-95fc-8af84b061970-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.892535 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf01307f-1529-4aa7-95fc-8af84b061970-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.892566 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-config-data\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.892595 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdwzm\" (UniqueName: \"kubernetes.io/projected/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-kube-api-access-rdwzm\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.892635 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.892667 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf01307f-1529-4aa7-95fc-8af84b061970-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.892867 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.894411 4632 scope.go:117] "RemoveContainer" containerID="a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a" Mar 13 10:29:34 crc kubenswrapper[4632]: E0313 10:29:34.894777 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a\": container with ID starting with a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a not found: ID does not exist" containerID="a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.894809 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a"} err="failed to get container status \"a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a\": rpc error: code = NotFound desc = could not find container \"a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a\": container with ID starting with a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a not found: ID does not exist" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.994973 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-logs\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.995234 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf01307f-1529-4aa7-95fc-8af84b061970-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.995366 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf01307f-1529-4aa7-95fc-8af84b061970-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.995481 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf01307f-1529-4aa7-95fc-8af84b061970-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.995567 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-logs\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.995583 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-config-data\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.995737 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwzm\" (UniqueName: \"kubernetes.io/projected/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-kube-api-access-rdwzm\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.995846 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.996006 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf01307f-1529-4aa7-95fc-8af84b061970-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.996154 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:34 crc kubenswrapper[4632]: I0313 10:29:34.996290 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9zdc\" (UniqueName: \"kubernetes.io/projected/bf01307f-1529-4aa7-95fc-8af84b061970-kube-api-access-j9zdc\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.001717 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf01307f-1529-4aa7-95fc-8af84b061970-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.001720 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf01307f-1529-4aa7-95fc-8af84b061970-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.014264 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.014870 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-config-data\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.018589 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.018698 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf01307f-1529-4aa7-95fc-8af84b061970-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.019098 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwzm\" (UniqueName: \"kubernetes.io/projected/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-kube-api-access-rdwzm\") pod \"nova-metadata-0\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " pod="openstack/nova-metadata-0" Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.020637 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf01307f-1529-4aa7-95fc-8af84b061970-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.035630 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9zdc\" (UniqueName: \"kubernetes.io/projected/bf01307f-1529-4aa7-95fc-8af84b061970-kube-api-access-j9zdc\") pod \"nova-cell1-novncproxy-0\" (UID: \"bf01307f-1529-4aa7-95fc-8af84b061970\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.182765 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.198312 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.399380 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.151:8443: connect: connection refused" Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.816794 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.833847 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 10:29:35 crc kubenswrapper[4632]: I0313 10:29:35.857220 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-689764498d-rg7vt" podUID="5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.152:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.152:8443: connect: connection refused" Mar 13 10:29:35 crc kubenswrapper[4632]: W0313 10:29:35.918407 4632 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b3f0cc6_72ae_4738_baee_ce7bc9ef2998.slice/crio-f1168e3921eed21e9e9f3fc46e51a896484996f89685c7e47f03c083dd2451a8.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b3f0cc6_72ae_4738_baee_ce7bc9ef2998.slice/crio-f1168e3921eed21e9e9f3fc46e51a896484996f89685c7e47f03c083dd2451a8.scope: no such file or directory Mar 13 10:29:35 crc kubenswrapper[4632]: W0313 10:29:35.918467 4632 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod229b0823_7d97_48cf_9b38_188a3f4ecde3.slice/crio-conmon-c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod229b0823_7d97_48cf_9b38_188a3f4ecde3.slice/crio-conmon-c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5.scope: no such file or directory Mar 13 10:29:35 crc kubenswrapper[4632]: W0313 10:29:35.918487 4632 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod229b0823_7d97_48cf_9b38_188a3f4ecde3.slice/crio-c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod229b0823_7d97_48cf_9b38_188a3f4ecde3.slice/crio-c37b72b8d98bcfa61d62d487a54c1380919b990c2a3301633445f9ec65005bf5.scope: no such file or directory Mar 13 10:29:35 crc kubenswrapper[4632]: W0313 10:29:35.918505 4632 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b3f0cc6_72ae_4738_baee_ce7bc9ef2998.slice/crio-conmon-29733198b90e2c4e70c2780c9fc89c7f6fe69e2a7c8898353f2770088741eed7.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b3f0cc6_72ae_4738_baee_ce7bc9ef2998.slice/crio-conmon-29733198b90e2c4e70c2780c9fc89c7f6fe69e2a7c8898353f2770088741eed7.scope: no such file or directory Mar 13 10:29:35 crc kubenswrapper[4632]: W0313 10:29:35.918521 4632 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b3f0cc6_72ae_4738_baee_ce7bc9ef2998.slice/crio-29733198b90e2c4e70c2780c9fc89c7f6fe69e2a7c8898353f2770088741eed7.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1b3f0cc6_72ae_4738_baee_ce7bc9ef2998.slice/crio-29733198b90e2c4e70c2780c9fc89c7f6fe69e2a7c8898353f2770088741eed7.scope: no such file or directory Mar 13 10:29:35 crc kubenswrapper[4632]: W0313 10:29:35.935519 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac666e99_860b_4f76_8b34_0ac5d3f67e9e.slice/crio-a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a.scope WatchSource:0}: Error finding container a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a: Status 404 returned error can't find the container with id a0afd63ece3b9f8f07f13d5cb18fecbc9a68899e19cf0a855e21117da83b326a Mar 13 10:29:35 crc kubenswrapper[4632]: W0313 10:29:35.945008 4632 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71d36cc3_a7e3_47e5_b98f_599dc669ccc5.slice": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod71d36cc3_a7e3_47e5_b98f_599dc669ccc5.slice: no such file or directory Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.072457 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="229b0823-7d97-48cf-9b38-188a3f4ecde3" path="/var/lib/kubelet/pods/229b0823-7d97-48cf-9b38-188a3f4ecde3/volumes" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.073289 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac666e99-860b-4f76-8b34-0ac5d3f67e9e" path="/var/lib/kubelet/pods/ac666e99-860b-4f76-8b34-0ac5d3f67e9e/volumes" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.316700 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.441899 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tntfg\" (UniqueName: \"kubernetes.io/projected/3548645f-4c72-4c75-b1bd-95116d47f6e2-kube-api-access-tntfg\") pod \"3548645f-4c72-4c75-b1bd-95116d47f6e2\" (UID: \"3548645f-4c72-4c75-b1bd-95116d47f6e2\") " Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.441981 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3548645f-4c72-4c75-b1bd-95116d47f6e2-combined-ca-bundle\") pod \"3548645f-4c72-4c75-b1bd-95116d47f6e2\" (UID: \"3548645f-4c72-4c75-b1bd-95116d47f6e2\") " Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.442029 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3548645f-4c72-4c75-b1bd-95116d47f6e2-config-data\") pod \"3548645f-4c72-4c75-b1bd-95116d47f6e2\" (UID: \"3548645f-4c72-4c75-b1bd-95116d47f6e2\") " Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.451078 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3548645f-4c72-4c75-b1bd-95116d47f6e2-kube-api-access-tntfg" (OuterVolumeSpecName: "kube-api-access-tntfg") pod "3548645f-4c72-4c75-b1bd-95116d47f6e2" (UID: "3548645f-4c72-4c75-b1bd-95116d47f6e2"). InnerVolumeSpecName "kube-api-access-tntfg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.484675 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3548645f-4c72-4c75-b1bd-95116d47f6e2-config-data" (OuterVolumeSpecName: "config-data") pod "3548645f-4c72-4c75-b1bd-95116d47f6e2" (UID: "3548645f-4c72-4c75-b1bd-95116d47f6e2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.487103 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3548645f-4c72-4c75-b1bd-95116d47f6e2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3548645f-4c72-4c75-b1bd-95116d47f6e2" (UID: "3548645f-4c72-4c75-b1bd-95116d47f6e2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.544017 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tntfg\" (UniqueName: \"kubernetes.io/projected/3548645f-4c72-4c75-b1bd-95116d47f6e2-kube-api-access-tntfg\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.544062 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3548645f-4c72-4c75-b1bd-95116d47f6e2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.544075 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3548645f-4c72-4c75-b1bd-95116d47f6e2-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.690407 4632 generic.go:334] "Generic (PLEG): container finished" podID="3548645f-4c72-4c75-b1bd-95116d47f6e2" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" exitCode=137 Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.690461 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3548645f-4c72-4c75-b1bd-95116d47f6e2","Type":"ContainerDied","Data":"cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713"} Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.690500 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"3548645f-4c72-4c75-b1bd-95116d47f6e2","Type":"ContainerDied","Data":"2e867316abee7ad3f39e6bce10fac89a2148bd2f0cb280bc0cfe2baf59687b11"} Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.690516 4632 scope.go:117] "RemoveContainer" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.690443 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.694354 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82ae3a46-0133-43f5-942d-0b9a5b4d59f4","Type":"ContainerStarted","Data":"a8f2d5b445eff9d0451f4c86782cfcac63ef30c3164366e7abf62cc09495ddd8"} Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.694535 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82ae3a46-0133-43f5-942d-0b9a5b4d59f4","Type":"ContainerStarted","Data":"519f8117c2e26f4a0b3f5e7e157a107d38c388ac29acd5faf2f4b4ecf121e55f"} Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.694621 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82ae3a46-0133-43f5-942d-0b9a5b4d59f4","Type":"ContainerStarted","Data":"d71e5927086b73530c1dba7fd1700212b2fab56fed475076dc30c04ba970bcf7"} Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.702641 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bf01307f-1529-4aa7-95fc-8af84b061970","Type":"ContainerStarted","Data":"d38babf4e97712ec891a6f38a9738c7611fa766a76e6a294604ae52bc653604f"} Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.702869 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"bf01307f-1529-4aa7-95fc-8af84b061970","Type":"ContainerStarted","Data":"38f25e4565386591341a7ca73af278dbdad469f10eb262b202099e00fa3206fb"} Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.727179 4632 scope.go:117] "RemoveContainer" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" Mar 13 10:29:36 crc kubenswrapper[4632]: E0313 10:29:36.727857 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713\": container with ID starting with cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713 not found: ID does not exist" containerID="cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.727902 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713"} err="failed to get container status \"cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713\": rpc error: code = NotFound desc = could not find container \"cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713\": container with ID starting with cd653c258a170cf57afe5c4a490b6a607d8ebe8fe6115c40d2e2f88f1ae5f713 not found: ID does not exist" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.740382 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.74036559 podStartE2EDuration="2.74036559s" podCreationTimestamp="2026-03-13 10:29:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:29:36.724281679 +0000 UTC m=+1550.746811822" watchObservedRunningTime="2026-03-13 10:29:36.74036559 +0000 UTC m=+1550.762895723" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.750253 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.772015 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.776668 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.776645732 podStartE2EDuration="2.776645732s" podCreationTimestamp="2026-03-13 10:29:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:29:36.757048906 +0000 UTC m=+1550.779579039" watchObservedRunningTime="2026-03-13 10:29:36.776645732 +0000 UTC m=+1550.799175865" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.822320 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:29:36 crc kubenswrapper[4632]: E0313 10:29:36.823775 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3548645f-4c72-4c75-b1bd-95116d47f6e2" containerName="nova-scheduler-scheduler" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.823796 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3548645f-4c72-4c75-b1bd-95116d47f6e2" containerName="nova-scheduler-scheduler" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.824053 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3548645f-4c72-4c75-b1bd-95116d47f6e2" containerName="nova-scheduler-scheduler" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.826618 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.836798 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.886731 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.953929 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d0c4f9f-780f-42d8-9eee-cb2201034218-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4d0c4f9f-780f-42d8-9eee-cb2201034218\") " pod="openstack/nova-scheduler-0" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.954173 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d0c4f9f-780f-42d8-9eee-cb2201034218-config-data\") pod \"nova-scheduler-0\" (UID: \"4d0c4f9f-780f-42d8-9eee-cb2201034218\") " pod="openstack/nova-scheduler-0" Mar 13 10:29:36 crc kubenswrapper[4632]: I0313 10:29:36.954219 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dmpm\" (UniqueName: \"kubernetes.io/projected/4d0c4f9f-780f-42d8-9eee-cb2201034218-kube-api-access-2dmpm\") pod \"nova-scheduler-0\" (UID: \"4d0c4f9f-780f-42d8-9eee-cb2201034218\") " pod="openstack/nova-scheduler-0" Mar 13 10:29:37 crc kubenswrapper[4632]: I0313 10:29:37.055711 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dmpm\" (UniqueName: \"kubernetes.io/projected/4d0c4f9f-780f-42d8-9eee-cb2201034218-kube-api-access-2dmpm\") pod \"nova-scheduler-0\" (UID: \"4d0c4f9f-780f-42d8-9eee-cb2201034218\") " pod="openstack/nova-scheduler-0" Mar 13 10:29:37 crc kubenswrapper[4632]: I0313 10:29:37.055817 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d0c4f9f-780f-42d8-9eee-cb2201034218-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4d0c4f9f-780f-42d8-9eee-cb2201034218\") " pod="openstack/nova-scheduler-0" Mar 13 10:29:37 crc kubenswrapper[4632]: I0313 10:29:37.056061 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d0c4f9f-780f-42d8-9eee-cb2201034218-config-data\") pod \"nova-scheduler-0\" (UID: \"4d0c4f9f-780f-42d8-9eee-cb2201034218\") " pod="openstack/nova-scheduler-0" Mar 13 10:29:37 crc kubenswrapper[4632]: I0313 10:29:37.060897 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d0c4f9f-780f-42d8-9eee-cb2201034218-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4d0c4f9f-780f-42d8-9eee-cb2201034218\") " pod="openstack/nova-scheduler-0" Mar 13 10:29:37 crc kubenswrapper[4632]: I0313 10:29:37.060969 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d0c4f9f-780f-42d8-9eee-cb2201034218-config-data\") pod \"nova-scheduler-0\" (UID: \"4d0c4f9f-780f-42d8-9eee-cb2201034218\") " pod="openstack/nova-scheduler-0" Mar 13 10:29:37 crc kubenswrapper[4632]: I0313 10:29:37.087595 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dmpm\" (UniqueName: \"kubernetes.io/projected/4d0c4f9f-780f-42d8-9eee-cb2201034218-kube-api-access-2dmpm\") pod \"nova-scheduler-0\" (UID: \"4d0c4f9f-780f-42d8-9eee-cb2201034218\") " pod="openstack/nova-scheduler-0" Mar 13 10:29:37 crc kubenswrapper[4632]: I0313 10:29:37.103928 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.214:3000/\": dial tcp 10.217.0.214:3000: connect: connection refused" Mar 13 10:29:37 crc kubenswrapper[4632]: I0313 10:29:37.181164 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 10:29:37 crc kubenswrapper[4632]: I0313 10:29:37.893561 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:29:38 crc kubenswrapper[4632]: I0313 10:29:38.082821 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3548645f-4c72-4c75-b1bd-95116d47f6e2" path="/var/lib/kubelet/pods/3548645f-4c72-4c75-b1bd-95116d47f6e2/volumes" Mar 13 10:29:38 crc kubenswrapper[4632]: I0313 10:29:38.413498 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n4z22" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="registry-server" probeResult="failure" output=< Mar 13 10:29:38 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:29:38 crc kubenswrapper[4632]: > Mar 13 10:29:38 crc kubenswrapper[4632]: I0313 10:29:38.727299 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d0c4f9f-780f-42d8-9eee-cb2201034218","Type":"ContainerStarted","Data":"9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58"} Mar 13 10:29:38 crc kubenswrapper[4632]: I0313 10:29:38.728260 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d0c4f9f-780f-42d8-9eee-cb2201034218","Type":"ContainerStarted","Data":"ccfc717e8149e75cbe225885927d06c595e9efeff8370ba3176af49fbdc5eb3d"} Mar 13 10:29:38 crc kubenswrapper[4632]: I0313 10:29:38.760109 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.760086338 podStartE2EDuration="2.760086338s" podCreationTimestamp="2026-03-13 10:29:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:29:38.752093583 +0000 UTC m=+1552.774623716" watchObservedRunningTime="2026-03-13 10:29:38.760086338 +0000 UTC m=+1552.782616471" Mar 13 10:29:40 crc kubenswrapper[4632]: I0313 10:29:40.183783 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 10:29:40 crc kubenswrapper[4632]: I0313 10:29:40.185108 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 10:29:40 crc kubenswrapper[4632]: I0313 10:29:40.198883 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:40 crc kubenswrapper[4632]: I0313 10:29:40.967640 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 10:29:40 crc kubenswrapper[4632]: I0313 10:29:40.968067 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 10:29:42 crc kubenswrapper[4632]: I0313 10:29:42.051177 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="de71c6bf-377b-44e8-a5fb-e654b259404f" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.218:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:42 crc kubenswrapper[4632]: I0313 10:29:42.051220 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="de71c6bf-377b-44e8-a5fb-e654b259404f" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.218:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:42 crc kubenswrapper[4632]: I0313 10:29:42.181754 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 10:29:43 crc kubenswrapper[4632]: E0313 10:29:43.281446 4632 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/d612d79249ad4601fcba1d21d312ec5600b48bd60b162893d856c65bd99ed9fb/diff" to get inode usage: stat /var/lib/containers/storage/overlay/d612d79249ad4601fcba1d21d312ec5600b48bd60b162893d856c65bd99ed9fb/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_nova-cell0-conductor-0_89c5451e-248e-46eb-ac20-f52c3e3bcdc4/nova-cell0-conductor-conductor/0.log" to get inode usage: stat /var/log/pods/openstack_nova-cell0-conductor-0_89c5451e-248e-46eb-ac20-f52c3e3bcdc4/nova-cell0-conductor-conductor/0.log: no such file or directory Mar 13 10:29:43 crc kubenswrapper[4632]: I0313 10:29:43.831053 4632 generic.go:334] "Generic (PLEG): container finished" podID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerID="b21cf0ac61b24fc322cab134fff86780fa6ec89891d057825ed89069f94de13b" exitCode=137 Mar 13 10:29:43 crc kubenswrapper[4632]: I0313 10:29:43.831186 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae6e23c-4223-4a5c-8074-0cdfb5d99e78","Type":"ContainerDied","Data":"b21cf0ac61b24fc322cab134fff86780fa6ec89891d057825ed89069f94de13b"} Mar 13 10:29:43 crc kubenswrapper[4632]: I0313 10:29:43.971963 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.011677 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vtzmr\" (UniqueName: \"kubernetes.io/projected/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-kube-api-access-vtzmr\") pod \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.011736 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-config-data\") pod \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.011870 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-run-httpd\") pod \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.012051 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-combined-ca-bundle\") pod \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.012096 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-log-httpd\") pod \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.012135 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-scripts\") pod \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.012161 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-sg-core-conf-yaml\") pod \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\" (UID: \"aae6e23c-4223-4a5c-8074-0cdfb5d99e78\") " Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.014222 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "aae6e23c-4223-4a5c-8074-0cdfb5d99e78" (UID: "aae6e23c-4223-4a5c-8074-0cdfb5d99e78"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.016576 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "aae6e23c-4223-4a5c-8074-0cdfb5d99e78" (UID: "aae6e23c-4223-4a5c-8074-0cdfb5d99e78"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.024809 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-kube-api-access-vtzmr" (OuterVolumeSpecName: "kube-api-access-vtzmr") pod "aae6e23c-4223-4a5c-8074-0cdfb5d99e78" (UID: "aae6e23c-4223-4a5c-8074-0cdfb5d99e78"). InnerVolumeSpecName "kube-api-access-vtzmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.033278 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-scripts" (OuterVolumeSpecName: "scripts") pod "aae6e23c-4223-4a5c-8074-0cdfb5d99e78" (UID: "aae6e23c-4223-4a5c-8074-0cdfb5d99e78"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.091843 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "aae6e23c-4223-4a5c-8074-0cdfb5d99e78" (UID: "aae6e23c-4223-4a5c-8074-0cdfb5d99e78"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.114672 4632 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.114716 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.114731 4632 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.114748 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vtzmr\" (UniqueName: \"kubernetes.io/projected/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-kube-api-access-vtzmr\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.114757 4632 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.130375 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aae6e23c-4223-4a5c-8074-0cdfb5d99e78" (UID: "aae6e23c-4223-4a5c-8074-0cdfb5d99e78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.186892 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-config-data" (OuterVolumeSpecName: "config-data") pod "aae6e23c-4223-4a5c-8074-0cdfb5d99e78" (UID: "aae6e23c-4223-4a5c-8074-0cdfb5d99e78"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.217022 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.217068 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aae6e23c-4223-4a5c-8074-0cdfb5d99e78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.843927 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"aae6e23c-4223-4a5c-8074-0cdfb5d99e78","Type":"ContainerDied","Data":"ac0a78784a347100c3f7eb503566ca534d51cff4225e41d7e119e9bf5dd8e6ea"} Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.844189 4632 scope.go:117] "RemoveContainer" containerID="604c723ae68fd3213f4723bdf3524877331344576a1c363f42185040543011d1" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.844314 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.898729 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.912782 4632 scope.go:117] "RemoveContainer" containerID="585bc9dbde7bcb94e64b2121baa3ef51bee81b44562ad73e38572b5035710787" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.915785 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.927302 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:29:44 crc kubenswrapper[4632]: E0313 10:29:44.927988 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="sg-core" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.928105 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="sg-core" Mar 13 10:29:44 crc kubenswrapper[4632]: E0313 10:29:44.928198 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="ceilometer-notification-agent" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.928302 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="ceilometer-notification-agent" Mar 13 10:29:44 crc kubenswrapper[4632]: E0313 10:29:44.928389 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="ceilometer-central-agent" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.928470 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="ceilometer-central-agent" Mar 13 10:29:44 crc kubenswrapper[4632]: E0313 10:29:44.928549 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="proxy-httpd" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.928615 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="proxy-httpd" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.929392 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="ceilometer-notification-agent" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.929533 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="sg-core" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.929625 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="ceilometer-central-agent" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.929708 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" containerName="proxy-httpd" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.932502 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.937621 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.937768 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.950670 4632 scope.go:117] "RemoveContainer" containerID="b21cf0ac61b24fc322cab134fff86780fa6ec89891d057825ed89069f94de13b" Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.954316 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:29:44 crc kubenswrapper[4632]: I0313 10:29:44.980638 4632 scope.go:117] "RemoveContainer" containerID="fb2da7115022d9a5776bf008164ce9b7e7fedf403dd765c92f96767951f31f8f" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.035202 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.035274 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff318cc9-cbe7-4357-971a-26c26e8bd269-run-httpd\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.035314 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qm5d\" (UniqueName: \"kubernetes.io/projected/ff318cc9-cbe7-4357-971a-26c26e8bd269-kube-api-access-2qm5d\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.035353 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-scripts\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.035433 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.035504 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-config-data\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.035529 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff318cc9-cbe7-4357-971a-26c26e8bd269-log-httpd\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.137278 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qm5d\" (UniqueName: \"kubernetes.io/projected/ff318cc9-cbe7-4357-971a-26c26e8bd269-kube-api-access-2qm5d\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.137349 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-scripts\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.137440 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.137538 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-config-data\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.137562 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff318cc9-cbe7-4357-971a-26c26e8bd269-log-httpd\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.137630 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.137660 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff318cc9-cbe7-4357-971a-26c26e8bd269-run-httpd\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.138195 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff318cc9-cbe7-4357-971a-26c26e8bd269-run-httpd\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.140735 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff318cc9-cbe7-4357-971a-26c26e8bd269-log-httpd\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.145325 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.146010 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-scripts\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.146970 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.151442 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-config-data\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.163891 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qm5d\" (UniqueName: \"kubernetes.io/projected/ff318cc9-cbe7-4357-971a-26c26e8bd269-kube-api-access-2qm5d\") pod \"ceilometer-0\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.183779 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.184028 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.199171 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.234197 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.255233 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.788667 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:29:45 crc kubenswrapper[4632]: W0313 10:29:45.799688 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff318cc9_cbe7_4357_971a_26c26e8bd269.slice/crio-2f6d41c40b5de2ff95100617c3719cca8da01adbe4c5436cbe7b0e955e7ff656 WatchSource:0}: Error finding container 2f6d41c40b5de2ff95100617c3719cca8da01adbe4c5436cbe7b0e955e7ff656: Status 404 returned error can't find the container with id 2f6d41c40b5de2ff95100617c3719cca8da01adbe4c5436cbe7b0e955e7ff656 Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.841356 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.864710 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff318cc9-cbe7-4357-971a-26c26e8bd269","Type":"ContainerStarted","Data":"2f6d41c40b5de2ff95100617c3719cca8da01adbe4c5436cbe7b0e955e7ff656"} Mar 13 10:29:45 crc kubenswrapper[4632]: I0313 10:29:45.889748 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.119138 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aae6e23c-4223-4a5c-8074-0cdfb5d99e78" path="/var/lib/kubelet/pods/aae6e23c-4223-4a5c-8074-0cdfb5d99e78/volumes" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.241115 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-ngzsx"] Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.243039 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.244142 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.219:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.244466 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.219:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.248340 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.248792 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.253339 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-ngzsx"] Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.278753 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-config-data\") pod \"nova-cell1-cell-mapping-ngzsx\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.278846 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4fj2\" (UniqueName: \"kubernetes.io/projected/601f3615-5015-486a-bbb5-04c683da6990-kube-api-access-q4fj2\") pod \"nova-cell1-cell-mapping-ngzsx\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.278870 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-scripts\") pod \"nova-cell1-cell-mapping-ngzsx\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.278888 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ngzsx\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.382668 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-config-data\") pod \"nova-cell1-cell-mapping-ngzsx\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.382890 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4fj2\" (UniqueName: \"kubernetes.io/projected/601f3615-5015-486a-bbb5-04c683da6990-kube-api-access-q4fj2\") pod \"nova-cell1-cell-mapping-ngzsx\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.382925 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-scripts\") pod \"nova-cell1-cell-mapping-ngzsx\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.383045 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ngzsx\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.396255 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-scripts\") pod \"nova-cell1-cell-mapping-ngzsx\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.399920 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-ngzsx\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.401267 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-config-data\") pod \"nova-cell1-cell-mapping-ngzsx\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.412071 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4fj2\" (UniqueName: \"kubernetes.io/projected/601f3615-5015-486a-bbb5-04c683da6990-kube-api-access-q4fj2\") pod \"nova-cell1-cell-mapping-ngzsx\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.642163 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:46 crc kubenswrapper[4632]: I0313 10:29:46.895683 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff318cc9-cbe7-4357-971a-26c26e8bd269","Type":"ContainerStarted","Data":"e19df6f77d1da54d66347bac1ac445e20a6a2fb793251d7691739401089385b1"} Mar 13 10:29:47 crc kubenswrapper[4632]: I0313 10:29:47.151888 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-ngzsx"] Mar 13 10:29:47 crc kubenswrapper[4632]: I0313 10:29:47.182128 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 10:29:47 crc kubenswrapper[4632]: I0313 10:29:47.219924 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 10:29:47 crc kubenswrapper[4632]: I0313 10:29:47.906398 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ngzsx" event={"ID":"601f3615-5015-486a-bbb5-04c683da6990","Type":"ContainerStarted","Data":"e181311595cfc3a50154df8d12fbc0793d907a3185d962d8a64fc357e0b6ee4f"} Mar 13 10:29:47 crc kubenswrapper[4632]: I0313 10:29:47.906788 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ngzsx" event={"ID":"601f3615-5015-486a-bbb5-04c683da6990","Type":"ContainerStarted","Data":"25ba33658addfd86f3973f0bbd3f7d31da0188bcd4762d3214f38a7e8af5297b"} Mar 13 10:29:47 crc kubenswrapper[4632]: I0313 10:29:47.912366 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff318cc9-cbe7-4357-971a-26c26e8bd269","Type":"ContainerStarted","Data":"ca42f2855aed10e6d2409a495733d8e631d4d8011dee434ee91ebcbee2777ca4"} Mar 13 10:29:47 crc kubenswrapper[4632]: I0313 10:29:47.912420 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff318cc9-cbe7-4357-971a-26c26e8bd269","Type":"ContainerStarted","Data":"cd9daf2ae40dbbc090305eb1d55ec458cd7302b2ff730d20e1f7272ead3f43c4"} Mar 13 10:29:47 crc kubenswrapper[4632]: I0313 10:29:47.930435 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-ngzsx" podStartSLOduration=1.930413961 podStartE2EDuration="1.930413961s" podCreationTimestamp="2026-03-13 10:29:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:29:47.925323789 +0000 UTC m=+1561.947853922" watchObservedRunningTime="2026-03-13 10:29:47.930413961 +0000 UTC m=+1561.952944094" Mar 13 10:29:47 crc kubenswrapper[4632]: I0313 10:29:47.942652 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 10:29:48 crc kubenswrapper[4632]: I0313 10:29:48.309840 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n4z22" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="registry-server" probeResult="failure" output=< Mar 13 10:29:48 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:29:48 crc kubenswrapper[4632]: > Mar 13 10:29:49 crc kubenswrapper[4632]: I0313 10:29:49.933252 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff318cc9-cbe7-4357-971a-26c26e8bd269","Type":"ContainerStarted","Data":"e81d93539be5ad8110d788d8abb89dd69059700a39392e03a27cde23af0ce544"} Mar 13 10:29:49 crc kubenswrapper[4632]: I0313 10:29:49.934683 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 10:29:49 crc kubenswrapper[4632]: I0313 10:29:49.970212 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.592511212 podStartE2EDuration="5.970189454s" podCreationTimestamp="2026-03-13 10:29:44 +0000 UTC" firstStartedPulling="2026-03-13 10:29:45.834730312 +0000 UTC m=+1559.857260445" lastFinishedPulling="2026-03-13 10:29:49.212408554 +0000 UTC m=+1563.234938687" observedRunningTime="2026-03-13 10:29:49.968278078 +0000 UTC m=+1563.990808231" watchObservedRunningTime="2026-03-13 10:29:49.970189454 +0000 UTC m=+1563.992719597" Mar 13 10:29:50 crc kubenswrapper[4632]: I0313 10:29:50.401184 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.151:8443/dashboard/auth/login/?next=/dashboard/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:50 crc kubenswrapper[4632]: I0313 10:29:50.401289 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:29:50 crc kubenswrapper[4632]: I0313 10:29:50.402593 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="horizon" containerStatusID={"Type":"cri-o","ID":"0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30"} pod="openstack/horizon-7bdb5f7878-ng2k2" containerMessage="Container horizon failed startup probe, will be restarted" Mar 13 10:29:50 crc kubenswrapper[4632]: I0313 10:29:50.402647 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" containerID="cri-o://0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30" gracePeriod=30 Mar 13 10:29:50 crc kubenswrapper[4632]: I0313 10:29:50.446395 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:29:50 crc kubenswrapper[4632]: I0313 10:29:50.973203 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 10:29:50 crc kubenswrapper[4632]: I0313 10:29:50.974158 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 10:29:50 crc kubenswrapper[4632]: I0313 10:29:50.977514 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 10:29:50 crc kubenswrapper[4632]: I0313 10:29:50.986873 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 10:29:51 crc kubenswrapper[4632]: I0313 10:29:51.952209 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 10:29:51 crc kubenswrapper[4632]: I0313 10:29:51.955738 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.184733 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-564797cccc-84dg2"] Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.186925 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.205140 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djmgg\" (UniqueName: \"kubernetes.io/projected/ac568760-fbe3-49ca-af4a-13f7780a1ad2-kube-api-access-djmgg\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.205206 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-ovsdbserver-nb\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.205237 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-config\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.205278 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-dns-swift-storage-0\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.205307 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-dns-svc\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.205342 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-ovsdbserver-sb\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.223290 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-564797cccc-84dg2"] Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.306745 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-config\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.306838 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-dns-swift-storage-0\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.306886 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-dns-svc\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.306957 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-ovsdbserver-sb\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.307078 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djmgg\" (UniqueName: \"kubernetes.io/projected/ac568760-fbe3-49ca-af4a-13f7780a1ad2-kube-api-access-djmgg\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.307145 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-ovsdbserver-nb\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.308540 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-config\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.308787 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-dns-svc\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.309016 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-ovsdbserver-sb\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.309218 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-dns-swift-storage-0\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.310182 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-ovsdbserver-nb\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.341258 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djmgg\" (UniqueName: \"kubernetes.io/projected/ac568760-fbe3-49ca-af4a-13f7780a1ad2-kube-api-access-djmgg\") pod \"dnsmasq-dns-564797cccc-84dg2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:52 crc kubenswrapper[4632]: I0313 10:29:52.539570 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:53 crc kubenswrapper[4632]: I0313 10:29:53.133310 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-564797cccc-84dg2"] Mar 13 10:29:53 crc kubenswrapper[4632]: I0313 10:29:53.974581 4632 generic.go:334] "Generic (PLEG): container finished" podID="ac568760-fbe3-49ca-af4a-13f7780a1ad2" containerID="7c9783dd40660c9e8665537c8ead9f633309987f7dedc616633d346075b3da86" exitCode=0 Mar 13 10:29:53 crc kubenswrapper[4632]: I0313 10:29:53.974640 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-564797cccc-84dg2" event={"ID":"ac568760-fbe3-49ca-af4a-13f7780a1ad2","Type":"ContainerDied","Data":"7c9783dd40660c9e8665537c8ead9f633309987f7dedc616633d346075b3da86"} Mar 13 10:29:53 crc kubenswrapper[4632]: I0313 10:29:53.974973 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-564797cccc-84dg2" event={"ID":"ac568760-fbe3-49ca-af4a-13f7780a1ad2","Type":"ContainerStarted","Data":"f114a94f0fb42ce2c1f69bef8fad045098717f536b5431da20286872b08fed02"} Mar 13 10:29:54 crc kubenswrapper[4632]: I0313 10:29:54.823018 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-689764498d-rg7vt" Mar 13 10:29:54 crc kubenswrapper[4632]: I0313 10:29:54.895480 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bdb5f7878-ng2k2"] Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.010331 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-564797cccc-84dg2" event={"ID":"ac568760-fbe3-49ca-af4a-13f7780a1ad2","Type":"ContainerStarted","Data":"3807149ca5beac08d142f3e5ffa3b80f5bf9a97b93a119f317229b5a8536c4a3"} Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.011733 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.029586 4632 generic.go:334] "Generic (PLEG): container finished" podID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerID="0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30" exitCode=0 Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.029632 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb5f7878-ng2k2" event={"ID":"3e4afb91-ce26-4325-89c9-2542da2ec48a","Type":"ContainerDied","Data":"0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30"} Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.029659 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb5f7878-ng2k2" event={"ID":"3e4afb91-ce26-4325-89c9-2542da2ec48a","Type":"ContainerStarted","Data":"d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35"} Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.029677 4632 scope.go:117] "RemoveContainer" containerID="2dbb3ede37abc9f5b483ae48b13ac3ed8913ac4529c34c39494b1541e21ce00b" Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.029906 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon-log" containerID="cri-o://0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589" gracePeriod=30 Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.030077 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7bdb5f7878-ng2k2" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" containerID="cri-o://d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35" gracePeriod=30 Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.049350 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-564797cccc-84dg2" podStartSLOduration=3.049331597 podStartE2EDuration="3.049331597s" podCreationTimestamp="2026-03-13 10:29:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:29:55.029458338 +0000 UTC m=+1569.051988471" watchObservedRunningTime="2026-03-13 10:29:55.049331597 +0000 UTC m=+1569.071861730" Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.192815 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.220456 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.220741 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="de71c6bf-377b-44e8-a5fb-e654b259404f" containerName="nova-api-log" containerID="cri-o://66957b9fbbc860bd8b0b4ba61ac2afb5edc2532051b2d051493e658595c97c89" gracePeriod=30 Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.221297 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="de71c6bf-377b-44e8-a5fb-e654b259404f" containerName="nova-api-api" containerID="cri-o://4fa9439436746dc39d49be92d41774ffb73ecdcce50f69453fca69442efcc0cf" gracePeriod=30 Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.223361 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.223626 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 10:29:55 crc kubenswrapper[4632]: I0313 10:29:55.394137 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:29:56 crc kubenswrapper[4632]: I0313 10:29:56.048342 4632 generic.go:334] "Generic (PLEG): container finished" podID="de71c6bf-377b-44e8-a5fb-e654b259404f" containerID="66957b9fbbc860bd8b0b4ba61ac2afb5edc2532051b2d051493e658595c97c89" exitCode=143 Mar 13 10:29:56 crc kubenswrapper[4632]: I0313 10:29:56.061151 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de71c6bf-377b-44e8-a5fb-e654b259404f","Type":"ContainerDied","Data":"66957b9fbbc860bd8b0b4ba61ac2afb5edc2532051b2d051493e658595c97c89"} Mar 13 10:29:56 crc kubenswrapper[4632]: I0313 10:29:56.146501 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 10:29:57 crc kubenswrapper[4632]: I0313 10:29:57.098139 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:29:57 crc kubenswrapper[4632]: I0313 10:29:57.098731 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="ceilometer-central-agent" containerID="cri-o://e19df6f77d1da54d66347bac1ac445e20a6a2fb793251d7691739401089385b1" gracePeriod=30 Mar 13 10:29:57 crc kubenswrapper[4632]: I0313 10:29:57.100643 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="proxy-httpd" containerID="cri-o://e81d93539be5ad8110d788d8abb89dd69059700a39392e03a27cde23af0ce544" gracePeriod=30 Mar 13 10:29:57 crc kubenswrapper[4632]: I0313 10:29:57.100683 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="ceilometer-notification-agent" containerID="cri-o://cd9daf2ae40dbbc090305eb1d55ec458cd7302b2ff730d20e1f7272ead3f43c4" gracePeriod=30 Mar 13 10:29:57 crc kubenswrapper[4632]: I0313 10:29:57.100638 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="sg-core" containerID="cri-o://ca42f2855aed10e6d2409a495733d8e631d4d8011dee434ee91ebcbee2777ca4" gracePeriod=30 Mar 13 10:29:58 crc kubenswrapper[4632]: I0313 10:29:58.116908 4632 generic.go:334] "Generic (PLEG): container finished" podID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerID="e81d93539be5ad8110d788d8abb89dd69059700a39392e03a27cde23af0ce544" exitCode=0 Mar 13 10:29:58 crc kubenswrapper[4632]: I0313 10:29:58.117203 4632 generic.go:334] "Generic (PLEG): container finished" podID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerID="ca42f2855aed10e6d2409a495733d8e631d4d8011dee434ee91ebcbee2777ca4" exitCode=2 Mar 13 10:29:58 crc kubenswrapper[4632]: I0313 10:29:58.117212 4632 generic.go:334] "Generic (PLEG): container finished" podID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerID="cd9daf2ae40dbbc090305eb1d55ec458cd7302b2ff730d20e1f7272ead3f43c4" exitCode=0 Mar 13 10:29:58 crc kubenswrapper[4632]: I0313 10:29:58.116973 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff318cc9-cbe7-4357-971a-26c26e8bd269","Type":"ContainerDied","Data":"e81d93539be5ad8110d788d8abb89dd69059700a39392e03a27cde23af0ce544"} Mar 13 10:29:58 crc kubenswrapper[4632]: I0313 10:29:58.117273 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff318cc9-cbe7-4357-971a-26c26e8bd269","Type":"ContainerDied","Data":"ca42f2855aed10e6d2409a495733d8e631d4d8011dee434ee91ebcbee2777ca4"} Mar 13 10:29:58 crc kubenswrapper[4632]: I0313 10:29:58.117287 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff318cc9-cbe7-4357-971a-26c26e8bd269","Type":"ContainerDied","Data":"cd9daf2ae40dbbc090305eb1d55ec458cd7302b2ff730d20e1f7272ead3f43c4"} Mar 13 10:29:58 crc kubenswrapper[4632]: I0313 10:29:58.121833 4632 generic.go:334] "Generic (PLEG): container finished" podID="601f3615-5015-486a-bbb5-04c683da6990" containerID="e181311595cfc3a50154df8d12fbc0793d907a3185d962d8a64fc357e0b6ee4f" exitCode=0 Mar 13 10:29:58 crc kubenswrapper[4632]: I0313 10:29:58.122066 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ngzsx" event={"ID":"601f3615-5015-486a-bbb5-04c683da6990","Type":"ContainerDied","Data":"e181311595cfc3a50154df8d12fbc0793d907a3185d962d8a64fc357e0b6ee4f"} Mar 13 10:29:58 crc kubenswrapper[4632]: I0313 10:29:58.307536 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n4z22" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="registry-server" probeResult="failure" output=< Mar 13 10:29:58 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:29:58 crc kubenswrapper[4632]: > Mar 13 10:29:59 crc kubenswrapper[4632]: I0313 10:29:59.133816 4632 generic.go:334] "Generic (PLEG): container finished" podID="de71c6bf-377b-44e8-a5fb-e654b259404f" containerID="4fa9439436746dc39d49be92d41774ffb73ecdcce50f69453fca69442efcc0cf" exitCode=0 Mar 13 10:29:59 crc kubenswrapper[4632]: I0313 10:29:59.134013 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de71c6bf-377b-44e8-a5fb-e654b259404f","Type":"ContainerDied","Data":"4fa9439436746dc39d49be92d41774ffb73ecdcce50f69453fca69442efcc0cf"} Mar 13 10:29:59 crc kubenswrapper[4632]: I0313 10:29:59.813580 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:29:59 crc kubenswrapper[4632]: I0313 10:29:59.828642 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-scripts\") pod \"601f3615-5015-486a-bbb5-04c683da6990\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " Mar 13 10:29:59 crc kubenswrapper[4632]: I0313 10:29:59.828751 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-config-data\") pod \"601f3615-5015-486a-bbb5-04c683da6990\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " Mar 13 10:29:59 crc kubenswrapper[4632]: I0313 10:29:59.828826 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-combined-ca-bundle\") pod \"601f3615-5015-486a-bbb5-04c683da6990\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " Mar 13 10:29:59 crc kubenswrapper[4632]: I0313 10:29:59.828867 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4fj2\" (UniqueName: \"kubernetes.io/projected/601f3615-5015-486a-bbb5-04c683da6990-kube-api-access-q4fj2\") pod \"601f3615-5015-486a-bbb5-04c683da6990\" (UID: \"601f3615-5015-486a-bbb5-04c683da6990\") " Mar 13 10:29:59 crc kubenswrapper[4632]: I0313 10:29:59.859350 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-scripts" (OuterVolumeSpecName: "scripts") pod "601f3615-5015-486a-bbb5-04c683da6990" (UID: "601f3615-5015-486a-bbb5-04c683da6990"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:29:59 crc kubenswrapper[4632]: I0313 10:29:59.881219 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/601f3615-5015-486a-bbb5-04c683da6990-kube-api-access-q4fj2" (OuterVolumeSpecName: "kube-api-access-q4fj2") pod "601f3615-5015-486a-bbb5-04c683da6990" (UID: "601f3615-5015-486a-bbb5-04c683da6990"). InnerVolumeSpecName "kube-api-access-q4fj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:29:59 crc kubenswrapper[4632]: I0313 10:29:59.946637 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4fj2\" (UniqueName: \"kubernetes.io/projected/601f3615-5015-486a-bbb5-04c683da6990-kube-api-access-q4fj2\") on node \"crc\" DevicePath \"\"" Mar 13 10:29:59 crc kubenswrapper[4632]: I0313 10:29:59.946686 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.001606 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-config-data" (OuterVolumeSpecName: "config-data") pod "601f3615-5015-486a-bbb5-04c683da6990" (UID: "601f3615-5015-486a-bbb5-04c683da6990"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.006899 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "601f3615-5015-486a-bbb5-04c683da6990" (UID: "601f3615-5015-486a-bbb5-04c683da6990"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.050364 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.051009 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/601f3615-5015-486a-bbb5-04c683da6990-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.121494 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.150103 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-ngzsx" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.150973 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-ngzsx" event={"ID":"601f3615-5015-486a-bbb5-04c683da6990","Type":"ContainerDied","Data":"25ba33658addfd86f3973f0bbd3f7d31da0188bcd4762d3214f38a7e8af5297b"} Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.151009 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="25ba33658addfd86f3973f0bbd3f7d31da0188bcd4762d3214f38a7e8af5297b" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.160738 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"de71c6bf-377b-44e8-a5fb-e654b259404f","Type":"ContainerDied","Data":"ee45ab9f33fdda93f7c890750739536f8547b9a2cf6264542af4cb74ce30fa4b"} Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.160810 4632 scope.go:117] "RemoveContainer" containerID="4fa9439436746dc39d49be92d41774ffb73ecdcce50f69453fca69442efcc0cf" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.161111 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.170430 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7"] Mar 13 10:30:00 crc kubenswrapper[4632]: E0313 10:30:00.171733 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de71c6bf-377b-44e8-a5fb-e654b259404f" containerName="nova-api-log" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.171791 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="de71c6bf-377b-44e8-a5fb-e654b259404f" containerName="nova-api-log" Mar 13 10:30:00 crc kubenswrapper[4632]: E0313 10:30:00.171861 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de71c6bf-377b-44e8-a5fb-e654b259404f" containerName="nova-api-api" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.171874 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="de71c6bf-377b-44e8-a5fb-e654b259404f" containerName="nova-api-api" Mar 13 10:30:00 crc kubenswrapper[4632]: E0313 10:30:00.171931 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="601f3615-5015-486a-bbb5-04c683da6990" containerName="nova-manage" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.171956 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="601f3615-5015-486a-bbb5-04c683da6990" containerName="nova-manage" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.172383 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="de71c6bf-377b-44e8-a5fb-e654b259404f" containerName="nova-api-log" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.172443 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="de71c6bf-377b-44e8-a5fb-e654b259404f" containerName="nova-api-api" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.172466 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="601f3615-5015-486a-bbb5-04c683da6990" containerName="nova-manage" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.180680 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.190194 4632 scope.go:117] "RemoveContainer" containerID="66957b9fbbc860bd8b0b4ba61ac2afb5edc2532051b2d051493e658595c97c89" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.190609 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.198660 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556630-kxrkn"] Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.198766 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.199995 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556630-kxrkn" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.203241 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.208718 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.213953 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.241906 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556630-kxrkn"] Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.255648 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de71c6bf-377b-44e8-a5fb-e654b259404f-logs\") pod \"de71c6bf-377b-44e8-a5fb-e654b259404f\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.255715 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de71c6bf-377b-44e8-a5fb-e654b259404f-config-data\") pod \"de71c6bf-377b-44e8-a5fb-e654b259404f\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.255870 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjszk\" (UniqueName: \"kubernetes.io/projected/de71c6bf-377b-44e8-a5fb-e654b259404f-kube-api-access-cjszk\") pod \"de71c6bf-377b-44e8-a5fb-e654b259404f\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.255924 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de71c6bf-377b-44e8-a5fb-e654b259404f-combined-ca-bundle\") pod \"de71c6bf-377b-44e8-a5fb-e654b259404f\" (UID: \"de71c6bf-377b-44e8-a5fb-e654b259404f\") " Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.260457 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de71c6bf-377b-44e8-a5fb-e654b259404f-logs" (OuterVolumeSpecName: "logs") pod "de71c6bf-377b-44e8-a5fb-e654b259404f" (UID: "de71c6bf-377b-44e8-a5fb-e654b259404f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.285292 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de71c6bf-377b-44e8-a5fb-e654b259404f-kube-api-access-cjszk" (OuterVolumeSpecName: "kube-api-access-cjszk") pod "de71c6bf-377b-44e8-a5fb-e654b259404f" (UID: "de71c6bf-377b-44e8-a5fb-e654b259404f"). InnerVolumeSpecName "kube-api-access-cjszk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.309207 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7"] Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.352756 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de71c6bf-377b-44e8-a5fb-e654b259404f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de71c6bf-377b-44e8-a5fb-e654b259404f" (UID: "de71c6bf-377b-44e8-a5fb-e654b259404f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.358628 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e912f5a7-eb85-4d19-9703-6cd7ff46c810-secret-volume\") pod \"collect-profiles-29556630-kpbz7\" (UID: \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.358703 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e912f5a7-eb85-4d19-9703-6cd7ff46c810-config-volume\") pod \"collect-profiles-29556630-kpbz7\" (UID: \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.358930 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4884\" (UniqueName: \"kubernetes.io/projected/b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e-kube-api-access-m4884\") pod \"auto-csr-approver-29556630-kxrkn\" (UID: \"b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e\") " pod="openshift-infra/auto-csr-approver-29556630-kxrkn" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.359008 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhzz5\" (UniqueName: \"kubernetes.io/projected/e912f5a7-eb85-4d19-9703-6cd7ff46c810-kube-api-access-bhzz5\") pod \"collect-profiles-29556630-kpbz7\" (UID: \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.359140 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/de71c6bf-377b-44e8-a5fb-e654b259404f-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.359157 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjszk\" (UniqueName: \"kubernetes.io/projected/de71c6bf-377b-44e8-a5fb-e654b259404f-kube-api-access-cjszk\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.359171 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de71c6bf-377b-44e8-a5fb-e654b259404f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.367065 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de71c6bf-377b-44e8-a5fb-e654b259404f-config-data" (OuterVolumeSpecName: "config-data") pod "de71c6bf-377b-44e8-a5fb-e654b259404f" (UID: "de71c6bf-377b-44e8-a5fb-e654b259404f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.460636 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4884\" (UniqueName: \"kubernetes.io/projected/b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e-kube-api-access-m4884\") pod \"auto-csr-approver-29556630-kxrkn\" (UID: \"b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e\") " pod="openshift-infra/auto-csr-approver-29556630-kxrkn" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.460699 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhzz5\" (UniqueName: \"kubernetes.io/projected/e912f5a7-eb85-4d19-9703-6cd7ff46c810-kube-api-access-bhzz5\") pod \"collect-profiles-29556630-kpbz7\" (UID: \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.460759 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e912f5a7-eb85-4d19-9703-6cd7ff46c810-secret-volume\") pod \"collect-profiles-29556630-kpbz7\" (UID: \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.460782 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e912f5a7-eb85-4d19-9703-6cd7ff46c810-config-volume\") pod \"collect-profiles-29556630-kpbz7\" (UID: \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.460882 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de71c6bf-377b-44e8-a5fb-e654b259404f-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.461752 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e912f5a7-eb85-4d19-9703-6cd7ff46c810-config-volume\") pod \"collect-profiles-29556630-kpbz7\" (UID: \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.477281 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.477515 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4d0c4f9f-780f-42d8-9eee-cb2201034218" containerName="nova-scheduler-scheduler" containerID="cri-o://9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58" gracePeriod=30 Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.517898 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.519198 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" containerName="nova-metadata-log" containerID="cri-o://519f8117c2e26f4a0b3f5e7e157a107d38c388ac29acd5faf2f4b4ecf121e55f" gracePeriod=30 Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.519783 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" containerName="nova-metadata-metadata" containerID="cri-o://a8f2d5b445eff9d0451f4c86782cfcac63ef30c3164366e7abf62cc09495ddd8" gracePeriod=30 Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.530230 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4884\" (UniqueName: \"kubernetes.io/projected/b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e-kube-api-access-m4884\") pod \"auto-csr-approver-29556630-kxrkn\" (UID: \"b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e\") " pod="openshift-infra/auto-csr-approver-29556630-kxrkn" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.540421 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhzz5\" (UniqueName: \"kubernetes.io/projected/e912f5a7-eb85-4d19-9703-6cd7ff46c810-kube-api-access-bhzz5\") pod \"collect-profiles-29556630-kpbz7\" (UID: \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.543048 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e912f5a7-eb85-4d19-9703-6cd7ff46c810-secret-volume\") pod \"collect-profiles-29556630-kpbz7\" (UID: \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.557372 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.579732 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556630-kxrkn" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.593511 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.627828 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.671140 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.676508 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.684252 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.684710 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.685435 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.702536 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.773862 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-config-data\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.773925 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr6fw\" (UniqueName: \"kubernetes.io/projected/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-kube-api-access-gr6fw\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.773969 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.774013 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-logs\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.774032 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-public-tls-certs\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.774105 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.877079 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.877252 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-config-data\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.877288 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr6fw\" (UniqueName: \"kubernetes.io/projected/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-kube-api-access-gr6fw\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.877308 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.877348 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-logs\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.877365 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-public-tls-certs\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.881707 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-logs\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.889902 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-public-tls-certs\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.895801 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.896677 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-config-data\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.897267 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-internal-tls-certs\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:00 crc kubenswrapper[4632]: I0313 10:30:00.926859 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr6fw\" (UniqueName: \"kubernetes.io/projected/3ef77ea1-fee2-432d-9aba-c0acfedb4e69-kube-api-access-gr6fw\") pod \"nova-api-0\" (UID: \"3ef77ea1-fee2-432d-9aba-c0acfedb4e69\") " pod="openstack/nova-api-0" Mar 13 10:30:01 crc kubenswrapper[4632]: I0313 10:30:01.040698 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 10:30:01 crc kubenswrapper[4632]: I0313 10:30:01.209164 4632 generic.go:334] "Generic (PLEG): container finished" podID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" containerID="519f8117c2e26f4a0b3f5e7e157a107d38c388ac29acd5faf2f4b4ecf121e55f" exitCode=143 Mar 13 10:30:01 crc kubenswrapper[4632]: I0313 10:30:01.209213 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82ae3a46-0133-43f5-942d-0b9a5b4d59f4","Type":"ContainerDied","Data":"519f8117c2e26f4a0b3f5e7e157a107d38c388ac29acd5faf2f4b4ecf121e55f"} Mar 13 10:30:01 crc kubenswrapper[4632]: I0313 10:30:01.470531 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7"] Mar 13 10:30:01 crc kubenswrapper[4632]: W0313 10:30:01.654321 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0ccb00a_40ce_4b3d_86e8_8f87354c1e8e.slice/crio-4d8d82c51a3c93ea917ffae6d282471f4227fa3a597385ba7ee78758a779abd9 WatchSource:0}: Error finding container 4d8d82c51a3c93ea917ffae6d282471f4227fa3a597385ba7ee78758a779abd9: Status 404 returned error can't find the container with id 4d8d82c51a3c93ea917ffae6d282471f4227fa3a597385ba7ee78758a779abd9 Mar 13 10:30:01 crc kubenswrapper[4632]: I0313 10:30:01.666877 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556630-kxrkn"] Mar 13 10:30:01 crc kubenswrapper[4632]: I0313 10:30:01.838633 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.087962 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de71c6bf-377b-44e8-a5fb-e654b259404f" path="/var/lib/kubelet/pods/de71c6bf-377b-44e8-a5fb-e654b259404f/volumes" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.150280 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: E0313 10:30:02.186442 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:30:02 crc kubenswrapper[4632]: E0313 10:30:02.191560 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:30:02 crc kubenswrapper[4632]: E0313 10:30:02.198837 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 10:30:02 crc kubenswrapper[4632]: E0313 10:30:02.198929 4632 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4d0c4f9f-780f-42d8-9eee-cb2201034218" containerName="nova-scheduler-scheduler" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.218506 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556630-kxrkn" event={"ID":"b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e","Type":"ContainerStarted","Data":"4d8d82c51a3c93ea917ffae6d282471f4227fa3a597385ba7ee78758a779abd9"} Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.220378 4632 generic.go:334] "Generic (PLEG): container finished" podID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerID="e19df6f77d1da54d66347bac1ac445e20a6a2fb793251d7691739401089385b1" exitCode=0 Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.220417 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff318cc9-cbe7-4357-971a-26c26e8bd269","Type":"ContainerDied","Data":"e19df6f77d1da54d66347bac1ac445e20a6a2fb793251d7691739401089385b1"} Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.220434 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ff318cc9-cbe7-4357-971a-26c26e8bd269","Type":"ContainerDied","Data":"2f6d41c40b5de2ff95100617c3719cca8da01adbe4c5436cbe7b0e955e7ff656"} Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.220450 4632 scope.go:117] "RemoveContainer" containerID="e81d93539be5ad8110d788d8abb89dd69059700a39392e03a27cde23af0ce544" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.220584 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.226744 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff318cc9-cbe7-4357-971a-26c26e8bd269-run-httpd\") pod \"ff318cc9-cbe7-4357-971a-26c26e8bd269\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.226884 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-config-data\") pod \"ff318cc9-cbe7-4357-971a-26c26e8bd269\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.226917 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff318cc9-cbe7-4357-971a-26c26e8bd269-log-httpd\") pod \"ff318cc9-cbe7-4357-971a-26c26e8bd269\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.227012 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-scripts\") pod \"ff318cc9-cbe7-4357-971a-26c26e8bd269\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.227043 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-sg-core-conf-yaml\") pod \"ff318cc9-cbe7-4357-971a-26c26e8bd269\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.227071 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-combined-ca-bundle\") pod \"ff318cc9-cbe7-4357-971a-26c26e8bd269\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.227140 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2qm5d\" (UniqueName: \"kubernetes.io/projected/ff318cc9-cbe7-4357-971a-26c26e8bd269-kube-api-access-2qm5d\") pod \"ff318cc9-cbe7-4357-971a-26c26e8bd269\" (UID: \"ff318cc9-cbe7-4357-971a-26c26e8bd269\") " Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.227914 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff318cc9-cbe7-4357-971a-26c26e8bd269-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ff318cc9-cbe7-4357-971a-26c26e8bd269" (UID: "ff318cc9-cbe7-4357-971a-26c26e8bd269"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.228077 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff318cc9-cbe7-4357-971a-26c26e8bd269-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ff318cc9-cbe7-4357-971a-26c26e8bd269" (UID: "ff318cc9-cbe7-4357-971a-26c26e8bd269"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.252015 4632 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff318cc9-cbe7-4357-971a-26c26e8bd269-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.253818 4632 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ff318cc9-cbe7-4357-971a-26c26e8bd269-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.254187 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" event={"ID":"e912f5a7-eb85-4d19-9703-6cd7ff46c810","Type":"ContainerStarted","Data":"020084ff22e9c174abe1865969844a8ece77dd4c848ac5f03af6af51bccf8643"} Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.254310 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" event={"ID":"e912f5a7-eb85-4d19-9703-6cd7ff46c810","Type":"ContainerStarted","Data":"7a016243d1468faa2397822528bac518deaada14c2e02f5ebb8a0907294cf98e"} Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.261351 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3ef77ea1-fee2-432d-9aba-c0acfedb4e69","Type":"ContainerStarted","Data":"da0b238c4fcc59e32638ff3d13934331efeb0e2693bb059a1f899fdb9cf426cf"} Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.267016 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-scripts" (OuterVolumeSpecName: "scripts") pod "ff318cc9-cbe7-4357-971a-26c26e8bd269" (UID: "ff318cc9-cbe7-4357-971a-26c26e8bd269"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.271280 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff318cc9-cbe7-4357-971a-26c26e8bd269-kube-api-access-2qm5d" (OuterVolumeSpecName: "kube-api-access-2qm5d") pod "ff318cc9-cbe7-4357-971a-26c26e8bd269" (UID: "ff318cc9-cbe7-4357-971a-26c26e8bd269"). InnerVolumeSpecName "kube-api-access-2qm5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.280381 4632 scope.go:117] "RemoveContainer" containerID="ca42f2855aed10e6d2409a495733d8e631d4d8011dee434ee91ebcbee2777ca4" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.293391 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" podStartSLOduration=2.293375857 podStartE2EDuration="2.293375857s" podCreationTimestamp="2026-03-13 10:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:30:02.291793618 +0000 UTC m=+1576.314323771" watchObservedRunningTime="2026-03-13 10:30:02.293375857 +0000 UTC m=+1576.315905990" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.344263 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ff318cc9-cbe7-4357-971a-26c26e8bd269" (UID: "ff318cc9-cbe7-4357-971a-26c26e8bd269"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.356365 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.356397 4632 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.356428 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2qm5d\" (UniqueName: \"kubernetes.io/projected/ff318cc9-cbe7-4357-971a-26c26e8bd269-kube-api-access-2qm5d\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.357676 4632 scope.go:117] "RemoveContainer" containerID="cd9daf2ae40dbbc090305eb1d55ec458cd7302b2ff730d20e1f7272ead3f43c4" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.414606 4632 scope.go:117] "RemoveContainer" containerID="e19df6f77d1da54d66347bac1ac445e20a6a2fb793251d7691739401089385b1" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.421296 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ff318cc9-cbe7-4357-971a-26c26e8bd269" (UID: "ff318cc9-cbe7-4357-971a-26c26e8bd269"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.452456 4632 scope.go:117] "RemoveContainer" containerID="e81d93539be5ad8110d788d8abb89dd69059700a39392e03a27cde23af0ce544" Mar 13 10:30:02 crc kubenswrapper[4632]: E0313 10:30:02.452974 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e81d93539be5ad8110d788d8abb89dd69059700a39392e03a27cde23af0ce544\": container with ID starting with e81d93539be5ad8110d788d8abb89dd69059700a39392e03a27cde23af0ce544 not found: ID does not exist" containerID="e81d93539be5ad8110d788d8abb89dd69059700a39392e03a27cde23af0ce544" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.453105 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e81d93539be5ad8110d788d8abb89dd69059700a39392e03a27cde23af0ce544"} err="failed to get container status \"e81d93539be5ad8110d788d8abb89dd69059700a39392e03a27cde23af0ce544\": rpc error: code = NotFound desc = could not find container \"e81d93539be5ad8110d788d8abb89dd69059700a39392e03a27cde23af0ce544\": container with ID starting with e81d93539be5ad8110d788d8abb89dd69059700a39392e03a27cde23af0ce544 not found: ID does not exist" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.453209 4632 scope.go:117] "RemoveContainer" containerID="ca42f2855aed10e6d2409a495733d8e631d4d8011dee434ee91ebcbee2777ca4" Mar 13 10:30:02 crc kubenswrapper[4632]: E0313 10:30:02.453501 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ca42f2855aed10e6d2409a495733d8e631d4d8011dee434ee91ebcbee2777ca4\": container with ID starting with ca42f2855aed10e6d2409a495733d8e631d4d8011dee434ee91ebcbee2777ca4 not found: ID does not exist" containerID="ca42f2855aed10e6d2409a495733d8e631d4d8011dee434ee91ebcbee2777ca4" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.453604 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ca42f2855aed10e6d2409a495733d8e631d4d8011dee434ee91ebcbee2777ca4"} err="failed to get container status \"ca42f2855aed10e6d2409a495733d8e631d4d8011dee434ee91ebcbee2777ca4\": rpc error: code = NotFound desc = could not find container \"ca42f2855aed10e6d2409a495733d8e631d4d8011dee434ee91ebcbee2777ca4\": container with ID starting with ca42f2855aed10e6d2409a495733d8e631d4d8011dee434ee91ebcbee2777ca4 not found: ID does not exist" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.453716 4632 scope.go:117] "RemoveContainer" containerID="cd9daf2ae40dbbc090305eb1d55ec458cd7302b2ff730d20e1f7272ead3f43c4" Mar 13 10:30:02 crc kubenswrapper[4632]: E0313 10:30:02.454774 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd9daf2ae40dbbc090305eb1d55ec458cd7302b2ff730d20e1f7272ead3f43c4\": container with ID starting with cd9daf2ae40dbbc090305eb1d55ec458cd7302b2ff730d20e1f7272ead3f43c4 not found: ID does not exist" containerID="cd9daf2ae40dbbc090305eb1d55ec458cd7302b2ff730d20e1f7272ead3f43c4" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.454878 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd9daf2ae40dbbc090305eb1d55ec458cd7302b2ff730d20e1f7272ead3f43c4"} err="failed to get container status \"cd9daf2ae40dbbc090305eb1d55ec458cd7302b2ff730d20e1f7272ead3f43c4\": rpc error: code = NotFound desc = could not find container \"cd9daf2ae40dbbc090305eb1d55ec458cd7302b2ff730d20e1f7272ead3f43c4\": container with ID starting with cd9daf2ae40dbbc090305eb1d55ec458cd7302b2ff730d20e1f7272ead3f43c4 not found: ID does not exist" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.455037 4632 scope.go:117] "RemoveContainer" containerID="e19df6f77d1da54d66347bac1ac445e20a6a2fb793251d7691739401089385b1" Mar 13 10:30:02 crc kubenswrapper[4632]: E0313 10:30:02.455899 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e19df6f77d1da54d66347bac1ac445e20a6a2fb793251d7691739401089385b1\": container with ID starting with e19df6f77d1da54d66347bac1ac445e20a6a2fb793251d7691739401089385b1 not found: ID does not exist" containerID="e19df6f77d1da54d66347bac1ac445e20a6a2fb793251d7691739401089385b1" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.455996 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e19df6f77d1da54d66347bac1ac445e20a6a2fb793251d7691739401089385b1"} err="failed to get container status \"e19df6f77d1da54d66347bac1ac445e20a6a2fb793251d7691739401089385b1\": rpc error: code = NotFound desc = could not find container \"e19df6f77d1da54d66347bac1ac445e20a6a2fb793251d7691739401089385b1\": container with ID starting with e19df6f77d1da54d66347bac1ac445e20a6a2fb793251d7691739401089385b1 not found: ID does not exist" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.462467 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.509747 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-config-data" (OuterVolumeSpecName: "config-data") pod "ff318cc9-cbe7-4357-971a-26c26e8bd269" (UID: "ff318cc9-cbe7-4357-971a-26c26e8bd269"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.542165 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.565903 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ff318cc9-cbe7-4357-971a-26c26e8bd269-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.566720 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.595032 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.624007 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:30:02 crc kubenswrapper[4632]: E0313 10:30:02.624436 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="ceilometer-central-agent" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.624453 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="ceilometer-central-agent" Mar 13 10:30:02 crc kubenswrapper[4632]: E0313 10:30:02.624481 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="proxy-httpd" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.624488 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="proxy-httpd" Mar 13 10:30:02 crc kubenswrapper[4632]: E0313 10:30:02.624499 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="sg-core" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.624505 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="sg-core" Mar 13 10:30:02 crc kubenswrapper[4632]: E0313 10:30:02.624522 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="ceilometer-notification-agent" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.624529 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="ceilometer-notification-agent" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.624687 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="ceilometer-central-agent" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.624713 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="proxy-httpd" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.624738 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="sg-core" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.624752 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" containerName="ceilometer-notification-agent" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.626920 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.631471 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.640667 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.641071 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.714011 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75788dd97c-r8qnr"] Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.714358 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" podUID="e0f17959-fde8-4cf1-b255-db5fc3325b70" containerName="dnsmasq-dns" containerID="cri-o://f957b291649cd64b5f0c12f7a4a8a32abd88e0067f00c5ae80a3e106aedde5a8" gracePeriod=10 Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.784022 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-scripts\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.784084 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-config-data\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.784223 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.784266 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvm2p\" (UniqueName: \"kubernetes.io/projected/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-kube-api-access-nvm2p\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.784298 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-run-httpd\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.784360 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-log-httpd\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.784384 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.890272 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.890365 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-run-httpd\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.890391 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvm2p\" (UniqueName: \"kubernetes.io/projected/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-kube-api-access-nvm2p\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.890432 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-log-httpd\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.890453 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.890529 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-scripts\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.890561 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-config-data\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.892538 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-log-httpd\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.892869 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-run-httpd\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.901124 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-scripts\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.901513 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.902830 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.927768 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvm2p\" (UniqueName: \"kubernetes.io/projected/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-kube-api-access-nvm2p\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.936416 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-config-data\") pod \"ceilometer-0\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " pod="openstack/ceilometer-0" Mar 13 10:30:02 crc kubenswrapper[4632]: I0313 10:30:02.967474 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.294044 4632 generic.go:334] "Generic (PLEG): container finished" podID="e912f5a7-eb85-4d19-9703-6cd7ff46c810" containerID="020084ff22e9c174abe1865969844a8ece77dd4c848ac5f03af6af51bccf8643" exitCode=0 Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.294220 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" event={"ID":"e912f5a7-eb85-4d19-9703-6cd7ff46c810","Type":"ContainerDied","Data":"020084ff22e9c174abe1865969844a8ece77dd4c848ac5f03af6af51bccf8643"} Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.307069 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3ef77ea1-fee2-432d-9aba-c0acfedb4e69","Type":"ContainerStarted","Data":"6593142dfefa299444df5fb08b472bc35abc684dfbe62b81f00b738f693d0298"} Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.307145 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"3ef77ea1-fee2-432d-9aba-c0acfedb4e69","Type":"ContainerStarted","Data":"369720f21c546cdcf03c1df03b4e7a408184d133e835b378f918587a8367d7e3"} Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.317983 4632 generic.go:334] "Generic (PLEG): container finished" podID="e0f17959-fde8-4cf1-b255-db5fc3325b70" containerID="f957b291649cd64b5f0c12f7a4a8a32abd88e0067f00c5ae80a3e106aedde5a8" exitCode=0 Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.318049 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" event={"ID":"e0f17959-fde8-4cf1-b255-db5fc3325b70","Type":"ContainerDied","Data":"f957b291649cd64b5f0c12f7a4a8a32abd88e0067f00c5ae80a3e106aedde5a8"} Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.359124 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.359105698 podStartE2EDuration="3.359105698s" podCreationTimestamp="2026-03-13 10:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:30:03.34720229 +0000 UTC m=+1577.369732423" watchObservedRunningTime="2026-03-13 10:30:03.359105698 +0000 UTC m=+1577.381635841" Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.718594 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.826544 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-dns-svc\") pod \"e0f17959-fde8-4cf1-b255-db5fc3325b70\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.826912 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-dns-swift-storage-0\") pod \"e0f17959-fde8-4cf1-b255-db5fc3325b70\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.827041 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-ovsdbserver-sb\") pod \"e0f17959-fde8-4cf1-b255-db5fc3325b70\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.827079 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc7jm\" (UniqueName: \"kubernetes.io/projected/e0f17959-fde8-4cf1-b255-db5fc3325b70-kube-api-access-pc7jm\") pod \"e0f17959-fde8-4cf1-b255-db5fc3325b70\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.827153 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-ovsdbserver-nb\") pod \"e0f17959-fde8-4cf1-b255-db5fc3325b70\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.827226 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-config\") pod \"e0f17959-fde8-4cf1-b255-db5fc3325b70\" (UID: \"e0f17959-fde8-4cf1-b255-db5fc3325b70\") " Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.871850 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0f17959-fde8-4cf1-b255-db5fc3325b70-kube-api-access-pc7jm" (OuterVolumeSpecName: "kube-api-access-pc7jm") pod "e0f17959-fde8-4cf1-b255-db5fc3325b70" (UID: "e0f17959-fde8-4cf1-b255-db5fc3325b70"). InnerVolumeSpecName "kube-api-access-pc7jm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.930797 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc7jm\" (UniqueName: \"kubernetes.io/projected/e0f17959-fde8-4cf1-b255-db5fc3325b70-kube-api-access-pc7jm\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.962209 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "e0f17959-fde8-4cf1-b255-db5fc3325b70" (UID: "e0f17959-fde8-4cf1-b255-db5fc3325b70"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:30:03 crc kubenswrapper[4632]: I0313 10:30:03.996417 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-config" (OuterVolumeSpecName: "config") pod "e0f17959-fde8-4cf1-b255-db5fc3325b70" (UID: "e0f17959-fde8-4cf1-b255-db5fc3325b70"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.005008 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e0f17959-fde8-4cf1-b255-db5fc3325b70" (UID: "e0f17959-fde8-4cf1-b255-db5fc3325b70"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.007505 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "e0f17959-fde8-4cf1-b255-db5fc3325b70" (UID: "e0f17959-fde8-4cf1-b255-db5fc3325b70"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.024546 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.034305 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.034349 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.034358 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.034367 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.075342 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "e0f17959-fde8-4cf1-b255-db5fc3325b70" (UID: "e0f17959-fde8-4cf1-b255-db5fc3325b70"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.105333 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff318cc9-cbe7-4357-971a-26c26e8bd269" path="/var/lib/kubelet/pods/ff318cc9-cbe7-4357-971a-26c26e8bd269/volumes" Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.136529 4632 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/e0f17959-fde8-4cf1-b255-db5fc3325b70-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.328545 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.328548 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75788dd97c-r8qnr" event={"ID":"e0f17959-fde8-4cf1-b255-db5fc3325b70","Type":"ContainerDied","Data":"a2b2fdcf6ef7efc2eca17a814eb5b4394c29b09fe6419666d04ee4759d7660a8"} Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.328664 4632 scope.go:117] "RemoveContainer" containerID="f957b291649cd64b5f0c12f7a4a8a32abd88e0067f00c5ae80a3e106aedde5a8" Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.330485 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc34b88a-a0cc-4ef1-8267-30f73d9712e7","Type":"ContainerStarted","Data":"43e2d3c1f3ad2fd1a1419876a0a3f1ee25556cf01d81619d7588168a228ea654"} Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.371852 4632 scope.go:117] "RemoveContainer" containerID="c6848744dc1fd449bb0df7b7ca2c04941331806f97abf20c11372e120fb30d31" Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.378203 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75788dd97c-r8qnr"] Mar 13 10:30:04 crc kubenswrapper[4632]: I0313 10:30:04.393539 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75788dd97c-r8qnr"] Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.026756 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.174991 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e912f5a7-eb85-4d19-9703-6cd7ff46c810-secret-volume\") pod \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\" (UID: \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\") " Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.175182 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e912f5a7-eb85-4d19-9703-6cd7ff46c810-config-volume\") pod \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\" (UID: \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\") " Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.175519 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bhzz5\" (UniqueName: \"kubernetes.io/projected/e912f5a7-eb85-4d19-9703-6cd7ff46c810-kube-api-access-bhzz5\") pod \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\" (UID: \"e912f5a7-eb85-4d19-9703-6cd7ff46c810\") " Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.175852 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e912f5a7-eb85-4d19-9703-6cd7ff46c810-config-volume" (OuterVolumeSpecName: "config-volume") pod "e912f5a7-eb85-4d19-9703-6cd7ff46c810" (UID: "e912f5a7-eb85-4d19-9703-6cd7ff46c810"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.176517 4632 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e912f5a7-eb85-4d19-9703-6cd7ff46c810-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.185553 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e912f5a7-eb85-4d19-9703-6cd7ff46c810-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e912f5a7-eb85-4d19-9703-6cd7ff46c810" (UID: "e912f5a7-eb85-4d19-9703-6cd7ff46c810"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.199777 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e912f5a7-eb85-4d19-9703-6cd7ff46c810-kube-api-access-bhzz5" (OuterVolumeSpecName: "kube-api-access-bhzz5") pod "e912f5a7-eb85-4d19-9703-6cd7ff46c810" (UID: "e912f5a7-eb85-4d19-9703-6cd7ff46c810"). InnerVolumeSpecName "kube-api-access-bhzz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.281160 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.282273 4632 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e912f5a7-eb85-4d19-9703-6cd7ff46c810-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.282315 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bhzz5\" (UniqueName: \"kubernetes.io/projected/e912f5a7-eb85-4d19-9703-6cd7ff46c810-kube-api-access-bhzz5\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.372570 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc34b88a-a0cc-4ef1-8267-30f73d9712e7","Type":"ContainerStarted","Data":"38c2fc76cd7277bb1d4be79e157389e249663d48647565e9e7553c7283f85329"} Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.378355 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" event={"ID":"e912f5a7-eb85-4d19-9703-6cd7ff46c810","Type":"ContainerDied","Data":"7a016243d1468faa2397822528bac518deaada14c2e02f5ebb8a0907294cf98e"} Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.378403 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7a016243d1468faa2397822528bac518deaada14c2e02f5ebb8a0907294cf98e" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.378475 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.384703 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-combined-ca-bundle\") pod \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.384804 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-nova-metadata-tls-certs\") pod \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.384927 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-config-data\") pod \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.385058 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rdwzm\" (UniqueName: \"kubernetes.io/projected/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-kube-api-access-rdwzm\") pod \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.385100 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-logs\") pod \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\" (UID: \"82ae3a46-0133-43f5-942d-0b9a5b4d59f4\") " Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.386910 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-logs" (OuterVolumeSpecName: "logs") pod "82ae3a46-0133-43f5-942d-0b9a5b4d59f4" (UID: "82ae3a46-0133-43f5-942d-0b9a5b4d59f4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.388019 4632 generic.go:334] "Generic (PLEG): container finished" podID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" containerID="a8f2d5b445eff9d0451f4c86782cfcac63ef30c3164366e7abf62cc09495ddd8" exitCode=0 Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.388071 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82ae3a46-0133-43f5-942d-0b9a5b4d59f4","Type":"ContainerDied","Data":"a8f2d5b445eff9d0451f4c86782cfcac63ef30c3164366e7abf62cc09495ddd8"} Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.388101 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82ae3a46-0133-43f5-942d-0b9a5b4d59f4","Type":"ContainerDied","Data":"d71e5927086b73530c1dba7fd1700212b2fab56fed475076dc30c04ba970bcf7"} Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.388123 4632 scope.go:117] "RemoveContainer" containerID="a8f2d5b445eff9d0451f4c86782cfcac63ef30c3164366e7abf62cc09495ddd8" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.388327 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.397227 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-kube-api-access-rdwzm" (OuterVolumeSpecName: "kube-api-access-rdwzm") pod "82ae3a46-0133-43f5-942d-0b9a5b4d59f4" (UID: "82ae3a46-0133-43f5-942d-0b9a5b4d59f4"). InnerVolumeSpecName "kube-api-access-rdwzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.457609 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-config-data" (OuterVolumeSpecName: "config-data") pod "82ae3a46-0133-43f5-942d-0b9a5b4d59f4" (UID: "82ae3a46-0133-43f5-942d-0b9a5b4d59f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.464187 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82ae3a46-0133-43f5-942d-0b9a5b4d59f4" (UID: "82ae3a46-0133-43f5-942d-0b9a5b4d59f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.480572 4632 scope.go:117] "RemoveContainer" containerID="519f8117c2e26f4a0b3f5e7e157a107d38c388ac29acd5faf2f4b4ecf121e55f" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.490385 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rdwzm\" (UniqueName: \"kubernetes.io/projected/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-kube-api-access-rdwzm\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.490444 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.490454 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.490465 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.539155 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "82ae3a46-0133-43f5-942d-0b9a5b4d59f4" (UID: "82ae3a46-0133-43f5-942d-0b9a5b4d59f4"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.558911 4632 scope.go:117] "RemoveContainer" containerID="a8f2d5b445eff9d0451f4c86782cfcac63ef30c3164366e7abf62cc09495ddd8" Mar 13 10:30:05 crc kubenswrapper[4632]: E0313 10:30:05.562451 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8f2d5b445eff9d0451f4c86782cfcac63ef30c3164366e7abf62cc09495ddd8\": container with ID starting with a8f2d5b445eff9d0451f4c86782cfcac63ef30c3164366e7abf62cc09495ddd8 not found: ID does not exist" containerID="a8f2d5b445eff9d0451f4c86782cfcac63ef30c3164366e7abf62cc09495ddd8" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.562499 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8f2d5b445eff9d0451f4c86782cfcac63ef30c3164366e7abf62cc09495ddd8"} err="failed to get container status \"a8f2d5b445eff9d0451f4c86782cfcac63ef30c3164366e7abf62cc09495ddd8\": rpc error: code = NotFound desc = could not find container \"a8f2d5b445eff9d0451f4c86782cfcac63ef30c3164366e7abf62cc09495ddd8\": container with ID starting with a8f2d5b445eff9d0451f4c86782cfcac63ef30c3164366e7abf62cc09495ddd8 not found: ID does not exist" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.562522 4632 scope.go:117] "RemoveContainer" containerID="519f8117c2e26f4a0b3f5e7e157a107d38c388ac29acd5faf2f4b4ecf121e55f" Mar 13 10:30:05 crc kubenswrapper[4632]: E0313 10:30:05.566083 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"519f8117c2e26f4a0b3f5e7e157a107d38c388ac29acd5faf2f4b4ecf121e55f\": container with ID starting with 519f8117c2e26f4a0b3f5e7e157a107d38c388ac29acd5faf2f4b4ecf121e55f not found: ID does not exist" containerID="519f8117c2e26f4a0b3f5e7e157a107d38c388ac29acd5faf2f4b4ecf121e55f" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.566121 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"519f8117c2e26f4a0b3f5e7e157a107d38c388ac29acd5faf2f4b4ecf121e55f"} err="failed to get container status \"519f8117c2e26f4a0b3f5e7e157a107d38c388ac29acd5faf2f4b4ecf121e55f\": rpc error: code = NotFound desc = could not find container \"519f8117c2e26f4a0b3f5e7e157a107d38c388ac29acd5faf2f4b4ecf121e55f\": container with ID starting with 519f8117c2e26f4a0b3f5e7e157a107d38c388ac29acd5faf2f4b4ecf121e55f not found: ID does not exist" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.592413 4632 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82ae3a46-0133-43f5-942d-0b9a5b4d59f4-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.728837 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.745423 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.754229 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:30:05 crc kubenswrapper[4632]: E0313 10:30:05.754618 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f17959-fde8-4cf1-b255-db5fc3325b70" containerName="dnsmasq-dns" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.754634 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f17959-fde8-4cf1-b255-db5fc3325b70" containerName="dnsmasq-dns" Mar 13 10:30:05 crc kubenswrapper[4632]: E0313 10:30:05.754645 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" containerName="nova-metadata-metadata" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.754652 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" containerName="nova-metadata-metadata" Mar 13 10:30:05 crc kubenswrapper[4632]: E0313 10:30:05.754666 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f17959-fde8-4cf1-b255-db5fc3325b70" containerName="init" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.754673 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f17959-fde8-4cf1-b255-db5fc3325b70" containerName="init" Mar 13 10:30:05 crc kubenswrapper[4632]: E0313 10:30:05.754696 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" containerName="nova-metadata-log" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.754702 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" containerName="nova-metadata-log" Mar 13 10:30:05 crc kubenswrapper[4632]: E0313 10:30:05.754715 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e912f5a7-eb85-4d19-9703-6cd7ff46c810" containerName="collect-profiles" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.754721 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e912f5a7-eb85-4d19-9703-6cd7ff46c810" containerName="collect-profiles" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.754898 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" containerName="nova-metadata-metadata" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.754912 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" containerName="nova-metadata-log" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.754931 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="e912f5a7-eb85-4d19-9703-6cd7ff46c810" containerName="collect-profiles" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.754960 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f17959-fde8-4cf1-b255-db5fc3325b70" containerName="dnsmasq-dns" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.755963 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.757847 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.762649 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.775985 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.897746 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b75084d0-782c-4f7e-8cc0-62ac424eec6f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.898093 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b75084d0-782c-4f7e-8cc0-62ac424eec6f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.898124 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzkds\" (UniqueName: \"kubernetes.io/projected/b75084d0-782c-4f7e-8cc0-62ac424eec6f-kube-api-access-pzkds\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.898177 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b75084d0-782c-4f7e-8cc0-62ac424eec6f-logs\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:05 crc kubenswrapper[4632]: I0313 10:30:05.898227 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b75084d0-782c-4f7e-8cc0-62ac424eec6f-config-data\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.000196 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b75084d0-782c-4f7e-8cc0-62ac424eec6f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.000310 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b75084d0-782c-4f7e-8cc0-62ac424eec6f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.000346 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzkds\" (UniqueName: \"kubernetes.io/projected/b75084d0-782c-4f7e-8cc0-62ac424eec6f-kube-api-access-pzkds\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.000415 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b75084d0-782c-4f7e-8cc0-62ac424eec6f-logs\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.000486 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b75084d0-782c-4f7e-8cc0-62ac424eec6f-config-data\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.001062 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b75084d0-782c-4f7e-8cc0-62ac424eec6f-logs\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.022031 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b75084d0-782c-4f7e-8cc0-62ac424eec6f-config-data\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.029601 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b75084d0-782c-4f7e-8cc0-62ac424eec6f-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.038734 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b75084d0-782c-4f7e-8cc0-62ac424eec6f-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.047074 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzkds\" (UniqueName: \"kubernetes.io/projected/b75084d0-782c-4f7e-8cc0-62ac424eec6f-kube-api-access-pzkds\") pod \"nova-metadata-0\" (UID: \"b75084d0-782c-4f7e-8cc0-62ac424eec6f\") " pod="openstack/nova-metadata-0" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.090336 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.124508 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" path="/var/lib/kubelet/pods/82ae3a46-0133-43f5-942d-0b9a5b4d59f4/volumes" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.131755 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0f17959-fde8-4cf1-b255-db5fc3325b70" path="/var/lib/kubelet/pods/e0f17959-fde8-4cf1-b255-db5fc3325b70/volumes" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.296730 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.423025 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc34b88a-a0cc-4ef1-8267-30f73d9712e7","Type":"ContainerStarted","Data":"8fc781f30b1bf552b69e65bd6fbe07dd0ebc73a78af0ab302a58fadf4aa1b637"} Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.443251 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556630-kxrkn" event={"ID":"b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e","Type":"ContainerStarted","Data":"dd1843e80da062d2b847859e60f624eed6f5f23e9e94519edc79cfc924e74d60"} Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.457122 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d0c4f9f-780f-42d8-9eee-cb2201034218-combined-ca-bundle\") pod \"4d0c4f9f-780f-42d8-9eee-cb2201034218\" (UID: \"4d0c4f9f-780f-42d8-9eee-cb2201034218\") " Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.457197 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d0c4f9f-780f-42d8-9eee-cb2201034218-config-data\") pod \"4d0c4f9f-780f-42d8-9eee-cb2201034218\" (UID: \"4d0c4f9f-780f-42d8-9eee-cb2201034218\") " Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.457228 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dmpm\" (UniqueName: \"kubernetes.io/projected/4d0c4f9f-780f-42d8-9eee-cb2201034218-kube-api-access-2dmpm\") pod \"4d0c4f9f-780f-42d8-9eee-cb2201034218\" (UID: \"4d0c4f9f-780f-42d8-9eee-cb2201034218\") " Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.472565 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d0c4f9f-780f-42d8-9eee-cb2201034218-kube-api-access-2dmpm" (OuterVolumeSpecName: "kube-api-access-2dmpm") pod "4d0c4f9f-780f-42d8-9eee-cb2201034218" (UID: "4d0c4f9f-780f-42d8-9eee-cb2201034218"). InnerVolumeSpecName "kube-api-access-2dmpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.500960 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556630-kxrkn" podStartSLOduration=4.016257042 podStartE2EDuration="6.500920796s" podCreationTimestamp="2026-03-13 10:30:00 +0000 UTC" firstStartedPulling="2026-03-13 10:30:01.661139541 +0000 UTC m=+1575.683669674" lastFinishedPulling="2026-03-13 10:30:04.145803295 +0000 UTC m=+1578.168333428" observedRunningTime="2026-03-13 10:30:06.466824554 +0000 UTC m=+1580.489354687" watchObservedRunningTime="2026-03-13 10:30:06.500920796 +0000 UTC m=+1580.523450929" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.532237 4632 generic.go:334] "Generic (PLEG): container finished" podID="4d0c4f9f-780f-42d8-9eee-cb2201034218" containerID="9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58" exitCode=0 Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.532503 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d0c4f9f-780f-42d8-9eee-cb2201034218","Type":"ContainerDied","Data":"9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58"} Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.532590 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4d0c4f9f-780f-42d8-9eee-cb2201034218","Type":"ContainerDied","Data":"ccfc717e8149e75cbe225885927d06c595e9efeff8370ba3176af49fbdc5eb3d"} Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.532659 4632 scope.go:117] "RemoveContainer" containerID="9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.532847 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.569067 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dmpm\" (UniqueName: \"kubernetes.io/projected/4d0c4f9f-780f-42d8-9eee-cb2201034218-kube-api-access-2dmpm\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.573116 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d0c4f9f-780f-42d8-9eee-cb2201034218-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4d0c4f9f-780f-42d8-9eee-cb2201034218" (UID: "4d0c4f9f-780f-42d8-9eee-cb2201034218"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.577920 4632 scope.go:117] "RemoveContainer" containerID="9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.588302 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d0c4f9f-780f-42d8-9eee-cb2201034218-config-data" (OuterVolumeSpecName: "config-data") pod "4d0c4f9f-780f-42d8-9eee-cb2201034218" (UID: "4d0c4f9f-780f-42d8-9eee-cb2201034218"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:06 crc kubenswrapper[4632]: E0313 10:30:06.589626 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58\": container with ID starting with 9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58 not found: ID does not exist" containerID="9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.589722 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58"} err="failed to get container status \"9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58\": rpc error: code = NotFound desc = could not find container \"9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58\": container with ID starting with 9ea4520ca12a1649f0ffe1aaf48fb3759b0ff4ddc87166ec50af120c6fe09b58 not found: ID does not exist" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.677423 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4d0c4f9f-780f-42d8-9eee-cb2201034218-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.677455 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4d0c4f9f-780f-42d8-9eee-cb2201034218-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:06 crc kubenswrapper[4632]: I0313 10:30:06.798854 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 10:30:06 crc kubenswrapper[4632]: W0313 10:30:06.803710 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb75084d0_782c_4f7e_8cc0_62ac424eec6f.slice/crio-f4c06c100d719e4d44cc6c6c52de3a749fc1ccff65408492392c702aaa713cc3 WatchSource:0}: Error finding container f4c06c100d719e4d44cc6c6c52de3a749fc1ccff65408492392c702aaa713cc3: Status 404 returned error can't find the container with id f4c06c100d719e4d44cc6c6c52de3a749fc1ccff65408492392c702aaa713cc3 Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.038397 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.085013 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.131806 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:30:07 crc kubenswrapper[4632]: E0313 10:30:07.136594 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d0c4f9f-780f-42d8-9eee-cb2201034218" containerName="nova-scheduler-scheduler" Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.136821 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d0c4f9f-780f-42d8-9eee-cb2201034218" containerName="nova-scheduler-scheduler" Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.138673 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d0c4f9f-780f-42d8-9eee-cb2201034218" containerName="nova-scheduler-scheduler" Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.140397 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.155486 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.229980 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.242761 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndpcn\" (UniqueName: \"kubernetes.io/projected/bd274a76-bf05-4f69-8d56-4844012a1fd1-kube-api-access-ndpcn\") pod \"nova-scheduler-0\" (UID: \"bd274a76-bf05-4f69-8d56-4844012a1fd1\") " pod="openstack/nova-scheduler-0" Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.242932 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd274a76-bf05-4f69-8d56-4844012a1fd1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bd274a76-bf05-4f69-8d56-4844012a1fd1\") " pod="openstack/nova-scheduler-0" Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.243093 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd274a76-bf05-4f69-8d56-4844012a1fd1-config-data\") pod \"nova-scheduler-0\" (UID: \"bd274a76-bf05-4f69-8d56-4844012a1fd1\") " pod="openstack/nova-scheduler-0" Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.347258 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd274a76-bf05-4f69-8d56-4844012a1fd1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bd274a76-bf05-4f69-8d56-4844012a1fd1\") " pod="openstack/nova-scheduler-0" Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.347376 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd274a76-bf05-4f69-8d56-4844012a1fd1-config-data\") pod \"nova-scheduler-0\" (UID: \"bd274a76-bf05-4f69-8d56-4844012a1fd1\") " pod="openstack/nova-scheduler-0" Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.347517 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndpcn\" (UniqueName: \"kubernetes.io/projected/bd274a76-bf05-4f69-8d56-4844012a1fd1-kube-api-access-ndpcn\") pod \"nova-scheduler-0\" (UID: \"bd274a76-bf05-4f69-8d56-4844012a1fd1\") " pod="openstack/nova-scheduler-0" Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.360198 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd274a76-bf05-4f69-8d56-4844012a1fd1-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"bd274a76-bf05-4f69-8d56-4844012a1fd1\") " pod="openstack/nova-scheduler-0" Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.361599 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd274a76-bf05-4f69-8d56-4844012a1fd1-config-data\") pod \"nova-scheduler-0\" (UID: \"bd274a76-bf05-4f69-8d56-4844012a1fd1\") " pod="openstack/nova-scheduler-0" Mar 13 10:30:07 crc kubenswrapper[4632]: E0313 10:30:07.383697 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4d0c4f9f_780f_42d8_9eee_cb2201034218.slice/crio-ccfc717e8149e75cbe225885927d06c595e9efeff8370ba3176af49fbdc5eb3d\": RecentStats: unable to find data in memory cache]" Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.387384 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndpcn\" (UniqueName: \"kubernetes.io/projected/bd274a76-bf05-4f69-8d56-4844012a1fd1-kube-api-access-ndpcn\") pod \"nova-scheduler-0\" (UID: \"bd274a76-bf05-4f69-8d56-4844012a1fd1\") " pod="openstack/nova-scheduler-0" Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.543720 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b75084d0-782c-4f7e-8cc0-62ac424eec6f","Type":"ContainerStarted","Data":"f4c06c100d719e4d44cc6c6c52de3a749fc1ccff65408492392c702aaa713cc3"} Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.546554 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc34b88a-a0cc-4ef1-8267-30f73d9712e7","Type":"ContainerStarted","Data":"2e3874df0599fa71baea0baa935a02f99ef35801e9c669049d94507f046a75da"} Mar 13 10:30:07 crc kubenswrapper[4632]: I0313 10:30:07.555164 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 10:30:08 crc kubenswrapper[4632]: I0313 10:30:08.065030 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d0c4f9f-780f-42d8-9eee-cb2201034218" path="/var/lib/kubelet/pods/4d0c4f9f-780f-42d8-9eee-cb2201034218/volumes" Mar 13 10:30:08 crc kubenswrapper[4632]: I0313 10:30:08.152995 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 10:30:08 crc kubenswrapper[4632]: W0313 10:30:08.163003 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd274a76_bf05_4f69_8d56_4844012a1fd1.slice/crio-39b042e495243a0fff2599056c78a54bf577bf285f31eae823e001fb8087bfca WatchSource:0}: Error finding container 39b042e495243a0fff2599056c78a54bf577bf285f31eae823e001fb8087bfca: Status 404 returned error can't find the container with id 39b042e495243a0fff2599056c78a54bf577bf285f31eae823e001fb8087bfca Mar 13 10:30:08 crc kubenswrapper[4632]: I0313 10:30:08.439847 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-n4z22" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="registry-server" probeResult="failure" output=< Mar 13 10:30:08 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:30:08 crc kubenswrapper[4632]: > Mar 13 10:30:08 crc kubenswrapper[4632]: I0313 10:30:08.562510 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bd274a76-bf05-4f69-8d56-4844012a1fd1","Type":"ContainerStarted","Data":"fedfdf64dd99493278404ee3a2fb9a63214432489ab8d05fc9862a8fc248f3f8"} Mar 13 10:30:08 crc kubenswrapper[4632]: I0313 10:30:08.562553 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"bd274a76-bf05-4f69-8d56-4844012a1fd1","Type":"ContainerStarted","Data":"39b042e495243a0fff2599056c78a54bf577bf285f31eae823e001fb8087bfca"} Mar 13 10:30:08 crc kubenswrapper[4632]: I0313 10:30:08.564967 4632 generic.go:334] "Generic (PLEG): container finished" podID="b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e" containerID="dd1843e80da062d2b847859e60f624eed6f5f23e9e94519edc79cfc924e74d60" exitCode=0 Mar 13 10:30:08 crc kubenswrapper[4632]: I0313 10:30:08.565033 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556630-kxrkn" event={"ID":"b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e","Type":"ContainerDied","Data":"dd1843e80da062d2b847859e60f624eed6f5f23e9e94519edc79cfc924e74d60"} Mar 13 10:30:08 crc kubenswrapper[4632]: I0313 10:30:08.566537 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b75084d0-782c-4f7e-8cc0-62ac424eec6f","Type":"ContainerStarted","Data":"b55f59548ce979d4ad220d89c90a338482b481e7c7175939010909518499f902"} Mar 13 10:30:08 crc kubenswrapper[4632]: I0313 10:30:08.566561 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b75084d0-782c-4f7e-8cc0-62ac424eec6f","Type":"ContainerStarted","Data":"a8ac6d9d06c83ccd83d308a76ad43bb6e8c9d0ff988e5799c63c26a39827efbc"} Mar 13 10:30:08 crc kubenswrapper[4632]: I0313 10:30:08.606411 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.6063949210000001 podStartE2EDuration="1.606394921s" podCreationTimestamp="2026-03-13 10:30:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:30:08.587919656 +0000 UTC m=+1582.610449789" watchObservedRunningTime="2026-03-13 10:30:08.606394921 +0000 UTC m=+1582.628925054" Mar 13 10:30:08 crc kubenswrapper[4632]: I0313 10:30:08.631601 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.631582258 podStartE2EDuration="3.631582258s" podCreationTimestamp="2026-03-13 10:30:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:30:08.626838824 +0000 UTC m=+1582.649368957" watchObservedRunningTime="2026-03-13 10:30:08.631582258 +0000 UTC m=+1582.654112391" Mar 13 10:30:09 crc kubenswrapper[4632]: I0313 10:30:09.578135 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc34b88a-a0cc-4ef1-8267-30f73d9712e7","Type":"ContainerStarted","Data":"aafa45b6b50dc62147a1600ded978371569f724c32856206cf7196aba295169c"} Mar 13 10:30:09 crc kubenswrapper[4632]: I0313 10:30:09.622840 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.894984538 podStartE2EDuration="7.622820694s" podCreationTimestamp="2026-03-13 10:30:02 +0000 UTC" firstStartedPulling="2026-03-13 10:30:04.082839488 +0000 UTC m=+1578.105369621" lastFinishedPulling="2026-03-13 10:30:08.810675644 +0000 UTC m=+1582.833205777" observedRunningTime="2026-03-13 10:30:09.609260648 +0000 UTC m=+1583.631790781" watchObservedRunningTime="2026-03-13 10:30:09.622820694 +0000 UTC m=+1583.645350827" Mar 13 10:30:10 crc kubenswrapper[4632]: I0313 10:30:10.095103 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556630-kxrkn" Mar 13 10:30:10 crc kubenswrapper[4632]: I0313 10:30:10.188718 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.219:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:10 crc kubenswrapper[4632]: I0313 10:30:10.189407 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="82ae3a46-0133-43f5-942d-0b9a5b4d59f4" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.219:8775/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:10 crc kubenswrapper[4632]: I0313 10:30:10.226498 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4884\" (UniqueName: \"kubernetes.io/projected/b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e-kube-api-access-m4884\") pod \"b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e\" (UID: \"b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e\") " Mar 13 10:30:10 crc kubenswrapper[4632]: I0313 10:30:10.243154 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e-kube-api-access-m4884" (OuterVolumeSpecName: "kube-api-access-m4884") pod "b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e" (UID: "b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e"). InnerVolumeSpecName "kube-api-access-m4884". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:30:10 crc kubenswrapper[4632]: I0313 10:30:10.328561 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4884\" (UniqueName: \"kubernetes.io/projected/b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e-kube-api-access-m4884\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:10 crc kubenswrapper[4632]: I0313 10:30:10.461408 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:30:10 crc kubenswrapper[4632]: I0313 10:30:10.461807 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:30:10 crc kubenswrapper[4632]: I0313 10:30:10.611686 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556630-kxrkn" event={"ID":"b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e","Type":"ContainerDied","Data":"4d8d82c51a3c93ea917ffae6d282471f4227fa3a597385ba7ee78758a779abd9"} Mar 13 10:30:10 crc kubenswrapper[4632]: I0313 10:30:10.614551 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d8d82c51a3c93ea917ffae6d282471f4227fa3a597385ba7ee78758a779abd9" Mar 13 10:30:10 crc kubenswrapper[4632]: I0313 10:30:10.614663 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 10:30:10 crc kubenswrapper[4632]: I0313 10:30:10.611733 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556630-kxrkn" Mar 13 10:30:10 crc kubenswrapper[4632]: I0313 10:30:10.692187 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556624-tnl4c"] Mar 13 10:30:10 crc kubenswrapper[4632]: I0313 10:30:10.700911 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556624-tnl4c"] Mar 13 10:30:11 crc kubenswrapper[4632]: I0313 10:30:11.041738 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 10:30:11 crc kubenswrapper[4632]: I0313 10:30:11.041798 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 10:30:11 crc kubenswrapper[4632]: I0313 10:30:11.091804 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 10:30:11 crc kubenswrapper[4632]: I0313 10:30:11.092716 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 10:30:12 crc kubenswrapper[4632]: I0313 10:30:12.054283 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3ef77ea1-fee2-432d-9aba-c0acfedb4e69" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.227:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4632]: I0313 10:30:12.054318 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="3ef77ea1-fee2-432d-9aba-c0acfedb4e69" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.227:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4632]: I0313 10:30:12.059760 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b2374d5-8d19-4837-8d91-79df0e65fc1f" path="/var/lib/kubelet/pods/5b2374d5-8d19-4837-8d91-79df0e65fc1f/volumes" Mar 13 10:30:12 crc kubenswrapper[4632]: I0313 10:30:12.555461 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 10:30:16 crc kubenswrapper[4632]: I0313 10:30:16.091393 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 10:30:16 crc kubenswrapper[4632]: I0313 10:30:16.091841 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 10:30:17 crc kubenswrapper[4632]: I0313 10:30:17.104217 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b75084d0-782c-4f7e-8cc0-62ac424eec6f" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.229:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4632]: I0313 10:30:17.104525 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b75084d0-782c-4f7e-8cc0-62ac424eec6f" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.229:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4632]: I0313 10:30:17.314891 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:30:17 crc kubenswrapper[4632]: I0313 10:30:17.394425 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:30:17 crc kubenswrapper[4632]: I0313 10:30:17.555632 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 10:30:17 crc kubenswrapper[4632]: I0313 10:30:17.586600 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 10:30:17 crc kubenswrapper[4632]: I0313 10:30:17.843320 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 10:30:18 crc kubenswrapper[4632]: I0313 10:30:18.178900 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n4z22"] Mar 13 10:30:18 crc kubenswrapper[4632]: I0313 10:30:18.819861 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-n4z22" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="registry-server" containerID="cri-o://d90ad363a11a672d65b86841626f5e66c7cb30326c75375f126c473ca51ca722" gracePeriod=2 Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.431910 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.547173 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0cabd29-ef3e-4808-8c92-3b032483789e-catalog-content\") pod \"d0cabd29-ef3e-4808-8c92-3b032483789e\" (UID: \"d0cabd29-ef3e-4808-8c92-3b032483789e\") " Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.547487 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpwh9\" (UniqueName: \"kubernetes.io/projected/d0cabd29-ef3e-4808-8c92-3b032483789e-kube-api-access-xpwh9\") pod \"d0cabd29-ef3e-4808-8c92-3b032483789e\" (UID: \"d0cabd29-ef3e-4808-8c92-3b032483789e\") " Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.547644 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0cabd29-ef3e-4808-8c92-3b032483789e-utilities\") pod \"d0cabd29-ef3e-4808-8c92-3b032483789e\" (UID: \"d0cabd29-ef3e-4808-8c92-3b032483789e\") " Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.550638 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0cabd29-ef3e-4808-8c92-3b032483789e-utilities" (OuterVolumeSpecName: "utilities") pod "d0cabd29-ef3e-4808-8c92-3b032483789e" (UID: "d0cabd29-ef3e-4808-8c92-3b032483789e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.559529 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0cabd29-ef3e-4808-8c92-3b032483789e-kube-api-access-xpwh9" (OuterVolumeSpecName: "kube-api-access-xpwh9") pod "d0cabd29-ef3e-4808-8c92-3b032483789e" (UID: "d0cabd29-ef3e-4808-8c92-3b032483789e"). InnerVolumeSpecName "kube-api-access-xpwh9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.649825 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpwh9\" (UniqueName: \"kubernetes.io/projected/d0cabd29-ef3e-4808-8c92-3b032483789e-kube-api-access-xpwh9\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.649862 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d0cabd29-ef3e-4808-8c92-3b032483789e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.742020 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d0cabd29-ef3e-4808-8c92-3b032483789e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d0cabd29-ef3e-4808-8c92-3b032483789e" (UID: "d0cabd29-ef3e-4808-8c92-3b032483789e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.752213 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d0cabd29-ef3e-4808-8c92-3b032483789e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.830485 4632 generic.go:334] "Generic (PLEG): container finished" podID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerID="d90ad363a11a672d65b86841626f5e66c7cb30326c75375f126c473ca51ca722" exitCode=0 Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.830537 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n4z22" event={"ID":"d0cabd29-ef3e-4808-8c92-3b032483789e","Type":"ContainerDied","Data":"d90ad363a11a672d65b86841626f5e66c7cb30326c75375f126c473ca51ca722"} Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.830565 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-n4z22" event={"ID":"d0cabd29-ef3e-4808-8c92-3b032483789e","Type":"ContainerDied","Data":"dd0fe42db5b99209dcd168810b0996ceb728a9055a395258ce5d2c5e8afe18b9"} Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.830581 4632 scope.go:117] "RemoveContainer" containerID="d90ad363a11a672d65b86841626f5e66c7cb30326c75375f126c473ca51ca722" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.830695 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-n4z22" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.867983 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-n4z22"] Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.876932 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-n4z22"] Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.881025 4632 scope.go:117] "RemoveContainer" containerID="71596437e6a2d325baac09e68c289bacf1de16594d0329b6a19a5f5d92099872" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.916747 4632 scope.go:117] "RemoveContainer" containerID="201e16e880cd2d90fce52e678a22a53fb8633269541a12fab257e648ab87ec2d" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.962841 4632 scope.go:117] "RemoveContainer" containerID="d90ad363a11a672d65b86841626f5e66c7cb30326c75375f126c473ca51ca722" Mar 13 10:30:19 crc kubenswrapper[4632]: E0313 10:30:19.963501 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d90ad363a11a672d65b86841626f5e66c7cb30326c75375f126c473ca51ca722\": container with ID starting with d90ad363a11a672d65b86841626f5e66c7cb30326c75375f126c473ca51ca722 not found: ID does not exist" containerID="d90ad363a11a672d65b86841626f5e66c7cb30326c75375f126c473ca51ca722" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.963557 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d90ad363a11a672d65b86841626f5e66c7cb30326c75375f126c473ca51ca722"} err="failed to get container status \"d90ad363a11a672d65b86841626f5e66c7cb30326c75375f126c473ca51ca722\": rpc error: code = NotFound desc = could not find container \"d90ad363a11a672d65b86841626f5e66c7cb30326c75375f126c473ca51ca722\": container with ID starting with d90ad363a11a672d65b86841626f5e66c7cb30326c75375f126c473ca51ca722 not found: ID does not exist" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.963585 4632 scope.go:117] "RemoveContainer" containerID="71596437e6a2d325baac09e68c289bacf1de16594d0329b6a19a5f5d92099872" Mar 13 10:30:19 crc kubenswrapper[4632]: E0313 10:30:19.964014 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71596437e6a2d325baac09e68c289bacf1de16594d0329b6a19a5f5d92099872\": container with ID starting with 71596437e6a2d325baac09e68c289bacf1de16594d0329b6a19a5f5d92099872 not found: ID does not exist" containerID="71596437e6a2d325baac09e68c289bacf1de16594d0329b6a19a5f5d92099872" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.964058 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71596437e6a2d325baac09e68c289bacf1de16594d0329b6a19a5f5d92099872"} err="failed to get container status \"71596437e6a2d325baac09e68c289bacf1de16594d0329b6a19a5f5d92099872\": rpc error: code = NotFound desc = could not find container \"71596437e6a2d325baac09e68c289bacf1de16594d0329b6a19a5f5d92099872\": container with ID starting with 71596437e6a2d325baac09e68c289bacf1de16594d0329b6a19a5f5d92099872 not found: ID does not exist" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.964094 4632 scope.go:117] "RemoveContainer" containerID="201e16e880cd2d90fce52e678a22a53fb8633269541a12fab257e648ab87ec2d" Mar 13 10:30:19 crc kubenswrapper[4632]: E0313 10:30:19.964443 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"201e16e880cd2d90fce52e678a22a53fb8633269541a12fab257e648ab87ec2d\": container with ID starting with 201e16e880cd2d90fce52e678a22a53fb8633269541a12fab257e648ab87ec2d not found: ID does not exist" containerID="201e16e880cd2d90fce52e678a22a53fb8633269541a12fab257e648ab87ec2d" Mar 13 10:30:19 crc kubenswrapper[4632]: I0313 10:30:19.964466 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"201e16e880cd2d90fce52e678a22a53fb8633269541a12fab257e648ab87ec2d"} err="failed to get container status \"201e16e880cd2d90fce52e678a22a53fb8633269541a12fab257e648ab87ec2d\": rpc error: code = NotFound desc = could not find container \"201e16e880cd2d90fce52e678a22a53fb8633269541a12fab257e648ab87ec2d\": container with ID starting with 201e16e880cd2d90fce52e678a22a53fb8633269541a12fab257e648ab87ec2d not found: ID does not exist" Mar 13 10:30:20 crc kubenswrapper[4632]: I0313 10:30:20.058136 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" path="/var/lib/kubelet/pods/d0cabd29-ef3e-4808-8c92-3b032483789e/volumes" Mar 13 10:30:21 crc kubenswrapper[4632]: I0313 10:30:21.049259 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 10:30:21 crc kubenswrapper[4632]: I0313 10:30:21.049644 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 10:30:21 crc kubenswrapper[4632]: I0313 10:30:21.050021 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 10:30:21 crc kubenswrapper[4632]: I0313 10:30:21.050063 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 10:30:21 crc kubenswrapper[4632]: I0313 10:30:21.058524 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 10:30:21 crc kubenswrapper[4632]: I0313 10:30:21.060843 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.522697 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.599762 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e4afb91-ce26-4325-89c9-2542da2ec48a-logs\") pod \"3e4afb91-ce26-4325-89c9-2542da2ec48a\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.599837 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-combined-ca-bundle\") pod \"3e4afb91-ce26-4325-89c9-2542da2ec48a\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.599865 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-horizon-secret-key\") pod \"3e4afb91-ce26-4325-89c9-2542da2ec48a\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.599931 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ntccx\" (UniqueName: \"kubernetes.io/projected/3e4afb91-ce26-4325-89c9-2542da2ec48a-kube-api-access-ntccx\") pod \"3e4afb91-ce26-4325-89c9-2542da2ec48a\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.600053 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e4afb91-ce26-4325-89c9-2542da2ec48a-scripts\") pod \"3e4afb91-ce26-4325-89c9-2542da2ec48a\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.600082 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e4afb91-ce26-4325-89c9-2542da2ec48a-config-data\") pod \"3e4afb91-ce26-4325-89c9-2542da2ec48a\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.600106 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-horizon-tls-certs\") pod \"3e4afb91-ce26-4325-89c9-2542da2ec48a\" (UID: \"3e4afb91-ce26-4325-89c9-2542da2ec48a\") " Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.603687 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e4afb91-ce26-4325-89c9-2542da2ec48a-logs" (OuterVolumeSpecName: "logs") pod "3e4afb91-ce26-4325-89c9-2542da2ec48a" (UID: "3e4afb91-ce26-4325-89c9-2542da2ec48a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.625203 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e4afb91-ce26-4325-89c9-2542da2ec48a-kube-api-access-ntccx" (OuterVolumeSpecName: "kube-api-access-ntccx") pod "3e4afb91-ce26-4325-89c9-2542da2ec48a" (UID: "3e4afb91-ce26-4325-89c9-2542da2ec48a"). InnerVolumeSpecName "kube-api-access-ntccx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.625298 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "3e4afb91-ce26-4325-89c9-2542da2ec48a" (UID: "3e4afb91-ce26-4325-89c9-2542da2ec48a"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.632341 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e4afb91-ce26-4325-89c9-2542da2ec48a-scripts" (OuterVolumeSpecName: "scripts") pod "3e4afb91-ce26-4325-89c9-2542da2ec48a" (UID: "3e4afb91-ce26-4325-89c9-2542da2ec48a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.639036 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e4afb91-ce26-4325-89c9-2542da2ec48a" (UID: "3e4afb91-ce26-4325-89c9-2542da2ec48a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.642253 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3e4afb91-ce26-4325-89c9-2542da2ec48a-config-data" (OuterVolumeSpecName: "config-data") pod "3e4afb91-ce26-4325-89c9-2542da2ec48a" (UID: "3e4afb91-ce26-4325-89c9-2542da2ec48a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.666551 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "3e4afb91-ce26-4325-89c9-2542da2ec48a" (UID: "3e4afb91-ce26-4325-89c9-2542da2ec48a"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.701553 4632 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3e4afb91-ce26-4325-89c9-2542da2ec48a-logs\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.701593 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.701604 4632 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.701644 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ntccx\" (UniqueName: \"kubernetes.io/projected/3e4afb91-ce26-4325-89c9-2542da2ec48a-kube-api-access-ntccx\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.701653 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3e4afb91-ce26-4325-89c9-2542da2ec48a-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.701661 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3e4afb91-ce26-4325-89c9-2542da2ec48a-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.701671 4632 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e4afb91-ce26-4325-89c9-2542da2ec48a-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.906708 4632 generic.go:334] "Generic (PLEG): container finished" podID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerID="d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35" exitCode=137 Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.906743 4632 generic.go:334] "Generic (PLEG): container finished" podID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerID="0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589" exitCode=137 Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.906764 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb5f7878-ng2k2" event={"ID":"3e4afb91-ce26-4325-89c9-2542da2ec48a","Type":"ContainerDied","Data":"d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35"} Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.906796 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb5f7878-ng2k2" event={"ID":"3e4afb91-ce26-4325-89c9-2542da2ec48a","Type":"ContainerDied","Data":"0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589"} Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.906795 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7bdb5f7878-ng2k2" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.906818 4632 scope.go:117] "RemoveContainer" containerID="d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35" Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.906807 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7bdb5f7878-ng2k2" event={"ID":"3e4afb91-ce26-4325-89c9-2542da2ec48a","Type":"ContainerDied","Data":"aaad122938f426786c8baabdc4555594b0ba0e55f0c39302b9bf84230f06cfd1"} Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.947880 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7bdb5f7878-ng2k2"] Mar 13 10:30:25 crc kubenswrapper[4632]: I0313 10:30:25.960022 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7bdb5f7878-ng2k2"] Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.068747 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" path="/var/lib/kubelet/pods/3e4afb91-ce26-4325-89c9-2542da2ec48a/volumes" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.105075 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.108170 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.115208 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.145366 4632 scope.go:117] "RemoveContainer" containerID="0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.381325 4632 scope.go:117] "RemoveContainer" containerID="0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.399545 4632 scope.go:117] "RemoveContainer" containerID="d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35" Mar 13 10:30:26 crc kubenswrapper[4632]: E0313 10:30:26.400116 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35\": container with ID starting with d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35 not found: ID does not exist" containerID="d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.400157 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35"} err="failed to get container status \"d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35\": rpc error: code = NotFound desc = could not find container \"d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35\": container with ID starting with d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35 not found: ID does not exist" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.400182 4632 scope.go:117] "RemoveContainer" containerID="0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30" Mar 13 10:30:26 crc kubenswrapper[4632]: E0313 10:30:26.400692 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30\": container with ID starting with 0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30 not found: ID does not exist" containerID="0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.400723 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30"} err="failed to get container status \"0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30\": rpc error: code = NotFound desc = could not find container \"0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30\": container with ID starting with 0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30 not found: ID does not exist" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.400742 4632 scope.go:117] "RemoveContainer" containerID="0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589" Mar 13 10:30:26 crc kubenswrapper[4632]: E0313 10:30:26.401066 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589\": container with ID starting with 0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589 not found: ID does not exist" containerID="0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.401089 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589"} err="failed to get container status \"0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589\": rpc error: code = NotFound desc = could not find container \"0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589\": container with ID starting with 0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589 not found: ID does not exist" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.401106 4632 scope.go:117] "RemoveContainer" containerID="d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.401333 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35"} err="failed to get container status \"d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35\": rpc error: code = NotFound desc = could not find container \"d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35\": container with ID starting with d57c6cccef4987c7003d38b1c8de63c00e37251e291f0b2d5a1b05218e53dd35 not found: ID does not exist" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.401348 4632 scope.go:117] "RemoveContainer" containerID="0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.401537 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30"} err="failed to get container status \"0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30\": rpc error: code = NotFound desc = could not find container \"0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30\": container with ID starting with 0d13a7ba01a78f7619f69655522000449193f4aa62fa0ed0a3c794480484af30 not found: ID does not exist" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.401758 4632 scope.go:117] "RemoveContainer" containerID="0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.402110 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589"} err="failed to get container status \"0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589\": rpc error: code = NotFound desc = could not find container \"0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589\": container with ID starting with 0848706fc94227d25a25441bdfc8b2affad934f1f484b5adb7c4470d7918e589 not found: ID does not exist" Mar 13 10:30:26 crc kubenswrapper[4632]: I0313 10:30:26.920153 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 10:30:32 crc kubenswrapper[4632]: I0313 10:30:32.980810 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 10:30:37 crc kubenswrapper[4632]: I0313 10:30:37.217159 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 10:30:37 crc kubenswrapper[4632]: I0313 10:30:37.217711 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="21ce0311-ff05-4626-9663-a373ae31eb56" containerName="kube-state-metrics" containerID="cri-o://1a16adc836bd27406a32b6f9e9672d40ce7e70e9caf414e0a9334fb34a8ec7ab" gracePeriod=30 Mar 13 10:30:37 crc kubenswrapper[4632]: I0313 10:30:37.918978 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 10:30:37 crc kubenswrapper[4632]: I0313 10:30:37.971559 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hfstv\" (UniqueName: \"kubernetes.io/projected/21ce0311-ff05-4626-9663-a373ae31eb56-kube-api-access-hfstv\") pod \"21ce0311-ff05-4626-9663-a373ae31eb56\" (UID: \"21ce0311-ff05-4626-9663-a373ae31eb56\") " Mar 13 10:30:37 crc kubenswrapper[4632]: I0313 10:30:37.990124 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21ce0311-ff05-4626-9663-a373ae31eb56-kube-api-access-hfstv" (OuterVolumeSpecName: "kube-api-access-hfstv") pod "21ce0311-ff05-4626-9663-a373ae31eb56" (UID: "21ce0311-ff05-4626-9663-a373ae31eb56"). InnerVolumeSpecName "kube-api-access-hfstv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.075479 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hfstv\" (UniqueName: \"kubernetes.io/projected/21ce0311-ff05-4626-9663-a373ae31eb56-kube-api-access-hfstv\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.079551 4632 generic.go:334] "Generic (PLEG): container finished" podID="21ce0311-ff05-4626-9663-a373ae31eb56" containerID="1a16adc836bd27406a32b6f9e9672d40ce7e70e9caf414e0a9334fb34a8ec7ab" exitCode=2 Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.079629 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.124629 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"21ce0311-ff05-4626-9663-a373ae31eb56","Type":"ContainerDied","Data":"1a16adc836bd27406a32b6f9e9672d40ce7e70e9caf414e0a9334fb34a8ec7ab"} Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.124715 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"21ce0311-ff05-4626-9663-a373ae31eb56","Type":"ContainerDied","Data":"96915bb97645358a6555ca60c9308596dd68c9b71a65da098dd5679653d9f202"} Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.124755 4632 scope.go:117] "RemoveContainer" containerID="1a16adc836bd27406a32b6f9e9672d40ce7e70e9caf414e0a9334fb34a8ec7ab" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.282860 4632 scope.go:117] "RemoveContainer" containerID="1a16adc836bd27406a32b6f9e9672d40ce7e70e9caf414e0a9334fb34a8ec7ab" Mar 13 10:30:38 crc kubenswrapper[4632]: E0313 10:30:38.288230 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a16adc836bd27406a32b6f9e9672d40ce7e70e9caf414e0a9334fb34a8ec7ab\": container with ID starting with 1a16adc836bd27406a32b6f9e9672d40ce7e70e9caf414e0a9334fb34a8ec7ab not found: ID does not exist" containerID="1a16adc836bd27406a32b6f9e9672d40ce7e70e9caf414e0a9334fb34a8ec7ab" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.288288 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a16adc836bd27406a32b6f9e9672d40ce7e70e9caf414e0a9334fb34a8ec7ab"} err="failed to get container status \"1a16adc836bd27406a32b6f9e9672d40ce7e70e9caf414e0a9334fb34a8ec7ab\": rpc error: code = NotFound desc = could not find container \"1a16adc836bd27406a32b6f9e9672d40ce7e70e9caf414e0a9334fb34a8ec7ab\": container with ID starting with 1a16adc836bd27406a32b6f9e9672d40ce7e70e9caf414e0a9334fb34a8ec7ab not found: ID does not exist" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.326510 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.336838 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347009 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 10:30:38 crc kubenswrapper[4632]: E0313 10:30:38.347497 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347516 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:38 crc kubenswrapper[4632]: E0313 10:30:38.347530 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="extract-utilities" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347537 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="extract-utilities" Mar 13 10:30:38 crc kubenswrapper[4632]: E0313 10:30:38.347558 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347565 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:38 crc kubenswrapper[4632]: E0313 10:30:38.347580 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e" containerName="oc" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347586 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e" containerName="oc" Mar 13 10:30:38 crc kubenswrapper[4632]: E0313 10:30:38.347596 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="registry-server" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347602 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="registry-server" Mar 13 10:30:38 crc kubenswrapper[4632]: E0313 10:30:38.347611 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="extract-content" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347619 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="extract-content" Mar 13 10:30:38 crc kubenswrapper[4632]: E0313 10:30:38.347631 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21ce0311-ff05-4626-9663-a373ae31eb56" containerName="kube-state-metrics" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347638 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="21ce0311-ff05-4626-9663-a373ae31eb56" containerName="kube-state-metrics" Mar 13 10:30:38 crc kubenswrapper[4632]: E0313 10:30:38.347650 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347656 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:38 crc kubenswrapper[4632]: E0313 10:30:38.347666 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347672 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:38 crc kubenswrapper[4632]: E0313 10:30:38.347680 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347686 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:38 crc kubenswrapper[4632]: E0313 10:30:38.347700 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon-log" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347706 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon-log" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347911 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347920 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347930 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0cabd29-ef3e-4808-8c92-3b032483789e" containerName="registry-server" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347952 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347962 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon-log" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347972 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e" containerName="oc" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.347983 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="21ce0311-ff05-4626-9663-a373ae31eb56" containerName="kube-state-metrics" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.348730 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.351635 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.351839 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.375078 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.406791 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26ce3314-15f1-490c-83e5-a1c609212437-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"26ce3314-15f1-490c-83e5-a1c609212437\") " pod="openstack/kube-state-metrics-0" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.407180 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/26ce3314-15f1-490c-83e5-a1c609212437-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"26ce3314-15f1-490c-83e5-a1c609212437\") " pod="openstack/kube-state-metrics-0" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.407380 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/26ce3314-15f1-490c-83e5-a1c609212437-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"26ce3314-15f1-490c-83e5-a1c609212437\") " pod="openstack/kube-state-metrics-0" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.407534 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nvb7r\" (UniqueName: \"kubernetes.io/projected/26ce3314-15f1-490c-83e5-a1c609212437-kube-api-access-nvb7r\") pod \"kube-state-metrics-0\" (UID: \"26ce3314-15f1-490c-83e5-a1c609212437\") " pod="openstack/kube-state-metrics-0" Mar 13 10:30:38 crc kubenswrapper[4632]: E0313 10:30:38.493877 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21ce0311_ff05_4626_9663_a373ae31eb56.slice/crio-96915bb97645358a6555ca60c9308596dd68c9b71a65da098dd5679653d9f202\": RecentStats: unable to find data in memory cache]" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.509678 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26ce3314-15f1-490c-83e5-a1c609212437-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"26ce3314-15f1-490c-83e5-a1c609212437\") " pod="openstack/kube-state-metrics-0" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.510065 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/26ce3314-15f1-490c-83e5-a1c609212437-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"26ce3314-15f1-490c-83e5-a1c609212437\") " pod="openstack/kube-state-metrics-0" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.510210 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/26ce3314-15f1-490c-83e5-a1c609212437-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"26ce3314-15f1-490c-83e5-a1c609212437\") " pod="openstack/kube-state-metrics-0" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.510320 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nvb7r\" (UniqueName: \"kubernetes.io/projected/26ce3314-15f1-490c-83e5-a1c609212437-kube-api-access-nvb7r\") pod \"kube-state-metrics-0\" (UID: \"26ce3314-15f1-490c-83e5-a1c609212437\") " pod="openstack/kube-state-metrics-0" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.519116 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26ce3314-15f1-490c-83e5-a1c609212437-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"26ce3314-15f1-490c-83e5-a1c609212437\") " pod="openstack/kube-state-metrics-0" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.528293 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/26ce3314-15f1-490c-83e5-a1c609212437-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"26ce3314-15f1-490c-83e5-a1c609212437\") " pod="openstack/kube-state-metrics-0" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.531625 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nvb7r\" (UniqueName: \"kubernetes.io/projected/26ce3314-15f1-490c-83e5-a1c609212437-kube-api-access-nvb7r\") pod \"kube-state-metrics-0\" (UID: \"26ce3314-15f1-490c-83e5-a1c609212437\") " pod="openstack/kube-state-metrics-0" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.549830 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/26ce3314-15f1-490c-83e5-a1c609212437-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"26ce3314-15f1-490c-83e5-a1c609212437\") " pod="openstack/kube-state-metrics-0" Mar 13 10:30:38 crc kubenswrapper[4632]: I0313 10:30:38.678864 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 10:30:39 crc kubenswrapper[4632]: I0313 10:30:39.201968 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 10:30:39 crc kubenswrapper[4632]: I0313 10:30:39.859251 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:30:39 crc kubenswrapper[4632]: I0313 10:30:39.859790 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="ceilometer-central-agent" containerID="cri-o://38c2fc76cd7277bb1d4be79e157389e249663d48647565e9e7553c7283f85329" gracePeriod=30 Mar 13 10:30:39 crc kubenswrapper[4632]: I0313 10:30:39.859921 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="proxy-httpd" containerID="cri-o://aafa45b6b50dc62147a1600ded978371569f724c32856206cf7196aba295169c" gracePeriod=30 Mar 13 10:30:39 crc kubenswrapper[4632]: I0313 10:30:39.859976 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="sg-core" containerID="cri-o://2e3874df0599fa71baea0baa935a02f99ef35801e9c669049d94507f046a75da" gracePeriod=30 Mar 13 10:30:39 crc kubenswrapper[4632]: I0313 10:30:39.860010 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="ceilometer-notification-agent" containerID="cri-o://8fc781f30b1bf552b69e65bd6fbe07dd0ebc73a78af0ab302a58fadf4aa1b637" gracePeriod=30 Mar 13 10:30:40 crc kubenswrapper[4632]: I0313 10:30:40.057142 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21ce0311-ff05-4626-9663-a373ae31eb56" path="/var/lib/kubelet/pods/21ce0311-ff05-4626-9663-a373ae31eb56/volumes" Mar 13 10:30:40 crc kubenswrapper[4632]: I0313 10:30:40.101264 4632 generic.go:334] "Generic (PLEG): container finished" podID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerID="aafa45b6b50dc62147a1600ded978371569f724c32856206cf7196aba295169c" exitCode=0 Mar 13 10:30:40 crc kubenswrapper[4632]: I0313 10:30:40.101515 4632 generic.go:334] "Generic (PLEG): container finished" podID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerID="2e3874df0599fa71baea0baa935a02f99ef35801e9c669049d94507f046a75da" exitCode=2 Mar 13 10:30:40 crc kubenswrapper[4632]: I0313 10:30:40.101748 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc34b88a-a0cc-4ef1-8267-30f73d9712e7","Type":"ContainerDied","Data":"aafa45b6b50dc62147a1600ded978371569f724c32856206cf7196aba295169c"} Mar 13 10:30:40 crc kubenswrapper[4632]: I0313 10:30:40.101776 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc34b88a-a0cc-4ef1-8267-30f73d9712e7","Type":"ContainerDied","Data":"2e3874df0599fa71baea0baa935a02f99ef35801e9c669049d94507f046a75da"} Mar 13 10:30:40 crc kubenswrapper[4632]: I0313 10:30:40.109478 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26ce3314-15f1-490c-83e5-a1c609212437","Type":"ContainerStarted","Data":"58a379ff336fdb7ea32918613cb19f466310efd4b6a05aea4a6b056ea6809ecf"} Mar 13 10:30:40 crc kubenswrapper[4632]: I0313 10:30:40.109531 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"26ce3314-15f1-490c-83e5-a1c609212437","Type":"ContainerStarted","Data":"2868b0d004bb34c44e511b76b488032e3fe9956857da8e812917313f0b5776f6"} Mar 13 10:30:40 crc kubenswrapper[4632]: I0313 10:30:40.109650 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 13 10:30:40 crc kubenswrapper[4632]: I0313 10:30:40.133480 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.744160599 podStartE2EDuration="2.13345633s" podCreationTimestamp="2026-03-13 10:30:38 +0000 UTC" firstStartedPulling="2026-03-13 10:30:39.23663199 +0000 UTC m=+1613.259162113" lastFinishedPulling="2026-03-13 10:30:39.625927711 +0000 UTC m=+1613.648457844" observedRunningTime="2026-03-13 10:30:40.124398992 +0000 UTC m=+1614.146929125" watchObservedRunningTime="2026-03-13 10:30:40.13345633 +0000 UTC m=+1614.155986473" Mar 13 10:30:40 crc kubenswrapper[4632]: I0313 10:30:40.461313 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:30:40 crc kubenswrapper[4632]: I0313 10:30:40.461370 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:30:41 crc kubenswrapper[4632]: I0313 10:30:41.119674 4632 generic.go:334] "Generic (PLEG): container finished" podID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerID="38c2fc76cd7277bb1d4be79e157389e249663d48647565e9e7553c7283f85329" exitCode=0 Mar 13 10:30:41 crc kubenswrapper[4632]: I0313 10:30:41.119750 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc34b88a-a0cc-4ef1-8267-30f73d9712e7","Type":"ContainerDied","Data":"38c2fc76cd7277bb1d4be79e157389e249663d48647565e9e7553c7283f85329"} Mar 13 10:30:41 crc kubenswrapper[4632]: I0313 10:30:41.746261 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:30:41 crc kubenswrapper[4632]: I0313 10:30:41.827493 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-log-httpd\") pod \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " Mar 13 10:30:41 crc kubenswrapper[4632]: I0313 10:30:41.827668 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-combined-ca-bundle\") pod \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " Mar 13 10:30:41 crc kubenswrapper[4632]: I0313 10:30:41.827772 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-config-data\") pod \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " Mar 13 10:30:41 crc kubenswrapper[4632]: I0313 10:30:41.827985 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-sg-core-conf-yaml\") pod \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " Mar 13 10:30:41 crc kubenswrapper[4632]: I0313 10:30:41.828180 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nvm2p\" (UniqueName: \"kubernetes.io/projected/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-kube-api-access-nvm2p\") pod \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " Mar 13 10:30:41 crc kubenswrapper[4632]: I0313 10:30:41.948545 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-scripts\") pod \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " Mar 13 10:30:41 crc kubenswrapper[4632]: I0313 10:30:41.948626 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-run-httpd\") pod \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\" (UID: \"bc34b88a-a0cc-4ef1-8267-30f73d9712e7\") " Mar 13 10:30:41 crc kubenswrapper[4632]: I0313 10:30:41.859638 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bc34b88a-a0cc-4ef1-8267-30f73d9712e7" (UID: "bc34b88a-a0cc-4ef1-8267-30f73d9712e7"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:30:41 crc kubenswrapper[4632]: I0313 10:30:41.954870 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bc34b88a-a0cc-4ef1-8267-30f73d9712e7" (UID: "bc34b88a-a0cc-4ef1-8267-30f73d9712e7"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:30:41 crc kubenswrapper[4632]: I0313 10:30:41.998659 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-scripts" (OuterVolumeSpecName: "scripts") pod "bc34b88a-a0cc-4ef1-8267-30f73d9712e7" (UID: "bc34b88a-a0cc-4ef1-8267-30f73d9712e7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.005343 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bc34b88a-a0cc-4ef1-8267-30f73d9712e7" (UID: "bc34b88a-a0cc-4ef1-8267-30f73d9712e7"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.014688 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-kube-api-access-nvm2p" (OuterVolumeSpecName: "kube-api-access-nvm2p") pod "bc34b88a-a0cc-4ef1-8267-30f73d9712e7" (UID: "bc34b88a-a0cc-4ef1-8267-30f73d9712e7"). InnerVolumeSpecName "kube-api-access-nvm2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.059836 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nvm2p\" (UniqueName: \"kubernetes.io/projected/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-kube-api-access-nvm2p\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.059869 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.059879 4632 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.059889 4632 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.059900 4632 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.163886 4632 generic.go:334] "Generic (PLEG): container finished" podID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerID="8fc781f30b1bf552b69e65bd6fbe07dd0ebc73a78af0ab302a58fadf4aa1b637" exitCode=0 Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.164002 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.170894 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc34b88a-a0cc-4ef1-8267-30f73d9712e7" (UID: "bc34b88a-a0cc-4ef1-8267-30f73d9712e7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.182518 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc34b88a-a0cc-4ef1-8267-30f73d9712e7","Type":"ContainerDied","Data":"8fc781f30b1bf552b69e65bd6fbe07dd0ebc73a78af0ab302a58fadf4aa1b637"} Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.182581 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc34b88a-a0cc-4ef1-8267-30f73d9712e7","Type":"ContainerDied","Data":"43e2d3c1f3ad2fd1a1419876a0a3f1ee25556cf01d81619d7588168a228ea654"} Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.182606 4632 scope.go:117] "RemoveContainer" containerID="aafa45b6b50dc62147a1600ded978371569f724c32856206cf7196aba295169c" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.217694 4632 scope.go:117] "RemoveContainer" containerID="2e3874df0599fa71baea0baa935a02f99ef35801e9c669049d94507f046a75da" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.256195 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-config-data" (OuterVolumeSpecName: "config-data") pod "bc34b88a-a0cc-4ef1-8267-30f73d9712e7" (UID: "bc34b88a-a0cc-4ef1-8267-30f73d9712e7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.266164 4632 scope.go:117] "RemoveContainer" containerID="8fc781f30b1bf552b69e65bd6fbe07dd0ebc73a78af0ab302a58fadf4aa1b637" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.269479 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.269526 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc34b88a-a0cc-4ef1-8267-30f73d9712e7-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.297528 4632 scope.go:117] "RemoveContainer" containerID="38c2fc76cd7277bb1d4be79e157389e249663d48647565e9e7553c7283f85329" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.333459 4632 scope.go:117] "RemoveContainer" containerID="aafa45b6b50dc62147a1600ded978371569f724c32856206cf7196aba295169c" Mar 13 10:30:42 crc kubenswrapper[4632]: E0313 10:30:42.333977 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aafa45b6b50dc62147a1600ded978371569f724c32856206cf7196aba295169c\": container with ID starting with aafa45b6b50dc62147a1600ded978371569f724c32856206cf7196aba295169c not found: ID does not exist" containerID="aafa45b6b50dc62147a1600ded978371569f724c32856206cf7196aba295169c" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.334026 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aafa45b6b50dc62147a1600ded978371569f724c32856206cf7196aba295169c"} err="failed to get container status \"aafa45b6b50dc62147a1600ded978371569f724c32856206cf7196aba295169c\": rpc error: code = NotFound desc = could not find container \"aafa45b6b50dc62147a1600ded978371569f724c32856206cf7196aba295169c\": container with ID starting with aafa45b6b50dc62147a1600ded978371569f724c32856206cf7196aba295169c not found: ID does not exist" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.334054 4632 scope.go:117] "RemoveContainer" containerID="2e3874df0599fa71baea0baa935a02f99ef35801e9c669049d94507f046a75da" Mar 13 10:30:42 crc kubenswrapper[4632]: E0313 10:30:42.334549 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e3874df0599fa71baea0baa935a02f99ef35801e9c669049d94507f046a75da\": container with ID starting with 2e3874df0599fa71baea0baa935a02f99ef35801e9c669049d94507f046a75da not found: ID does not exist" containerID="2e3874df0599fa71baea0baa935a02f99ef35801e9c669049d94507f046a75da" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.334586 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e3874df0599fa71baea0baa935a02f99ef35801e9c669049d94507f046a75da"} err="failed to get container status \"2e3874df0599fa71baea0baa935a02f99ef35801e9c669049d94507f046a75da\": rpc error: code = NotFound desc = could not find container \"2e3874df0599fa71baea0baa935a02f99ef35801e9c669049d94507f046a75da\": container with ID starting with 2e3874df0599fa71baea0baa935a02f99ef35801e9c669049d94507f046a75da not found: ID does not exist" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.334604 4632 scope.go:117] "RemoveContainer" containerID="8fc781f30b1bf552b69e65bd6fbe07dd0ebc73a78af0ab302a58fadf4aa1b637" Mar 13 10:30:42 crc kubenswrapper[4632]: E0313 10:30:42.334932 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fc781f30b1bf552b69e65bd6fbe07dd0ebc73a78af0ab302a58fadf4aa1b637\": container with ID starting with 8fc781f30b1bf552b69e65bd6fbe07dd0ebc73a78af0ab302a58fadf4aa1b637 not found: ID does not exist" containerID="8fc781f30b1bf552b69e65bd6fbe07dd0ebc73a78af0ab302a58fadf4aa1b637" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.335064 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fc781f30b1bf552b69e65bd6fbe07dd0ebc73a78af0ab302a58fadf4aa1b637"} err="failed to get container status \"8fc781f30b1bf552b69e65bd6fbe07dd0ebc73a78af0ab302a58fadf4aa1b637\": rpc error: code = NotFound desc = could not find container \"8fc781f30b1bf552b69e65bd6fbe07dd0ebc73a78af0ab302a58fadf4aa1b637\": container with ID starting with 8fc781f30b1bf552b69e65bd6fbe07dd0ebc73a78af0ab302a58fadf4aa1b637 not found: ID does not exist" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.335084 4632 scope.go:117] "RemoveContainer" containerID="38c2fc76cd7277bb1d4be79e157389e249663d48647565e9e7553c7283f85329" Mar 13 10:30:42 crc kubenswrapper[4632]: E0313 10:30:42.335383 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38c2fc76cd7277bb1d4be79e157389e249663d48647565e9e7553c7283f85329\": container with ID starting with 38c2fc76cd7277bb1d4be79e157389e249663d48647565e9e7553c7283f85329 not found: ID does not exist" containerID="38c2fc76cd7277bb1d4be79e157389e249663d48647565e9e7553c7283f85329" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.335418 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38c2fc76cd7277bb1d4be79e157389e249663d48647565e9e7553c7283f85329"} err="failed to get container status \"38c2fc76cd7277bb1d4be79e157389e249663d48647565e9e7553c7283f85329\": rpc error: code = NotFound desc = could not find container \"38c2fc76cd7277bb1d4be79e157389e249663d48647565e9e7553c7283f85329\": container with ID starting with 38c2fc76cd7277bb1d4be79e157389e249663d48647565e9e7553c7283f85329 not found: ID does not exist" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.512692 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.521075 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.547209 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:30:42 crc kubenswrapper[4632]: E0313 10:30:42.547722 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="ceilometer-notification-agent" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.547749 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="ceilometer-notification-agent" Mar 13 10:30:42 crc kubenswrapper[4632]: E0313 10:30:42.547770 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="sg-core" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.547782 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="sg-core" Mar 13 10:30:42 crc kubenswrapper[4632]: E0313 10:30:42.547797 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="proxy-httpd" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.547806 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="proxy-httpd" Mar 13 10:30:42 crc kubenswrapper[4632]: E0313 10:30:42.547832 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="ceilometer-central-agent" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.547840 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="ceilometer-central-agent" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.548086 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.548111 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e4afb91-ce26-4325-89c9-2542da2ec48a" containerName="horizon" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.548132 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="ceilometer-central-agent" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.548148 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="ceilometer-notification-agent" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.548168 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="proxy-httpd" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.548180 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" containerName="sg-core" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.555172 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.561842 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.562348 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.562682 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.582649 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.677629 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.677704 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.677744 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac97dc03-9537-4f95-bb79-5bb60a99089d-run-httpd\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.677783 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-scripts\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.677872 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-config-data\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.677907 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22tlw\" (UniqueName: \"kubernetes.io/projected/ac97dc03-9537-4f95-bb79-5bb60a99089d-kube-api-access-22tlw\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.677969 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac97dc03-9537-4f95-bb79-5bb60a99089d-log-httpd\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.678078 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.779835 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-config-data\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.779908 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22tlw\" (UniqueName: \"kubernetes.io/projected/ac97dc03-9537-4f95-bb79-5bb60a99089d-kube-api-access-22tlw\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.779974 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac97dc03-9537-4f95-bb79-5bb60a99089d-log-httpd\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.780049 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.780085 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.780113 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.780133 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac97dc03-9537-4f95-bb79-5bb60a99089d-run-httpd\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.780820 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac97dc03-9537-4f95-bb79-5bb60a99089d-run-httpd\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.780866 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-scripts\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.781493 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac97dc03-9537-4f95-bb79-5bb60a99089d-log-httpd\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.785074 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.788367 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.790477 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.793437 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-scripts\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.795649 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-config-data\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.819009 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22tlw\" (UniqueName: \"kubernetes.io/projected/ac97dc03-9537-4f95-bb79-5bb60a99089d-kube-api-access-22tlw\") pod \"ceilometer-0\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " pod="openstack/ceilometer-0" Mar 13 10:30:42 crc kubenswrapper[4632]: I0313 10:30:42.904302 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:30:43 crc kubenswrapper[4632]: I0313 10:30:43.530438 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:30:43 crc kubenswrapper[4632]: W0313 10:30:43.543251 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac97dc03_9537_4f95_bb79_5bb60a99089d.slice/crio-6c35a65f59ec813bb19b2b3e4862d24780f1cf59570c0c358308767506eead20 WatchSource:0}: Error finding container 6c35a65f59ec813bb19b2b3e4862d24780f1cf59570c0c358308767506eead20: Status 404 returned error can't find the container with id 6c35a65f59ec813bb19b2b3e4862d24780f1cf59570c0c358308767506eead20 Mar 13 10:30:44 crc kubenswrapper[4632]: I0313 10:30:44.060306 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc34b88a-a0cc-4ef1-8267-30f73d9712e7" path="/var/lib/kubelet/pods/bc34b88a-a0cc-4ef1-8267-30f73d9712e7/volumes" Mar 13 10:30:44 crc kubenswrapper[4632]: I0313 10:30:44.193614 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac97dc03-9537-4f95-bb79-5bb60a99089d","Type":"ContainerStarted","Data":"3da76186915cfbbbe688750a6110b1e64143d37e61c44ef62a9740eabb32c983"} Mar 13 10:30:44 crc kubenswrapper[4632]: I0313 10:30:44.193680 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac97dc03-9537-4f95-bb79-5bb60a99089d","Type":"ContainerStarted","Data":"6c35a65f59ec813bb19b2b3e4862d24780f1cf59570c0c358308767506eead20"} Mar 13 10:30:45 crc kubenswrapper[4632]: I0313 10:30:45.221916 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac97dc03-9537-4f95-bb79-5bb60a99089d","Type":"ContainerStarted","Data":"5d34565e3f3d53e4eb4eec7fc127b7d0ef95db5c894a8b9fbc65ec70d12e4d20"} Mar 13 10:30:46 crc kubenswrapper[4632]: I0313 10:30:46.238543 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac97dc03-9537-4f95-bb79-5bb60a99089d","Type":"ContainerStarted","Data":"7784ac325dc1b12d740a758a07a8e9e03da012db50eef1bc62b207161880f530"} Mar 13 10:30:47 crc kubenswrapper[4632]: I0313 10:30:47.425893 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 10:30:48 crc kubenswrapper[4632]: I0313 10:30:48.266992 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac97dc03-9537-4f95-bb79-5bb60a99089d","Type":"ContainerStarted","Data":"9a2b2ece3b9e850a4d4ebe5776040511a71de7bd0fba43340538aa166e80ade2"} Mar 13 10:30:48 crc kubenswrapper[4632]: I0313 10:30:48.267386 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 10:30:48 crc kubenswrapper[4632]: I0313 10:30:48.292580 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.794955469 podStartE2EDuration="6.292554821s" podCreationTimestamp="2026-03-13 10:30:42 +0000 UTC" firstStartedPulling="2026-03-13 10:30:43.546552506 +0000 UTC m=+1617.569082639" lastFinishedPulling="2026-03-13 10:30:47.044151868 +0000 UTC m=+1621.066681991" observedRunningTime="2026-03-13 10:30:48.286647559 +0000 UTC m=+1622.309177692" watchObservedRunningTime="2026-03-13 10:30:48.292554821 +0000 UTC m=+1622.315084964" Mar 13 10:30:48 crc kubenswrapper[4632]: I0313 10:30:48.646994 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 10:30:48 crc kubenswrapper[4632]: I0313 10:30:48.964165 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 13 10:30:53 crc kubenswrapper[4632]: I0313 10:30:53.858705 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="211718f0-f29c-457b-bc2b-487bb76d4801" containerName="rabbitmq" containerID="cri-o://40d92cf95f1cc26685e0359414b43dbdc31eeb90ab4b39c564b241d3fcc263fe" gracePeriod=604794 Mar 13 10:30:55 crc kubenswrapper[4632]: I0313 10:30:55.013765 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="159c6cee-c82b-4725-82d6-dbd27216f53c" containerName="rabbitmq" containerID="cri-o://d8fa91cb90a686638520d703bb5ab925cd9f40c680cdbe53067f753945b6ae3f" gracePeriod=604794 Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.514148 4632 generic.go:334] "Generic (PLEG): container finished" podID="211718f0-f29c-457b-bc2b-487bb76d4801" containerID="40d92cf95f1cc26685e0359414b43dbdc31eeb90ab4b39c564b241d3fcc263fe" exitCode=0 Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.514372 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"211718f0-f29c-457b-bc2b-487bb76d4801","Type":"ContainerDied","Data":"40d92cf95f1cc26685e0359414b43dbdc31eeb90ab4b39c564b241d3fcc263fe"} Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.514724 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"211718f0-f29c-457b-bc2b-487bb76d4801","Type":"ContainerDied","Data":"fd0dcad1534e2c23d238622a824c4e32c97444e16220054d2406cb0e89183756"} Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.514741 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd0dcad1534e2c23d238622a824c4e32c97444e16220054d2406cb0e89183756" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.529759 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.634168 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="159c6cee-c82b-4725-82d6-dbd27216f53c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.722650 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-server-conf\") pod \"211718f0-f29c-457b-bc2b-487bb76d4801\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.722727 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-confd\") pod \"211718f0-f29c-457b-bc2b-487bb76d4801\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.722751 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-plugins-conf\") pod \"211718f0-f29c-457b-bc2b-487bb76d4801\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.722793 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"211718f0-f29c-457b-bc2b-487bb76d4801\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.722829 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-plugins\") pod \"211718f0-f29c-457b-bc2b-487bb76d4801\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.722957 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/211718f0-f29c-457b-bc2b-487bb76d4801-erlang-cookie-secret\") pod \"211718f0-f29c-457b-bc2b-487bb76d4801\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.722992 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kfgh\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-kube-api-access-6kfgh\") pod \"211718f0-f29c-457b-bc2b-487bb76d4801\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.723011 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-tls\") pod \"211718f0-f29c-457b-bc2b-487bb76d4801\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.723316 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "211718f0-f29c-457b-bc2b-487bb76d4801" (UID: "211718f0-f29c-457b-bc2b-487bb76d4801"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.723384 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-erlang-cookie\") pod \"211718f0-f29c-457b-bc2b-487bb76d4801\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.723431 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-config-data\") pod \"211718f0-f29c-457b-bc2b-487bb76d4801\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.723470 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "211718f0-f29c-457b-bc2b-487bb76d4801" (UID: "211718f0-f29c-457b-bc2b-487bb76d4801"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.723489 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/211718f0-f29c-457b-bc2b-487bb76d4801-pod-info\") pod \"211718f0-f29c-457b-bc2b-487bb76d4801\" (UID: \"211718f0-f29c-457b-bc2b-487bb76d4801\") " Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.724414 4632 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.724436 4632 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.729308 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "211718f0-f29c-457b-bc2b-487bb76d4801" (UID: "211718f0-f29c-457b-bc2b-487bb76d4801"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.740257 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/211718f0-f29c-457b-bc2b-487bb76d4801-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "211718f0-f29c-457b-bc2b-487bb76d4801" (UID: "211718f0-f29c-457b-bc2b-487bb76d4801"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.740455 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "211718f0-f29c-457b-bc2b-487bb76d4801" (UID: "211718f0-f29c-457b-bc2b-487bb76d4801"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.743991 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "211718f0-f29c-457b-bc2b-487bb76d4801" (UID: "211718f0-f29c-457b-bc2b-487bb76d4801"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.749867 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-kube-api-access-6kfgh" (OuterVolumeSpecName: "kube-api-access-6kfgh") pod "211718f0-f29c-457b-bc2b-487bb76d4801" (UID: "211718f0-f29c-457b-bc2b-487bb76d4801"). InnerVolumeSpecName "kube-api-access-6kfgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.750053 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/211718f0-f29c-457b-bc2b-487bb76d4801-pod-info" (OuterVolumeSpecName: "pod-info") pod "211718f0-f29c-457b-bc2b-487bb76d4801" (UID: "211718f0-f29c-457b-bc2b-487bb76d4801"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.819119 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-config-data" (OuterVolumeSpecName: "config-data") pod "211718f0-f29c-457b-bc2b-487bb76d4801" (UID: "211718f0-f29c-457b-bc2b-487bb76d4801"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.836659 4632 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.836691 4632 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/211718f0-f29c-457b-bc2b-487bb76d4801-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.836704 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kfgh\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-kube-api-access-6kfgh\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.836713 4632 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.836723 4632 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.836733 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.836741 4632 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/211718f0-f29c-457b-bc2b-487bb76d4801-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.838370 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-server-conf" (OuterVolumeSpecName: "server-conf") pod "211718f0-f29c-457b-bc2b-487bb76d4801" (UID: "211718f0-f29c-457b-bc2b-487bb76d4801"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.887517 4632 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.940267 4632 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/211718f0-f29c-457b-bc2b-487bb76d4801-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.940306 4632 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:00 crc kubenswrapper[4632]: I0313 10:31:00.954987 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "211718f0-f29c-457b-bc2b-487bb76d4801" (UID: "211718f0-f29c-457b-bc2b-487bb76d4801"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.050423 4632 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/211718f0-f29c-457b-bc2b-487bb76d4801-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.555907 4632 generic.go:334] "Generic (PLEG): container finished" podID="159c6cee-c82b-4725-82d6-dbd27216f53c" containerID="d8fa91cb90a686638520d703bb5ab925cd9f40c680cdbe53067f753945b6ae3f" exitCode=0 Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.556278 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.558094 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"159c6cee-c82b-4725-82d6-dbd27216f53c","Type":"ContainerDied","Data":"d8fa91cb90a686638520d703bb5ab925cd9f40c680cdbe53067f753945b6ae3f"} Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.648783 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.677857 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.739722 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 10:31:01 crc kubenswrapper[4632]: E0313 10:31:01.742909 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="211718f0-f29c-457b-bc2b-487bb76d4801" containerName="setup-container" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.742980 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="211718f0-f29c-457b-bc2b-487bb76d4801" containerName="setup-container" Mar 13 10:31:01 crc kubenswrapper[4632]: E0313 10:31:01.743044 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="211718f0-f29c-457b-bc2b-487bb76d4801" containerName="rabbitmq" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.743054 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="211718f0-f29c-457b-bc2b-487bb76d4801" containerName="rabbitmq" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.743524 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="211718f0-f29c-457b-bc2b-487bb76d4801" containerName="rabbitmq" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.745768 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.750854 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.751520 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.751779 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.752082 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.752220 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.752435 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-x424t" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.753033 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.760958 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.874201 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-config-data\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.874495 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.874590 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.874733 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.874869 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vfbl\" (UniqueName: \"kubernetes.io/projected/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-kube-api-access-9vfbl\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.875024 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.875134 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.875257 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.875408 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.875555 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.875662 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.879790 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.977179 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-plugins-conf\") pod \"159c6cee-c82b-4725-82d6-dbd27216f53c\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.977290 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8hmx\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-kube-api-access-k8hmx\") pod \"159c6cee-c82b-4725-82d6-dbd27216f53c\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.978603 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/159c6cee-c82b-4725-82d6-dbd27216f53c-pod-info\") pod \"159c6cee-c82b-4725-82d6-dbd27216f53c\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.978674 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-confd\") pod \"159c6cee-c82b-4725-82d6-dbd27216f53c\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.978701 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/159c6cee-c82b-4725-82d6-dbd27216f53c-erlang-cookie-secret\") pod \"159c6cee-c82b-4725-82d6-dbd27216f53c\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.978735 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"159c6cee-c82b-4725-82d6-dbd27216f53c\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.978811 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-tls\") pod \"159c6cee-c82b-4725-82d6-dbd27216f53c\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.978847 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-plugins\") pod \"159c6cee-c82b-4725-82d6-dbd27216f53c\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.978987 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-server-conf\") pod \"159c6cee-c82b-4725-82d6-dbd27216f53c\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.979026 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-erlang-cookie\") pod \"159c6cee-c82b-4725-82d6-dbd27216f53c\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.979079 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-config-data\") pod \"159c6cee-c82b-4725-82d6-dbd27216f53c\" (UID: \"159c6cee-c82b-4725-82d6-dbd27216f53c\") " Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.979361 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "159c6cee-c82b-4725-82d6-dbd27216f53c" (UID: "159c6cee-c82b-4725-82d6-dbd27216f53c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.979418 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.979673 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.979744 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.979875 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-config-data\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.980078 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.980138 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.980221 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.980321 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vfbl\" (UniqueName: \"kubernetes.io/projected/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-kube-api-access-9vfbl\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.980421 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.980489 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.980618 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.980789 4632 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.986574 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "159c6cee-c82b-4725-82d6-dbd27216f53c" (UID: "159c6cee-c82b-4725-82d6-dbd27216f53c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.995898 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.997265 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-server-conf\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:01 crc kubenswrapper[4632]: I0313 10:31:01.997713 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-config-data\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.001829 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.002090 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "159c6cee-c82b-4725-82d6-dbd27216f53c" (UID: "159c6cee-c82b-4725-82d6-dbd27216f53c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.003513 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.005589 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.005762 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "159c6cee-c82b-4725-82d6-dbd27216f53c" (UID: "159c6cee-c82b-4725-82d6-dbd27216f53c"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.011102 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "159c6cee-c82b-4725-82d6-dbd27216f53c" (UID: "159c6cee-c82b-4725-82d6-dbd27216f53c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.020777 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-kube-api-access-k8hmx" (OuterVolumeSpecName: "kube-api-access-k8hmx") pod "159c6cee-c82b-4725-82d6-dbd27216f53c" (UID: "159c6cee-c82b-4725-82d6-dbd27216f53c"). InnerVolumeSpecName "kube-api-access-k8hmx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.020925 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.026838 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-pod-info\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.026900 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.027289 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/159c6cee-c82b-4725-82d6-dbd27216f53c-pod-info" (OuterVolumeSpecName: "pod-info") pod "159c6cee-c82b-4725-82d6-dbd27216f53c" (UID: "159c6cee-c82b-4725-82d6-dbd27216f53c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.031769 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.038807 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/159c6cee-c82b-4725-82d6-dbd27216f53c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "159c6cee-c82b-4725-82d6-dbd27216f53c" (UID: "159c6cee-c82b-4725-82d6-dbd27216f53c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.082848 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vfbl\" (UniqueName: \"kubernetes.io/projected/c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e-kube-api-access-9vfbl\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.100249 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="211718f0-f29c-457b-bc2b-487bb76d4801" path="/var/lib/kubelet/pods/211718f0-f29c-457b-bc2b-487bb76d4801/volumes" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.100652 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8hmx\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-kube-api-access-k8hmx\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.100684 4632 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/159c6cee-c82b-4725-82d6-dbd27216f53c-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.100696 4632 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/159c6cee-c82b-4725-82d6-dbd27216f53c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.100732 4632 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.100742 4632 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.100752 4632 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.100762 4632 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.156057 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-config-data" (OuterVolumeSpecName: "config-data") pod "159c6cee-c82b-4725-82d6-dbd27216f53c" (UID: "159c6cee-c82b-4725-82d6-dbd27216f53c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.170231 4632 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.171593 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-server-0\" (UID: \"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e\") " pod="openstack/rabbitmq-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.202341 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.202382 4632 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.205448 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.229398 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-server-conf" (OuterVolumeSpecName: "server-conf") pod "159c6cee-c82b-4725-82d6-dbd27216f53c" (UID: "159c6cee-c82b-4725-82d6-dbd27216f53c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.305471 4632 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/159c6cee-c82b-4725-82d6-dbd27216f53c-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.325311 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "159c6cee-c82b-4725-82d6-dbd27216f53c" (UID: "159c6cee-c82b-4725-82d6-dbd27216f53c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.408388 4632 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/159c6cee-c82b-4725-82d6-dbd27216f53c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.593330 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"159c6cee-c82b-4725-82d6-dbd27216f53c","Type":"ContainerDied","Data":"06613fdc2799f04ea62de7d5a6995bb48161830d28a55edb1ede1542c640e10e"} Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.593736 4632 scope.go:117] "RemoveContainer" containerID="d8fa91cb90a686638520d703bb5ab925cd9f40c680cdbe53067f753945b6ae3f" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.593407 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.684022 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.705468 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.711174 4632 scope.go:117] "RemoveContainer" containerID="d5bd67d741203861cfd1afa23ec3f20fd6236a99625563ac3c10816dbb2a6677" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.744035 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 10:31:02 crc kubenswrapper[4632]: E0313 10:31:02.744624 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="159c6cee-c82b-4725-82d6-dbd27216f53c" containerName="setup-container" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.744642 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="159c6cee-c82b-4725-82d6-dbd27216f53c" containerName="setup-container" Mar 13 10:31:02 crc kubenswrapper[4632]: E0313 10:31:02.744654 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="159c6cee-c82b-4725-82d6-dbd27216f53c" containerName="rabbitmq" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.744664 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="159c6cee-c82b-4725-82d6-dbd27216f53c" containerName="rabbitmq" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.744911 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="159c6cee-c82b-4725-82d6-dbd27216f53c" containerName="rabbitmq" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.752444 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.767015 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.768808 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.768995 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.769102 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.769163 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.769288 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-m5r4h" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.769390 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.768813 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.818960 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.819051 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.820415 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.820474 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqgkl\" (UniqueName: \"kubernetes.io/projected/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-kube-api-access-fqgkl\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.820508 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.820565 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.820690 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.820784 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.820850 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.820901 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.820968 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.925802 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.926240 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.926375 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.926624 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.926800 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.927076 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.927234 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.927275 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.927574 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.927624 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.928171 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.927643 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqgkl\" (UniqueName: \"kubernetes.io/projected/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-kube-api-access-fqgkl\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.928310 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.928356 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.928517 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.928531 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.935185 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.960551 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.961652 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.962774 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.963019 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqgkl\" (UniqueName: \"kubernetes.io/projected/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-kube-api-access-fqgkl\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.969176 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.971111 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a3d80d9f-c956-40f5-b2e1-8aea2f136b6e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:02 crc kubenswrapper[4632]: I0313 10:31:02.996049 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:03 crc kubenswrapper[4632]: I0313 10:31:03.095902 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:03 crc kubenswrapper[4632]: I0313 10:31:03.634540 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e","Type":"ContainerStarted","Data":"66423cf2bd8e9a56e0d8a5e063485d987fb7b493b018dca40689aaf9e621933f"} Mar 13 10:31:03 crc kubenswrapper[4632]: I0313 10:31:03.800555 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 10:31:04 crc kubenswrapper[4632]: I0313 10:31:04.057043 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="159c6cee-c82b-4725-82d6-dbd27216f53c" path="/var/lib/kubelet/pods/159c6cee-c82b-4725-82d6-dbd27216f53c/volumes" Mar 13 10:31:04 crc kubenswrapper[4632]: I0313 10:31:04.665507 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e","Type":"ContainerStarted","Data":"221855cb8b608a0fa79bdaa1a68882b74336883a0052291e98bb95a394015359"} Mar 13 10:31:05 crc kubenswrapper[4632]: I0313 10:31:05.376460 4632 scope.go:117] "RemoveContainer" containerID="92d546a480b1e583e7b11dc48ab2d570a4a8d7af0616de2352d72ca175520f17" Mar 13 10:31:05 crc kubenswrapper[4632]: I0313 10:31:05.431411 4632 scope.go:117] "RemoveContainer" containerID="40d92cf95f1cc26685e0359414b43dbdc31eeb90ab4b39c564b241d3fcc263fe" Mar 13 10:31:05 crc kubenswrapper[4632]: I0313 10:31:05.485789 4632 scope.go:117] "RemoveContainer" containerID="9011fe3e8ff19daa76b8d8bddf336d224d69f10272938404d994caa9a1a4d6ee" Mar 13 10:31:05 crc kubenswrapper[4632]: I0313 10:31:05.598583 4632 scope.go:117] "RemoveContainer" containerID="c6b6fdf02c5b942ff5eb86fa09449efd1927d429db47c31ad2d68c9602235d4f" Mar 13 10:31:05 crc kubenswrapper[4632]: I0313 10:31:05.643928 4632 scope.go:117] "RemoveContainer" containerID="572bb794023bd7d53a23050c721933f004db547126df9eaf9b5f8e767603f2d3" Mar 13 10:31:05 crc kubenswrapper[4632]: I0313 10:31:05.693447 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e","Type":"ContainerStarted","Data":"f69527ec2be6e4df808bf875b701d62b874066c28a489a38ef573bdae2b131dc"} Mar 13 10:31:05 crc kubenswrapper[4632]: I0313 10:31:05.763256 4632 scope.go:117] "RemoveContainer" containerID="6257821be47ec7e5943095f3b1d29a6e6fd0a1190515cb74642f7cb762d806d1" Mar 13 10:31:05 crc kubenswrapper[4632]: I0313 10:31:05.900512 4632 scope.go:117] "RemoveContainer" containerID="98a44d8e524895de3db65a2da91c25a6875681d7e31dfa6eb205635df601d593" Mar 13 10:31:05 crc kubenswrapper[4632]: I0313 10:31:05.969836 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57b9bf8b5-98n78"] Mar 13 10:31:05 crc kubenswrapper[4632]: I0313 10:31:05.983076 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:05 crc kubenswrapper[4632]: I0313 10:31:05.991630 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.045291 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57b9bf8b5-98n78"] Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.058269 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5pxf\" (UniqueName: \"kubernetes.io/projected/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-kube-api-access-v5pxf\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.058379 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-ovsdbserver-nb\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.058478 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-dns-swift-storage-0\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.058502 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-config\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.058553 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-ovsdbserver-sb\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.058575 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-dns-svc\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.058643 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-openstack-edpm-ipam\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.163038 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5pxf\" (UniqueName: \"kubernetes.io/projected/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-kube-api-access-v5pxf\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.163122 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-ovsdbserver-nb\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.163247 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-dns-swift-storage-0\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.163282 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-config\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.163338 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-ovsdbserver-sb\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.163371 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-dns-svc\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.163424 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-openstack-edpm-ipam\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.164641 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-config\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.164798 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-openstack-edpm-ipam\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.164898 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-ovsdbserver-nb\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.165204 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-ovsdbserver-sb\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.165285 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-dns-svc\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.165682 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-dns-swift-storage-0\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.211075 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5pxf\" (UniqueName: \"kubernetes.io/projected/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-kube-api-access-v5pxf\") pod \"dnsmasq-dns-57b9bf8b5-98n78\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.316509 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:06 crc kubenswrapper[4632]: I0313 10:31:06.755034 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e","Type":"ContainerStarted","Data":"14c062fec112b9ff9dd032399f07c596e52d395c51c6e56a8d5ba1fc6a94ca9a"} Mar 13 10:31:07 crc kubenswrapper[4632]: I0313 10:31:07.029746 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57b9bf8b5-98n78"] Mar 13 10:31:07 crc kubenswrapper[4632]: W0313 10:31:07.039843 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c3ff36b_8bef_403f_b6ac_ec88e26e924f.slice/crio-11f7381d58b27be1d13563ddb1c5a150b98cb261eb7c5049caee80961374ea23 WatchSource:0}: Error finding container 11f7381d58b27be1d13563ddb1c5a150b98cb261eb7c5049caee80961374ea23: Status 404 returned error can't find the container with id 11f7381d58b27be1d13563ddb1c5a150b98cb261eb7c5049caee80961374ea23 Mar 13 10:31:07 crc kubenswrapper[4632]: I0313 10:31:07.778924 4632 generic.go:334] "Generic (PLEG): container finished" podID="7c3ff36b-8bef-403f-b6ac-ec88e26e924f" containerID="90dfbecc999c31c0a51b0624874627a8f3c0659cb11e205820b8e9aab659a4a1" exitCode=0 Mar 13 10:31:07 crc kubenswrapper[4632]: I0313 10:31:07.782698 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" event={"ID":"7c3ff36b-8bef-403f-b6ac-ec88e26e924f","Type":"ContainerDied","Data":"90dfbecc999c31c0a51b0624874627a8f3c0659cb11e205820b8e9aab659a4a1"} Mar 13 10:31:07 crc kubenswrapper[4632]: I0313 10:31:07.782878 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" event={"ID":"7c3ff36b-8bef-403f-b6ac-ec88e26e924f","Type":"ContainerStarted","Data":"11f7381d58b27be1d13563ddb1c5a150b98cb261eb7c5049caee80961374ea23"} Mar 13 10:31:08 crc kubenswrapper[4632]: I0313 10:31:08.791598 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" event={"ID":"7c3ff36b-8bef-403f-b6ac-ec88e26e924f","Type":"ContainerStarted","Data":"1e8d2b5aecd08236cabb2c50425d69df7147e32b58dae758550f96994f27f434"} Mar 13 10:31:08 crc kubenswrapper[4632]: I0313 10:31:08.792660 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:08 crc kubenswrapper[4632]: I0313 10:31:08.831636 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" podStartSLOduration=3.8316164820000003 podStartE2EDuration="3.831616482s" podCreationTimestamp="2026-03-13 10:31:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:31:08.817712787 +0000 UTC m=+1642.840242920" watchObservedRunningTime="2026-03-13 10:31:08.831616482 +0000 UTC m=+1642.854146615" Mar 13 10:31:10 crc kubenswrapper[4632]: I0313 10:31:10.461317 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:31:10 crc kubenswrapper[4632]: I0313 10:31:10.461655 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:31:10 crc kubenswrapper[4632]: I0313 10:31:10.461716 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:31:10 crc kubenswrapper[4632]: I0313 10:31:10.462423 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:31:10 crc kubenswrapper[4632]: I0313 10:31:10.462489 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" gracePeriod=600 Mar 13 10:31:10 crc kubenswrapper[4632]: E0313 10:31:10.590546 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:31:10 crc kubenswrapper[4632]: I0313 10:31:10.835412 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" exitCode=0 Mar 13 10:31:10 crc kubenswrapper[4632]: I0313 10:31:10.836816 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f"} Mar 13 10:31:10 crc kubenswrapper[4632]: I0313 10:31:10.836965 4632 scope.go:117] "RemoveContainer" containerID="a148dfa9ef48de458189e9fda19ce88937bedd25c3ec76e22d14f43a4745805f" Mar 13 10:31:10 crc kubenswrapper[4632]: I0313 10:31:10.837612 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:31:10 crc kubenswrapper[4632]: E0313 10:31:10.838093 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:31:12 crc kubenswrapper[4632]: I0313 10:31:12.916316 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.318104 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.390426 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-564797cccc-84dg2"] Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.390853 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-564797cccc-84dg2" podUID="ac568760-fbe3-49ca-af4a-13f7780a1ad2" containerName="dnsmasq-dns" containerID="cri-o://3807149ca5beac08d142f3e5ffa3b80f5bf9a97b93a119f317229b5a8536c4a3" gracePeriod=10 Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.620239 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7b457785b5-7hzp6"] Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.621775 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.643698 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b457785b5-7hzp6"] Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.803719 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-openstack-edpm-ipam\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.804101 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-config\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.804146 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-dns-svc\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.804189 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-dns-swift-storage-0\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.804215 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-ovsdbserver-sb\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.804272 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-ovsdbserver-nb\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.804326 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6km2\" (UniqueName: \"kubernetes.io/projected/1aca78bb-c923-4964-9b4c-5f7fb50badba-kube-api-access-s6km2\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.898645 4632 generic.go:334] "Generic (PLEG): container finished" podID="ac568760-fbe3-49ca-af4a-13f7780a1ad2" containerID="3807149ca5beac08d142f3e5ffa3b80f5bf9a97b93a119f317229b5a8536c4a3" exitCode=0 Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.898697 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-564797cccc-84dg2" event={"ID":"ac568760-fbe3-49ca-af4a-13f7780a1ad2","Type":"ContainerDied","Data":"3807149ca5beac08d142f3e5ffa3b80f5bf9a97b93a119f317229b5a8536c4a3"} Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.906584 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6km2\" (UniqueName: \"kubernetes.io/projected/1aca78bb-c923-4964-9b4c-5f7fb50badba-kube-api-access-s6km2\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.906647 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-openstack-edpm-ipam\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.906680 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-config\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.906716 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-dns-svc\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.906756 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-dns-swift-storage-0\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.906778 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-ovsdbserver-sb\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.906837 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-ovsdbserver-nb\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.910185 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-dns-svc\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.910705 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-ovsdbserver-nb\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.910781 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-ovsdbserver-sb\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.911058 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-dns-swift-storage-0\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.913522 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-openstack-edpm-ipam\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.922679 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1aca78bb-c923-4964-9b4c-5f7fb50badba-config\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:16 crc kubenswrapper[4632]: I0313 10:31:16.963377 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6km2\" (UniqueName: \"kubernetes.io/projected/1aca78bb-c923-4964-9b4c-5f7fb50badba-kube-api-access-s6km2\") pod \"dnsmasq-dns-7b457785b5-7hzp6\" (UID: \"1aca78bb-c923-4964-9b4c-5f7fb50badba\") " pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.079311 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.112362 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-dns-swift-storage-0\") pod \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.112424 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-ovsdbserver-nb\") pod \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.112460 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-ovsdbserver-sb\") pod \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.112506 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-dns-svc\") pod \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.202387 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ac568760-fbe3-49ca-af4a-13f7780a1ad2" (UID: "ac568760-fbe3-49ca-af4a-13f7780a1ad2"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.213797 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djmgg\" (UniqueName: \"kubernetes.io/projected/ac568760-fbe3-49ca-af4a-13f7780a1ad2-kube-api-access-djmgg\") pod \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.214147 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-config\") pod \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\" (UID: \"ac568760-fbe3-49ca-af4a-13f7780a1ad2\") " Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.214581 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.217054 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ac568760-fbe3-49ca-af4a-13f7780a1ad2" (UID: "ac568760-fbe3-49ca-af4a-13f7780a1ad2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.227335 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ac568760-fbe3-49ca-af4a-13f7780a1ad2" (UID: "ac568760-fbe3-49ca-af4a-13f7780a1ad2"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.227610 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ac568760-fbe3-49ca-af4a-13f7780a1ad2" (UID: "ac568760-fbe3-49ca-af4a-13f7780a1ad2"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.228371 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac568760-fbe3-49ca-af4a-13f7780a1ad2-kube-api-access-djmgg" (OuterVolumeSpecName: "kube-api-access-djmgg") pod "ac568760-fbe3-49ca-af4a-13f7780a1ad2" (UID: "ac568760-fbe3-49ca-af4a-13f7780a1ad2"). InnerVolumeSpecName "kube-api-access-djmgg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.242420 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.270184 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-config" (OuterVolumeSpecName: "config") pod "ac568760-fbe3-49ca-af4a-13f7780a1ad2" (UID: "ac568760-fbe3-49ca-af4a-13f7780a1ad2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.315789 4632 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.315835 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.315849 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.315863 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac568760-fbe3-49ca-af4a-13f7780a1ad2-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.315874 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djmgg\" (UniqueName: \"kubernetes.io/projected/ac568760-fbe3-49ca-af4a-13f7780a1ad2-kube-api-access-djmgg\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.754016 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7b457785b5-7hzp6"] Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.914117 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" event={"ID":"1aca78bb-c923-4964-9b4c-5f7fb50badba","Type":"ContainerStarted","Data":"13b59ba479e5d6efa177855cf3835521bdc7ec51c03d2dc65872228d4234c924"} Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.917753 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-564797cccc-84dg2" event={"ID":"ac568760-fbe3-49ca-af4a-13f7780a1ad2","Type":"ContainerDied","Data":"f114a94f0fb42ce2c1f69bef8fad045098717f536b5431da20286872b08fed02"} Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.917807 4632 scope.go:117] "RemoveContainer" containerID="3807149ca5beac08d142f3e5ffa3b80f5bf9a97b93a119f317229b5a8536c4a3" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.917984 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-564797cccc-84dg2" Mar 13 10:31:17 crc kubenswrapper[4632]: I0313 10:31:17.971290 4632 scope.go:117] "RemoveContainer" containerID="7c9783dd40660c9e8665537c8ead9f633309987f7dedc616633d346075b3da86" Mar 13 10:31:18 crc kubenswrapper[4632]: I0313 10:31:18.024181 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-564797cccc-84dg2"] Mar 13 10:31:18 crc kubenswrapper[4632]: I0313 10:31:18.036452 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-564797cccc-84dg2"] Mar 13 10:31:18 crc kubenswrapper[4632]: I0313 10:31:18.091689 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac568760-fbe3-49ca-af4a-13f7780a1ad2" path="/var/lib/kubelet/pods/ac568760-fbe3-49ca-af4a-13f7780a1ad2/volumes" Mar 13 10:31:18 crc kubenswrapper[4632]: I0313 10:31:18.929518 4632 generic.go:334] "Generic (PLEG): container finished" podID="1aca78bb-c923-4964-9b4c-5f7fb50badba" containerID="cf9f6f4ace996138ffbc8df4b970fa3faa1ff37d0ed951c237b3de812effd2da" exitCode=0 Mar 13 10:31:18 crc kubenswrapper[4632]: I0313 10:31:18.929888 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" event={"ID":"1aca78bb-c923-4964-9b4c-5f7fb50badba","Type":"ContainerDied","Data":"cf9f6f4ace996138ffbc8df4b970fa3faa1ff37d0ed951c237b3de812effd2da"} Mar 13 10:31:19 crc kubenswrapper[4632]: I0313 10:31:19.944991 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" event={"ID":"1aca78bb-c923-4964-9b4c-5f7fb50badba","Type":"ContainerStarted","Data":"71a2fc806710f00a6f21a003ef29d764436a754a876865e6540b7ecd4a886549"} Mar 13 10:31:19 crc kubenswrapper[4632]: I0313 10:31:19.945473 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:19 crc kubenswrapper[4632]: I0313 10:31:19.967443 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" podStartSLOduration=3.967422941 podStartE2EDuration="3.967422941s" podCreationTimestamp="2026-03-13 10:31:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:31:19.963459726 +0000 UTC m=+1653.985989859" watchObservedRunningTime="2026-03-13 10:31:19.967422941 +0000 UTC m=+1653.989953074" Mar 13 10:31:22 crc kubenswrapper[4632]: I0313 10:31:22.044594 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:31:22 crc kubenswrapper[4632]: E0313 10:31:22.045230 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:31:27 crc kubenswrapper[4632]: I0313 10:31:27.244124 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7b457785b5-7hzp6" Mar 13 10:31:27 crc kubenswrapper[4632]: I0313 10:31:27.348587 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57b9bf8b5-98n78"] Mar 13 10:31:27 crc kubenswrapper[4632]: I0313 10:31:27.348880 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" podUID="7c3ff36b-8bef-403f-b6ac-ec88e26e924f" containerName="dnsmasq-dns" containerID="cri-o://1e8d2b5aecd08236cabb2c50425d69df7147e32b58dae758550f96994f27f434" gracePeriod=10 Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.026830 4632 generic.go:334] "Generic (PLEG): container finished" podID="7c3ff36b-8bef-403f-b6ac-ec88e26e924f" containerID="1e8d2b5aecd08236cabb2c50425d69df7147e32b58dae758550f96994f27f434" exitCode=0 Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.027445 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" event={"ID":"7c3ff36b-8bef-403f-b6ac-ec88e26e924f","Type":"ContainerDied","Data":"1e8d2b5aecd08236cabb2c50425d69df7147e32b58dae758550f96994f27f434"} Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.027474 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" event={"ID":"7c3ff36b-8bef-403f-b6ac-ec88e26e924f","Type":"ContainerDied","Data":"11f7381d58b27be1d13563ddb1c5a150b98cb261eb7c5049caee80961374ea23"} Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.027488 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11f7381d58b27be1d13563ddb1c5a150b98cb261eb7c5049caee80961374ea23" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.042691 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.061630 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-dns-svc\") pod \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.061688 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-ovsdbserver-sb\") pod \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.061736 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5pxf\" (UniqueName: \"kubernetes.io/projected/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-kube-api-access-v5pxf\") pod \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.061771 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-openstack-edpm-ipam\") pod \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.061793 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-config\") pod \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.061822 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-ovsdbserver-nb\") pod \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.061863 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-dns-swift-storage-0\") pod \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\" (UID: \"7c3ff36b-8bef-403f-b6ac-ec88e26e924f\") " Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.100695 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-kube-api-access-v5pxf" (OuterVolumeSpecName: "kube-api-access-v5pxf") pod "7c3ff36b-8bef-403f-b6ac-ec88e26e924f" (UID: "7c3ff36b-8bef-403f-b6ac-ec88e26e924f"). InnerVolumeSpecName "kube-api-access-v5pxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.165160 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5pxf\" (UniqueName: \"kubernetes.io/projected/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-kube-api-access-v5pxf\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.217576 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7c3ff36b-8bef-403f-b6ac-ec88e26e924f" (UID: "7c3ff36b-8bef-403f-b6ac-ec88e26e924f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.242973 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7c3ff36b-8bef-403f-b6ac-ec88e26e924f" (UID: "7c3ff36b-8bef-403f-b6ac-ec88e26e924f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.243082 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7c3ff36b-8bef-403f-b6ac-ec88e26e924f" (UID: "7c3ff36b-8bef-403f-b6ac-ec88e26e924f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.245683 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "7c3ff36b-8bef-403f-b6ac-ec88e26e924f" (UID: "7c3ff36b-8bef-403f-b6ac-ec88e26e924f"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.266680 4632 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.266725 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.266740 4632 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.266753 4632 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.271409 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7c3ff36b-8bef-403f-b6ac-ec88e26e924f" (UID: "7c3ff36b-8bef-403f-b6ac-ec88e26e924f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.274506 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-config" (OuterVolumeSpecName: "config") pod "7c3ff36b-8bef-403f-b6ac-ec88e26e924f" (UID: "7c3ff36b-8bef-403f-b6ac-ec88e26e924f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.369166 4632 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:28 crc kubenswrapper[4632]: I0313 10:31:28.369211 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7c3ff36b-8bef-403f-b6ac-ec88e26e924f-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:29 crc kubenswrapper[4632]: I0313 10:31:29.038085 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57b9bf8b5-98n78" Mar 13 10:31:29 crc kubenswrapper[4632]: I0313 10:31:29.084739 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57b9bf8b5-98n78"] Mar 13 10:31:29 crc kubenswrapper[4632]: I0313 10:31:29.098444 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57b9bf8b5-98n78"] Mar 13 10:31:30 crc kubenswrapper[4632]: I0313 10:31:30.104352 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c3ff36b-8bef-403f-b6ac-ec88e26e924f" path="/var/lib/kubelet/pods/7c3ff36b-8bef-403f-b6ac-ec88e26e924f/volumes" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.025837 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gjdvf"] Mar 13 10:31:34 crc kubenswrapper[4632]: E0313 10:31:34.026469 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3ff36b-8bef-403f-b6ac-ec88e26e924f" containerName="dnsmasq-dns" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.026481 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3ff36b-8bef-403f-b6ac-ec88e26e924f" containerName="dnsmasq-dns" Mar 13 10:31:34 crc kubenswrapper[4632]: E0313 10:31:34.026495 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac568760-fbe3-49ca-af4a-13f7780a1ad2" containerName="init" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.026501 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac568760-fbe3-49ca-af4a-13f7780a1ad2" containerName="init" Mar 13 10:31:34 crc kubenswrapper[4632]: E0313 10:31:34.026515 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3ff36b-8bef-403f-b6ac-ec88e26e924f" containerName="init" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.026521 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3ff36b-8bef-403f-b6ac-ec88e26e924f" containerName="init" Mar 13 10:31:34 crc kubenswrapper[4632]: E0313 10:31:34.026534 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac568760-fbe3-49ca-af4a-13f7780a1ad2" containerName="dnsmasq-dns" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.026540 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac568760-fbe3-49ca-af4a-13f7780a1ad2" containerName="dnsmasq-dns" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.026712 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c3ff36b-8bef-403f-b6ac-ec88e26e924f" containerName="dnsmasq-dns" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.026731 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac568760-fbe3-49ca-af4a-13f7780a1ad2" containerName="dnsmasq-dns" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.028042 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.042059 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gjdvf"] Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.183961 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxklg\" (UniqueName: \"kubernetes.io/projected/03215c5e-aa7f-4865-8e14-7adb79cc6daa-kube-api-access-zxklg\") pod \"community-operators-gjdvf\" (UID: \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\") " pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.184078 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03215c5e-aa7f-4865-8e14-7adb79cc6daa-catalog-content\") pod \"community-operators-gjdvf\" (UID: \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\") " pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.184143 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03215c5e-aa7f-4865-8e14-7adb79cc6daa-utilities\") pod \"community-operators-gjdvf\" (UID: \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\") " pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.285855 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03215c5e-aa7f-4865-8e14-7adb79cc6daa-catalog-content\") pod \"community-operators-gjdvf\" (UID: \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\") " pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.286246 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03215c5e-aa7f-4865-8e14-7adb79cc6daa-utilities\") pod \"community-operators-gjdvf\" (UID: \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\") " pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.286328 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxklg\" (UniqueName: \"kubernetes.io/projected/03215c5e-aa7f-4865-8e14-7adb79cc6daa-kube-api-access-zxklg\") pod \"community-operators-gjdvf\" (UID: \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\") " pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.286738 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03215c5e-aa7f-4865-8e14-7adb79cc6daa-catalog-content\") pod \"community-operators-gjdvf\" (UID: \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\") " pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.287140 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03215c5e-aa7f-4865-8e14-7adb79cc6daa-utilities\") pod \"community-operators-gjdvf\" (UID: \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\") " pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.304906 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxklg\" (UniqueName: \"kubernetes.io/projected/03215c5e-aa7f-4865-8e14-7adb79cc6daa-kube-api-access-zxklg\") pod \"community-operators-gjdvf\" (UID: \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\") " pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.403874 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:31:34 crc kubenswrapper[4632]: I0313 10:31:34.935845 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gjdvf"] Mar 13 10:31:35 crc kubenswrapper[4632]: I0313 10:31:35.126693 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gjdvf" event={"ID":"03215c5e-aa7f-4865-8e14-7adb79cc6daa","Type":"ContainerStarted","Data":"49ad0dc835d9a21740d64cc609bbe889c1787b45abac10c16f9550eca410cb5e"} Mar 13 10:31:35 crc kubenswrapper[4632]: I0313 10:31:35.127025 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gjdvf" event={"ID":"03215c5e-aa7f-4865-8e14-7adb79cc6daa","Type":"ContainerStarted","Data":"5bfc882deb40abae4190d943339bb6bc98f0f88a95b8725fe0fdc29c76d7cf9f"} Mar 13 10:31:36 crc kubenswrapper[4632]: I0313 10:31:36.142728 4632 generic.go:334] "Generic (PLEG): container finished" podID="03215c5e-aa7f-4865-8e14-7adb79cc6daa" containerID="49ad0dc835d9a21740d64cc609bbe889c1787b45abac10c16f9550eca410cb5e" exitCode=0 Mar 13 10:31:36 crc kubenswrapper[4632]: I0313 10:31:36.142791 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gjdvf" event={"ID":"03215c5e-aa7f-4865-8e14-7adb79cc6daa","Type":"ContainerDied","Data":"49ad0dc835d9a21740d64cc609bbe889c1787b45abac10c16f9550eca410cb5e"} Mar 13 10:31:37 crc kubenswrapper[4632]: I0313 10:31:37.044775 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:31:37 crc kubenswrapper[4632]: E0313 10:31:37.046009 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:31:37 crc kubenswrapper[4632]: I0313 10:31:37.152409 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gjdvf" event={"ID":"03215c5e-aa7f-4865-8e14-7adb79cc6daa","Type":"ContainerStarted","Data":"81f430399eabd7810f83b975907d2a562229273cc3df439ab9b33cbfde5ddc35"} Mar 13 10:31:38 crc kubenswrapper[4632]: I0313 10:31:38.164103 4632 generic.go:334] "Generic (PLEG): container finished" podID="c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e" containerID="f69527ec2be6e4df808bf875b701d62b874066c28a489a38ef573bdae2b131dc" exitCode=0 Mar 13 10:31:38 crc kubenswrapper[4632]: I0313 10:31:38.164158 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e","Type":"ContainerDied","Data":"f69527ec2be6e4df808bf875b701d62b874066c28a489a38ef573bdae2b131dc"} Mar 13 10:31:39 crc kubenswrapper[4632]: I0313 10:31:39.178322 4632 generic.go:334] "Generic (PLEG): container finished" podID="03215c5e-aa7f-4865-8e14-7adb79cc6daa" containerID="81f430399eabd7810f83b975907d2a562229273cc3df439ab9b33cbfde5ddc35" exitCode=0 Mar 13 10:31:39 crc kubenswrapper[4632]: I0313 10:31:39.178490 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gjdvf" event={"ID":"03215c5e-aa7f-4865-8e14-7adb79cc6daa","Type":"ContainerDied","Data":"81f430399eabd7810f83b975907d2a562229273cc3df439ab9b33cbfde5ddc35"} Mar 13 10:31:39 crc kubenswrapper[4632]: I0313 10:31:39.182985 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e","Type":"ContainerStarted","Data":"6477bb63f9230d4eb7ed71b9808b2a7459fbb37920ec3edee2c2e8ce3382ade9"} Mar 13 10:31:39 crc kubenswrapper[4632]: I0313 10:31:39.183203 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 13 10:31:39 crc kubenswrapper[4632]: I0313 10:31:39.189065 4632 generic.go:334] "Generic (PLEG): container finished" podID="a3d80d9f-c956-40f5-b2e1-8aea2f136b6e" containerID="14c062fec112b9ff9dd032399f07c596e52d395c51c6e56a8d5ba1fc6a94ca9a" exitCode=0 Mar 13 10:31:39 crc kubenswrapper[4632]: I0313 10:31:39.189119 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e","Type":"ContainerDied","Data":"14c062fec112b9ff9dd032399f07c596e52d395c51c6e56a8d5ba1fc6a94ca9a"} Mar 13 10:31:39 crc kubenswrapper[4632]: I0313 10:31:39.336717 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=38.336693404 podStartE2EDuration="38.336693404s" podCreationTimestamp="2026-03-13 10:31:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:31:39.296024304 +0000 UTC m=+1673.318554437" watchObservedRunningTime="2026-03-13 10:31:39.336693404 +0000 UTC m=+1673.359223547" Mar 13 10:31:40 crc kubenswrapper[4632]: I0313 10:31:40.200615 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gjdvf" event={"ID":"03215c5e-aa7f-4865-8e14-7adb79cc6daa","Type":"ContainerStarted","Data":"f3faae6eab76531261d4e9489936f30a1f2265caad99779ef7b44da581399110"} Mar 13 10:31:40 crc kubenswrapper[4632]: I0313 10:31:40.204010 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"a3d80d9f-c956-40f5-b2e1-8aea2f136b6e","Type":"ContainerStarted","Data":"4acfb0604b12245729351effb8cd4a294fb3df90c4eb8d0f98d89deda4f3f3dc"} Mar 13 10:31:40 crc kubenswrapper[4632]: I0313 10:31:40.236923 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gjdvf" podStartSLOduration=2.57190617 podStartE2EDuration="6.236901406s" podCreationTimestamp="2026-03-13 10:31:34 +0000 UTC" firstStartedPulling="2026-03-13 10:31:36.144628434 +0000 UTC m=+1670.167158567" lastFinishedPulling="2026-03-13 10:31:39.80962367 +0000 UTC m=+1673.832153803" observedRunningTime="2026-03-13 10:31:40.23041819 +0000 UTC m=+1674.252948333" watchObservedRunningTime="2026-03-13 10:31:40.236901406 +0000 UTC m=+1674.259431529" Mar 13 10:31:40 crc kubenswrapper[4632]: I0313 10:31:40.260456 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=38.260436413 podStartE2EDuration="38.260436413s" podCreationTimestamp="2026-03-13 10:31:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:31:40.256953449 +0000 UTC m=+1674.279483592" watchObservedRunningTime="2026-03-13 10:31:40.260436413 +0000 UTC m=+1674.282966546" Mar 13 10:31:43 crc kubenswrapper[4632]: I0313 10:31:43.097238 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.261579 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9"] Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.263105 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.267401 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.268345 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.268405 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.269707 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.283588 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9"] Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.369531 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.369734 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.370066 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.370178 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7vn8\" (UniqueName: \"kubernetes.io/projected/0ea59acf-3206-492e-a7a8-bf855823d92c-kube-api-access-w7vn8\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.406038 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.406085 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.472473 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.472583 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7vn8\" (UniqueName: \"kubernetes.io/projected/0ea59acf-3206-492e-a7a8-bf855823d92c-kube-api-access-w7vn8\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.472636 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.472849 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.478264 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.483215 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.488614 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.537841 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7vn8\" (UniqueName: \"kubernetes.io/projected/0ea59acf-3206-492e-a7a8-bf855823d92c-kube-api-access-w7vn8\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:31:44 crc kubenswrapper[4632]: I0313 10:31:44.588427 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:31:45 crc kubenswrapper[4632]: I0313 10:31:45.308045 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9"] Mar 13 10:31:45 crc kubenswrapper[4632]: I0313 10:31:45.454576 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-gjdvf" podUID="03215c5e-aa7f-4865-8e14-7adb79cc6daa" containerName="registry-server" probeResult="failure" output=< Mar 13 10:31:45 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:31:45 crc kubenswrapper[4632]: > Mar 13 10:31:46 crc kubenswrapper[4632]: I0313 10:31:46.280473 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" event={"ID":"0ea59acf-3206-492e-a7a8-bf855823d92c","Type":"ContainerStarted","Data":"3b44df298dbf44fed38d413586ce5557104e682a2fdc9a900e9bbae85c8951c4"} Mar 13 10:31:50 crc kubenswrapper[4632]: I0313 10:31:50.045694 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:31:50 crc kubenswrapper[4632]: E0313 10:31:50.046693 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:31:52 crc kubenswrapper[4632]: I0313 10:31:52.221169 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 13 10:31:53 crc kubenswrapper[4632]: I0313 10:31:53.101775 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 13 10:31:55 crc kubenswrapper[4632]: I0313 10:31:55.464041 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-gjdvf" podUID="03215c5e-aa7f-4865-8e14-7adb79cc6daa" containerName="registry-server" probeResult="failure" output=< Mar 13 10:31:55 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:31:55 crc kubenswrapper[4632]: > Mar 13 10:31:58 crc kubenswrapper[4632]: I0313 10:31:58.849394 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:31:59 crc kubenswrapper[4632]: I0313 10:31:59.523620 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" event={"ID":"0ea59acf-3206-492e-a7a8-bf855823d92c","Type":"ContainerStarted","Data":"2a45f7f396ff9ac0f8fe934eb95d769681cb829e23350a9d92e18b2aeedef144"} Mar 13 10:31:59 crc kubenswrapper[4632]: I0313 10:31:59.558192 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" podStartSLOduration=2.029747859 podStartE2EDuration="15.558164163s" podCreationTimestamp="2026-03-13 10:31:44 +0000 UTC" firstStartedPulling="2026-03-13 10:31:45.318732775 +0000 UTC m=+1679.341262908" lastFinishedPulling="2026-03-13 10:31:58.847149069 +0000 UTC m=+1692.869679212" observedRunningTime="2026-03-13 10:31:59.544181336 +0000 UTC m=+1693.566711479" watchObservedRunningTime="2026-03-13 10:31:59.558164163 +0000 UTC m=+1693.580694296" Mar 13 10:32:00 crc kubenswrapper[4632]: I0313 10:32:00.155966 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556632-sr4l5"] Mar 13 10:32:00 crc kubenswrapper[4632]: I0313 10:32:00.157835 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556632-sr4l5" Mar 13 10:32:00 crc kubenswrapper[4632]: I0313 10:32:00.160487 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:32:00 crc kubenswrapper[4632]: I0313 10:32:00.160926 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:32:00 crc kubenswrapper[4632]: I0313 10:32:00.161227 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:32:00 crc kubenswrapper[4632]: I0313 10:32:00.168165 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556632-sr4l5"] Mar 13 10:32:00 crc kubenswrapper[4632]: I0313 10:32:00.209993 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7q2t\" (UniqueName: \"kubernetes.io/projected/009f055c-d442-4b23-8f55-52a43362bbb2-kube-api-access-w7q2t\") pod \"auto-csr-approver-29556632-sr4l5\" (UID: \"009f055c-d442-4b23-8f55-52a43362bbb2\") " pod="openshift-infra/auto-csr-approver-29556632-sr4l5" Mar 13 10:32:00 crc kubenswrapper[4632]: I0313 10:32:00.312861 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7q2t\" (UniqueName: \"kubernetes.io/projected/009f055c-d442-4b23-8f55-52a43362bbb2-kube-api-access-w7q2t\") pod \"auto-csr-approver-29556632-sr4l5\" (UID: \"009f055c-d442-4b23-8f55-52a43362bbb2\") " pod="openshift-infra/auto-csr-approver-29556632-sr4l5" Mar 13 10:32:00 crc kubenswrapper[4632]: I0313 10:32:00.338788 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7q2t\" (UniqueName: \"kubernetes.io/projected/009f055c-d442-4b23-8f55-52a43362bbb2-kube-api-access-w7q2t\") pod \"auto-csr-approver-29556632-sr4l5\" (UID: \"009f055c-d442-4b23-8f55-52a43362bbb2\") " pod="openshift-infra/auto-csr-approver-29556632-sr4l5" Mar 13 10:32:00 crc kubenswrapper[4632]: I0313 10:32:00.482604 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556632-sr4l5" Mar 13 10:32:01 crc kubenswrapper[4632]: I0313 10:32:01.474135 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556632-sr4l5"] Mar 13 10:32:01 crc kubenswrapper[4632]: W0313 10:32:01.486072 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod009f055c_d442_4b23_8f55_52a43362bbb2.slice/crio-98b91929707278b3c3c358e0d943597f90b34c6e4534e5a39b399c489440ea9a WatchSource:0}: Error finding container 98b91929707278b3c3c358e0d943597f90b34c6e4534e5a39b399c489440ea9a: Status 404 returned error can't find the container with id 98b91929707278b3c3c358e0d943597f90b34c6e4534e5a39b399c489440ea9a Mar 13 10:32:01 crc kubenswrapper[4632]: I0313 10:32:01.543780 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556632-sr4l5" event={"ID":"009f055c-d442-4b23-8f55-52a43362bbb2","Type":"ContainerStarted","Data":"98b91929707278b3c3c358e0d943597f90b34c6e4534e5a39b399c489440ea9a"} Mar 13 10:32:03 crc kubenswrapper[4632]: I0313 10:32:03.044363 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:32:03 crc kubenswrapper[4632]: E0313 10:32:03.045011 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:32:03 crc kubenswrapper[4632]: I0313 10:32:03.564170 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556632-sr4l5" event={"ID":"009f055c-d442-4b23-8f55-52a43362bbb2","Type":"ContainerStarted","Data":"4ccfb76824c418f1c761a43ca7732c6a7a69b7b1944ea2ee35bd45c569e7d7c6"} Mar 13 10:32:03 crc kubenswrapper[4632]: I0313 10:32:03.583029 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556632-sr4l5" podStartSLOduration=2.408523498 podStartE2EDuration="3.583011419s" podCreationTimestamp="2026-03-13 10:32:00 +0000 UTC" firstStartedPulling="2026-03-13 10:32:01.489179825 +0000 UTC m=+1695.511709958" lastFinishedPulling="2026-03-13 10:32:02.663667746 +0000 UTC m=+1696.686197879" observedRunningTime="2026-03-13 10:32:03.580284133 +0000 UTC m=+1697.602814276" watchObservedRunningTime="2026-03-13 10:32:03.583011419 +0000 UTC m=+1697.605541542" Mar 13 10:32:04 crc kubenswrapper[4632]: I0313 10:32:04.466307 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:32:04 crc kubenswrapper[4632]: I0313 10:32:04.518772 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:32:04 crc kubenswrapper[4632]: I0313 10:32:04.577039 4632 generic.go:334] "Generic (PLEG): container finished" podID="009f055c-d442-4b23-8f55-52a43362bbb2" containerID="4ccfb76824c418f1c761a43ca7732c6a7a69b7b1944ea2ee35bd45c569e7d7c6" exitCode=0 Mar 13 10:32:04 crc kubenswrapper[4632]: I0313 10:32:04.577098 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556632-sr4l5" event={"ID":"009f055c-d442-4b23-8f55-52a43362bbb2","Type":"ContainerDied","Data":"4ccfb76824c418f1c761a43ca7732c6a7a69b7b1944ea2ee35bd45c569e7d7c6"} Mar 13 10:32:05 crc kubenswrapper[4632]: I0313 10:32:05.230361 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gjdvf"] Mar 13 10:32:05 crc kubenswrapper[4632]: I0313 10:32:05.586605 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gjdvf" podUID="03215c5e-aa7f-4865-8e14-7adb79cc6daa" containerName="registry-server" containerID="cri-o://f3faae6eab76531261d4e9489936f30a1f2265caad99779ef7b44da581399110" gracePeriod=2 Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.173004 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556632-sr4l5" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.180424 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.236482 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7q2t\" (UniqueName: \"kubernetes.io/projected/009f055c-d442-4b23-8f55-52a43362bbb2-kube-api-access-w7q2t\") pod \"009f055c-d442-4b23-8f55-52a43362bbb2\" (UID: \"009f055c-d442-4b23-8f55-52a43362bbb2\") " Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.236630 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03215c5e-aa7f-4865-8e14-7adb79cc6daa-utilities\") pod \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\" (UID: \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\") " Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.236731 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxklg\" (UniqueName: \"kubernetes.io/projected/03215c5e-aa7f-4865-8e14-7adb79cc6daa-kube-api-access-zxklg\") pod \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\" (UID: \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\") " Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.236919 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03215c5e-aa7f-4865-8e14-7adb79cc6daa-catalog-content\") pod \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\" (UID: \"03215c5e-aa7f-4865-8e14-7adb79cc6daa\") " Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.237911 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03215c5e-aa7f-4865-8e14-7adb79cc6daa-utilities" (OuterVolumeSpecName: "utilities") pod "03215c5e-aa7f-4865-8e14-7adb79cc6daa" (UID: "03215c5e-aa7f-4865-8e14-7adb79cc6daa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.242810 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03215c5e-aa7f-4865-8e14-7adb79cc6daa-kube-api-access-zxklg" (OuterVolumeSpecName: "kube-api-access-zxklg") pod "03215c5e-aa7f-4865-8e14-7adb79cc6daa" (UID: "03215c5e-aa7f-4865-8e14-7adb79cc6daa"). InnerVolumeSpecName "kube-api-access-zxklg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.242872 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/009f055c-d442-4b23-8f55-52a43362bbb2-kube-api-access-w7q2t" (OuterVolumeSpecName: "kube-api-access-w7q2t") pod "009f055c-d442-4b23-8f55-52a43362bbb2" (UID: "009f055c-d442-4b23-8f55-52a43362bbb2"). InnerVolumeSpecName "kube-api-access-w7q2t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.308252 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/03215c5e-aa7f-4865-8e14-7adb79cc6daa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "03215c5e-aa7f-4865-8e14-7adb79cc6daa" (UID: "03215c5e-aa7f-4865-8e14-7adb79cc6daa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.338866 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/03215c5e-aa7f-4865-8e14-7adb79cc6daa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.338899 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7q2t\" (UniqueName: \"kubernetes.io/projected/009f055c-d442-4b23-8f55-52a43362bbb2-kube-api-access-w7q2t\") on node \"crc\" DevicePath \"\"" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.338910 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/03215c5e-aa7f-4865-8e14-7adb79cc6daa-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.338919 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxklg\" (UniqueName: \"kubernetes.io/projected/03215c5e-aa7f-4865-8e14-7adb79cc6daa-kube-api-access-zxklg\") on node \"crc\" DevicePath \"\"" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.592520 4632 scope.go:117] "RemoveContainer" containerID="19adb417107921a77df964ab1bd8c8cf0029e40afcac705a66952307655b68b9" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.598273 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556632-sr4l5" event={"ID":"009f055c-d442-4b23-8f55-52a43362bbb2","Type":"ContainerDied","Data":"98b91929707278b3c3c358e0d943597f90b34c6e4534e5a39b399c489440ea9a"} Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.598315 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98b91929707278b3c3c358e0d943597f90b34c6e4534e5a39b399c489440ea9a" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.598529 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556632-sr4l5" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.613242 4632 generic.go:334] "Generic (PLEG): container finished" podID="03215c5e-aa7f-4865-8e14-7adb79cc6daa" containerID="f3faae6eab76531261d4e9489936f30a1f2265caad99779ef7b44da581399110" exitCode=0 Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.613299 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gjdvf" event={"ID":"03215c5e-aa7f-4865-8e14-7adb79cc6daa","Type":"ContainerDied","Data":"f3faae6eab76531261d4e9489936f30a1f2265caad99779ef7b44da581399110"} Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.613325 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gjdvf" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.613341 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gjdvf" event={"ID":"03215c5e-aa7f-4865-8e14-7adb79cc6daa","Type":"ContainerDied","Data":"5bfc882deb40abae4190d943339bb6bc98f0f88a95b8725fe0fdc29c76d7cf9f"} Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.613364 4632 scope.go:117] "RemoveContainer" containerID="f3faae6eab76531261d4e9489936f30a1f2265caad99779ef7b44da581399110" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.660910 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556626-z45rd"] Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.666863 4632 scope.go:117] "RemoveContainer" containerID="81f430399eabd7810f83b975907d2a562229273cc3df439ab9b33cbfde5ddc35" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.674257 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556626-z45rd"] Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.686429 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gjdvf"] Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.700899 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gjdvf"] Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.706759 4632 scope.go:117] "RemoveContainer" containerID="49ad0dc835d9a21740d64cc609bbe889c1787b45abac10c16f9550eca410cb5e" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.762790 4632 scope.go:117] "RemoveContainer" containerID="f3faae6eab76531261d4e9489936f30a1f2265caad99779ef7b44da581399110" Mar 13 10:32:06 crc kubenswrapper[4632]: E0313 10:32:06.763583 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3faae6eab76531261d4e9489936f30a1f2265caad99779ef7b44da581399110\": container with ID starting with f3faae6eab76531261d4e9489936f30a1f2265caad99779ef7b44da581399110 not found: ID does not exist" containerID="f3faae6eab76531261d4e9489936f30a1f2265caad99779ef7b44da581399110" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.763630 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3faae6eab76531261d4e9489936f30a1f2265caad99779ef7b44da581399110"} err="failed to get container status \"f3faae6eab76531261d4e9489936f30a1f2265caad99779ef7b44da581399110\": rpc error: code = NotFound desc = could not find container \"f3faae6eab76531261d4e9489936f30a1f2265caad99779ef7b44da581399110\": container with ID starting with f3faae6eab76531261d4e9489936f30a1f2265caad99779ef7b44da581399110 not found: ID does not exist" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.763651 4632 scope.go:117] "RemoveContainer" containerID="81f430399eabd7810f83b975907d2a562229273cc3df439ab9b33cbfde5ddc35" Mar 13 10:32:06 crc kubenswrapper[4632]: E0313 10:32:06.763980 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81f430399eabd7810f83b975907d2a562229273cc3df439ab9b33cbfde5ddc35\": container with ID starting with 81f430399eabd7810f83b975907d2a562229273cc3df439ab9b33cbfde5ddc35 not found: ID does not exist" containerID="81f430399eabd7810f83b975907d2a562229273cc3df439ab9b33cbfde5ddc35" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.764055 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81f430399eabd7810f83b975907d2a562229273cc3df439ab9b33cbfde5ddc35"} err="failed to get container status \"81f430399eabd7810f83b975907d2a562229273cc3df439ab9b33cbfde5ddc35\": rpc error: code = NotFound desc = could not find container \"81f430399eabd7810f83b975907d2a562229273cc3df439ab9b33cbfde5ddc35\": container with ID starting with 81f430399eabd7810f83b975907d2a562229273cc3df439ab9b33cbfde5ddc35 not found: ID does not exist" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.764119 4632 scope.go:117] "RemoveContainer" containerID="49ad0dc835d9a21740d64cc609bbe889c1787b45abac10c16f9550eca410cb5e" Mar 13 10:32:06 crc kubenswrapper[4632]: E0313 10:32:06.764381 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49ad0dc835d9a21740d64cc609bbe889c1787b45abac10c16f9550eca410cb5e\": container with ID starting with 49ad0dc835d9a21740d64cc609bbe889c1787b45abac10c16f9550eca410cb5e not found: ID does not exist" containerID="49ad0dc835d9a21740d64cc609bbe889c1787b45abac10c16f9550eca410cb5e" Mar 13 10:32:06 crc kubenswrapper[4632]: I0313 10:32:06.764420 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49ad0dc835d9a21740d64cc609bbe889c1787b45abac10c16f9550eca410cb5e"} err="failed to get container status \"49ad0dc835d9a21740d64cc609bbe889c1787b45abac10c16f9550eca410cb5e\": rpc error: code = NotFound desc = could not find container \"49ad0dc835d9a21740d64cc609bbe889c1787b45abac10c16f9550eca410cb5e\": container with ID starting with 49ad0dc835d9a21740d64cc609bbe889c1787b45abac10c16f9550eca410cb5e not found: ID does not exist" Mar 13 10:32:08 crc kubenswrapper[4632]: I0313 10:32:08.055168 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03215c5e-aa7f-4865-8e14-7adb79cc6daa" path="/var/lib/kubelet/pods/03215c5e-aa7f-4865-8e14-7adb79cc6daa/volumes" Mar 13 10:32:08 crc kubenswrapper[4632]: I0313 10:32:08.056551 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27" path="/var/lib/kubelet/pods/8d59f1c4-0990-423d-ab4f-ecf2d0a1ac27/volumes" Mar 13 10:32:12 crc kubenswrapper[4632]: I0313 10:32:12.668161 4632 generic.go:334] "Generic (PLEG): container finished" podID="0ea59acf-3206-492e-a7a8-bf855823d92c" containerID="2a45f7f396ff9ac0f8fe934eb95d769681cb829e23350a9d92e18b2aeedef144" exitCode=0 Mar 13 10:32:12 crc kubenswrapper[4632]: I0313 10:32:12.668311 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" event={"ID":"0ea59acf-3206-492e-a7a8-bf855823d92c","Type":"ContainerDied","Data":"2a45f7f396ff9ac0f8fe934eb95d769681cb829e23350a9d92e18b2aeedef144"} Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.043882 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:32:14 crc kubenswrapper[4632]: E0313 10:32:14.044689 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.355147 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.498481 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-repo-setup-combined-ca-bundle\") pod \"0ea59acf-3206-492e-a7a8-bf855823d92c\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.499740 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7vn8\" (UniqueName: \"kubernetes.io/projected/0ea59acf-3206-492e-a7a8-bf855823d92c-kube-api-access-w7vn8\") pod \"0ea59acf-3206-492e-a7a8-bf855823d92c\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.500336 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-inventory\") pod \"0ea59acf-3206-492e-a7a8-bf855823d92c\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.500585 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-ssh-key-openstack-edpm-ipam\") pod \"0ea59acf-3206-492e-a7a8-bf855823d92c\" (UID: \"0ea59acf-3206-492e-a7a8-bf855823d92c\") " Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.505873 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ea59acf-3206-492e-a7a8-bf855823d92c-kube-api-access-w7vn8" (OuterVolumeSpecName: "kube-api-access-w7vn8") pod "0ea59acf-3206-492e-a7a8-bf855823d92c" (UID: "0ea59acf-3206-492e-a7a8-bf855823d92c"). InnerVolumeSpecName "kube-api-access-w7vn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.511284 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "0ea59acf-3206-492e-a7a8-bf855823d92c" (UID: "0ea59acf-3206-492e-a7a8-bf855823d92c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.529638 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0ea59acf-3206-492e-a7a8-bf855823d92c" (UID: "0ea59acf-3206-492e-a7a8-bf855823d92c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.545146 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-inventory" (OuterVolumeSpecName: "inventory") pod "0ea59acf-3206-492e-a7a8-bf855823d92c" (UID: "0ea59acf-3206-492e-a7a8-bf855823d92c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.604343 4632 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.604394 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7vn8\" (UniqueName: \"kubernetes.io/projected/0ea59acf-3206-492e-a7a8-bf855823d92c-kube-api-access-w7vn8\") on node \"crc\" DevicePath \"\"" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.604409 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.604422 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0ea59acf-3206-492e-a7a8-bf855823d92c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.688811 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" event={"ID":"0ea59acf-3206-492e-a7a8-bf855823d92c","Type":"ContainerDied","Data":"3b44df298dbf44fed38d413586ce5557104e682a2fdc9a900e9bbae85c8951c4"} Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.688878 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b44df298dbf44fed38d413586ce5557104e682a2fdc9a900e9bbae85c8951c4" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.688836 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.792290 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h"] Mar 13 10:32:14 crc kubenswrapper[4632]: E0313 10:32:14.792741 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03215c5e-aa7f-4865-8e14-7adb79cc6daa" containerName="registry-server" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.792763 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="03215c5e-aa7f-4865-8e14-7adb79cc6daa" containerName="registry-server" Mar 13 10:32:14 crc kubenswrapper[4632]: E0313 10:32:14.792774 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="009f055c-d442-4b23-8f55-52a43362bbb2" containerName="oc" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.792781 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="009f055c-d442-4b23-8f55-52a43362bbb2" containerName="oc" Mar 13 10:32:14 crc kubenswrapper[4632]: E0313 10:32:14.792816 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ea59acf-3206-492e-a7a8-bf855823d92c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.795079 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ea59acf-3206-492e-a7a8-bf855823d92c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 13 10:32:14 crc kubenswrapper[4632]: E0313 10:32:14.795104 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03215c5e-aa7f-4865-8e14-7adb79cc6daa" containerName="extract-content" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.795113 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="03215c5e-aa7f-4865-8e14-7adb79cc6daa" containerName="extract-content" Mar 13 10:32:14 crc kubenswrapper[4632]: E0313 10:32:14.795127 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03215c5e-aa7f-4865-8e14-7adb79cc6daa" containerName="extract-utilities" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.795136 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="03215c5e-aa7f-4865-8e14-7adb79cc6daa" containerName="extract-utilities" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.795394 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ea59acf-3206-492e-a7a8-bf855823d92c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.795434 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="03215c5e-aa7f-4865-8e14-7adb79cc6daa" containerName="registry-server" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.795447 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="009f055c-d442-4b23-8f55-52a43362bbb2" containerName="oc" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.796207 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.798824 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.799015 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.799382 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.799502 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.847393 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h"] Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.908974 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gncwc\" (UniqueName: \"kubernetes.io/projected/1dc9191f-32b9-45b9-b49f-fd704075f0a5-kube-api-access-gncwc\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5s64h\" (UID: \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.909047 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc9191f-32b9-45b9-b49f-fd704075f0a5-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5s64h\" (UID: \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" Mar 13 10:32:14 crc kubenswrapper[4632]: I0313 10:32:14.909071 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1dc9191f-32b9-45b9-b49f-fd704075f0a5-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5s64h\" (UID: \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" Mar 13 10:32:15 crc kubenswrapper[4632]: I0313 10:32:15.011111 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gncwc\" (UniqueName: \"kubernetes.io/projected/1dc9191f-32b9-45b9-b49f-fd704075f0a5-kube-api-access-gncwc\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5s64h\" (UID: \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" Mar 13 10:32:15 crc kubenswrapper[4632]: I0313 10:32:15.011188 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc9191f-32b9-45b9-b49f-fd704075f0a5-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5s64h\" (UID: \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" Mar 13 10:32:15 crc kubenswrapper[4632]: I0313 10:32:15.011208 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1dc9191f-32b9-45b9-b49f-fd704075f0a5-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5s64h\" (UID: \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" Mar 13 10:32:15 crc kubenswrapper[4632]: I0313 10:32:15.016882 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc9191f-32b9-45b9-b49f-fd704075f0a5-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5s64h\" (UID: \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" Mar 13 10:32:15 crc kubenswrapper[4632]: I0313 10:32:15.023035 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1dc9191f-32b9-45b9-b49f-fd704075f0a5-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5s64h\" (UID: \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" Mar 13 10:32:15 crc kubenswrapper[4632]: I0313 10:32:15.029231 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gncwc\" (UniqueName: \"kubernetes.io/projected/1dc9191f-32b9-45b9-b49f-fd704075f0a5-kube-api-access-gncwc\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5s64h\" (UID: \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" Mar 13 10:32:15 crc kubenswrapper[4632]: I0313 10:32:15.112507 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" Mar 13 10:32:15 crc kubenswrapper[4632]: I0313 10:32:15.682345 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h"] Mar 13 10:32:16 crc kubenswrapper[4632]: I0313 10:32:16.718130 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" event={"ID":"1dc9191f-32b9-45b9-b49f-fd704075f0a5","Type":"ContainerStarted","Data":"77d6861fd1b459222aac820c3526b629eeb8d651fa3c13e5ff7bf6c45b935373"} Mar 13 10:32:16 crc kubenswrapper[4632]: I0313 10:32:16.718405 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" event={"ID":"1dc9191f-32b9-45b9-b49f-fd704075f0a5","Type":"ContainerStarted","Data":"7daf0ba28488eb10767fe67cb2c89de198570e079f99b96b008361e72e107851"} Mar 13 10:32:16 crc kubenswrapper[4632]: I0313 10:32:16.743388 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" podStartSLOduration=2.324840511 podStartE2EDuration="2.743371536s" podCreationTimestamp="2026-03-13 10:32:14 +0000 UTC" firstStartedPulling="2026-03-13 10:32:15.702879262 +0000 UTC m=+1709.725409395" lastFinishedPulling="2026-03-13 10:32:16.121410287 +0000 UTC m=+1710.143940420" observedRunningTime="2026-03-13 10:32:16.73903799 +0000 UTC m=+1710.761568123" watchObservedRunningTime="2026-03-13 10:32:16.743371536 +0000 UTC m=+1710.765901669" Mar 13 10:32:19 crc kubenswrapper[4632]: I0313 10:32:19.754352 4632 generic.go:334] "Generic (PLEG): container finished" podID="1dc9191f-32b9-45b9-b49f-fd704075f0a5" containerID="77d6861fd1b459222aac820c3526b629eeb8d651fa3c13e5ff7bf6c45b935373" exitCode=0 Mar 13 10:32:19 crc kubenswrapper[4632]: I0313 10:32:19.754410 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" event={"ID":"1dc9191f-32b9-45b9-b49f-fd704075f0a5","Type":"ContainerDied","Data":"77d6861fd1b459222aac820c3526b629eeb8d651fa3c13e5ff7bf6c45b935373"} Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.223992 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.424015 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1dc9191f-32b9-45b9-b49f-fd704075f0a5-ssh-key-openstack-edpm-ipam\") pod \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\" (UID: \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\") " Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.424080 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gncwc\" (UniqueName: \"kubernetes.io/projected/1dc9191f-32b9-45b9-b49f-fd704075f0a5-kube-api-access-gncwc\") pod \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\" (UID: \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\") " Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.424153 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc9191f-32b9-45b9-b49f-fd704075f0a5-inventory\") pod \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\" (UID: \"1dc9191f-32b9-45b9-b49f-fd704075f0a5\") " Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.429858 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dc9191f-32b9-45b9-b49f-fd704075f0a5-kube-api-access-gncwc" (OuterVolumeSpecName: "kube-api-access-gncwc") pod "1dc9191f-32b9-45b9-b49f-fd704075f0a5" (UID: "1dc9191f-32b9-45b9-b49f-fd704075f0a5"). InnerVolumeSpecName "kube-api-access-gncwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.456122 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dc9191f-32b9-45b9-b49f-fd704075f0a5-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1dc9191f-32b9-45b9-b49f-fd704075f0a5" (UID: "1dc9191f-32b9-45b9-b49f-fd704075f0a5"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.457746 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1dc9191f-32b9-45b9-b49f-fd704075f0a5-inventory" (OuterVolumeSpecName: "inventory") pod "1dc9191f-32b9-45b9-b49f-fd704075f0a5" (UID: "1dc9191f-32b9-45b9-b49f-fd704075f0a5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.526165 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1dc9191f-32b9-45b9-b49f-fd704075f0a5-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.526415 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gncwc\" (UniqueName: \"kubernetes.io/projected/1dc9191f-32b9-45b9-b49f-fd704075f0a5-kube-api-access-gncwc\") on node \"crc\" DevicePath \"\"" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.526530 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1dc9191f-32b9-45b9-b49f-fd704075f0a5-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.777717 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" event={"ID":"1dc9191f-32b9-45b9-b49f-fd704075f0a5","Type":"ContainerDied","Data":"7daf0ba28488eb10767fe67cb2c89de198570e079f99b96b008361e72e107851"} Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.777803 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7daf0ba28488eb10767fe67cb2c89de198570e079f99b96b008361e72e107851" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.777767 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5s64h" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.878658 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp"] Mar 13 10:32:21 crc kubenswrapper[4632]: E0313 10:32:21.879720 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dc9191f-32b9-45b9-b49f-fd704075f0a5" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.879827 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dc9191f-32b9-45b9-b49f-fd704075f0a5" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.880227 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dc9191f-32b9-45b9-b49f-fd704075f0a5" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.881109 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.883341 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.883749 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.883933 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.894816 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:32:21 crc kubenswrapper[4632]: I0313 10:32:21.897626 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp"] Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.037022 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2z5j\" (UniqueName: \"kubernetes.io/projected/684a2658-ba02-40cf-a371-ec2a8934c0d3-kube-api-access-p2z5j\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.037088 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.037252 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.037295 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.138836 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2z5j\" (UniqueName: \"kubernetes.io/projected/684a2658-ba02-40cf-a371-ec2a8934c0d3-kube-api-access-p2z5j\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.138908 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.139108 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.139171 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.149865 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.149864 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.150582 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.155751 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2z5j\" (UniqueName: \"kubernetes.io/projected/684a2658-ba02-40cf-a371-ec2a8934c0d3-kube-api-access-p2z5j\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.197202 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.757921 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp"] Mar 13 10:32:22 crc kubenswrapper[4632]: I0313 10:32:22.787826 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" event={"ID":"684a2658-ba02-40cf-a371-ec2a8934c0d3","Type":"ContainerStarted","Data":"7e733b330c11a117b4bd0ac4f5dc54f5cdcd79d6738005ec9435833371e5bb1e"} Mar 13 10:32:23 crc kubenswrapper[4632]: I0313 10:32:23.799237 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" event={"ID":"684a2658-ba02-40cf-a371-ec2a8934c0d3","Type":"ContainerStarted","Data":"2a49e9be958588623abc6e501e323438380932e75041c2f3f5b099060c297811"} Mar 13 10:32:23 crc kubenswrapper[4632]: I0313 10:32:23.825022 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" podStartSLOduration=2.364078684 podStartE2EDuration="2.825002681s" podCreationTimestamp="2026-03-13 10:32:21 +0000 UTC" firstStartedPulling="2026-03-13 10:32:22.769561559 +0000 UTC m=+1716.792091692" lastFinishedPulling="2026-03-13 10:32:23.230485566 +0000 UTC m=+1717.253015689" observedRunningTime="2026-03-13 10:32:23.817119092 +0000 UTC m=+1717.839649225" watchObservedRunningTime="2026-03-13 10:32:23.825002681 +0000 UTC m=+1717.847532814" Mar 13 10:32:25 crc kubenswrapper[4632]: I0313 10:32:25.044714 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:32:25 crc kubenswrapper[4632]: E0313 10:32:25.045028 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:32:38 crc kubenswrapper[4632]: I0313 10:32:38.054237 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:32:38 crc kubenswrapper[4632]: E0313 10:32:38.055063 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:32:49 crc kubenswrapper[4632]: I0313 10:32:49.045171 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:32:49 crc kubenswrapper[4632]: E0313 10:32:49.046263 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:33:02 crc kubenswrapper[4632]: I0313 10:33:02.044810 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:33:02 crc kubenswrapper[4632]: E0313 10:33:02.045504 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:33:06 crc kubenswrapper[4632]: I0313 10:33:06.740819 4632 scope.go:117] "RemoveContainer" containerID="e8fc7f9526396e3f4333f93ccef86f72aee3214939c63a5e8145c990bbf9d938" Mar 13 10:33:15 crc kubenswrapper[4632]: I0313 10:33:15.044933 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:33:15 crc kubenswrapper[4632]: E0313 10:33:15.045731 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:33:29 crc kubenswrapper[4632]: I0313 10:33:29.044470 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:33:29 crc kubenswrapper[4632]: E0313 10:33:29.045236 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:33:41 crc kubenswrapper[4632]: I0313 10:33:41.045988 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:33:41 crc kubenswrapper[4632]: E0313 10:33:41.046853 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:33:54 crc kubenswrapper[4632]: I0313 10:33:54.044024 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:33:54 crc kubenswrapper[4632]: E0313 10:33:54.045006 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:34:00 crc kubenswrapper[4632]: I0313 10:34:00.149140 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556634-6n98g"] Mar 13 10:34:00 crc kubenswrapper[4632]: I0313 10:34:00.152990 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556634-6n98g" Mar 13 10:34:00 crc kubenswrapper[4632]: I0313 10:34:00.155905 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:34:00 crc kubenswrapper[4632]: I0313 10:34:00.155919 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:34:00 crc kubenswrapper[4632]: I0313 10:34:00.157296 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:34:00 crc kubenswrapper[4632]: I0313 10:34:00.162914 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556634-6n98g"] Mar 13 10:34:00 crc kubenswrapper[4632]: I0313 10:34:00.297716 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mh7l9\" (UniqueName: \"kubernetes.io/projected/155ba738-4ba0-424a-a1d7-067786728969-kube-api-access-mh7l9\") pod \"auto-csr-approver-29556634-6n98g\" (UID: \"155ba738-4ba0-424a-a1d7-067786728969\") " pod="openshift-infra/auto-csr-approver-29556634-6n98g" Mar 13 10:34:00 crc kubenswrapper[4632]: I0313 10:34:00.400282 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mh7l9\" (UniqueName: \"kubernetes.io/projected/155ba738-4ba0-424a-a1d7-067786728969-kube-api-access-mh7l9\") pod \"auto-csr-approver-29556634-6n98g\" (UID: \"155ba738-4ba0-424a-a1d7-067786728969\") " pod="openshift-infra/auto-csr-approver-29556634-6n98g" Mar 13 10:34:00 crc kubenswrapper[4632]: I0313 10:34:00.421404 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mh7l9\" (UniqueName: \"kubernetes.io/projected/155ba738-4ba0-424a-a1d7-067786728969-kube-api-access-mh7l9\") pod \"auto-csr-approver-29556634-6n98g\" (UID: \"155ba738-4ba0-424a-a1d7-067786728969\") " pod="openshift-infra/auto-csr-approver-29556634-6n98g" Mar 13 10:34:00 crc kubenswrapper[4632]: I0313 10:34:00.485019 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556634-6n98g" Mar 13 10:34:01 crc kubenswrapper[4632]: I0313 10:34:01.020358 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556634-6n98g"] Mar 13 10:34:01 crc kubenswrapper[4632]: I0313 10:34:01.769761 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556634-6n98g" event={"ID":"155ba738-4ba0-424a-a1d7-067786728969","Type":"ContainerStarted","Data":"162ef96ec7c09d383ed757ecece575f1df9b067804234a90b5649e29f280fcfd"} Mar 13 10:34:03 crc kubenswrapper[4632]: I0313 10:34:03.788576 4632 generic.go:334] "Generic (PLEG): container finished" podID="155ba738-4ba0-424a-a1d7-067786728969" containerID="f95b291d052d44a477db7fca5558efb7e90f20270d66ae208043b37111d582be" exitCode=0 Mar 13 10:34:03 crc kubenswrapper[4632]: I0313 10:34:03.788794 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556634-6n98g" event={"ID":"155ba738-4ba0-424a-a1d7-067786728969","Type":"ContainerDied","Data":"f95b291d052d44a477db7fca5558efb7e90f20270d66ae208043b37111d582be"} Mar 13 10:34:05 crc kubenswrapper[4632]: I0313 10:34:05.045613 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:34:05 crc kubenswrapper[4632]: E0313 10:34:05.046229 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:34:05 crc kubenswrapper[4632]: I0313 10:34:05.159759 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556634-6n98g" Mar 13 10:34:05 crc kubenswrapper[4632]: I0313 10:34:05.296233 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mh7l9\" (UniqueName: \"kubernetes.io/projected/155ba738-4ba0-424a-a1d7-067786728969-kube-api-access-mh7l9\") pod \"155ba738-4ba0-424a-a1d7-067786728969\" (UID: \"155ba738-4ba0-424a-a1d7-067786728969\") " Mar 13 10:34:05 crc kubenswrapper[4632]: I0313 10:34:05.313819 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/155ba738-4ba0-424a-a1d7-067786728969-kube-api-access-mh7l9" (OuterVolumeSpecName: "kube-api-access-mh7l9") pod "155ba738-4ba0-424a-a1d7-067786728969" (UID: "155ba738-4ba0-424a-a1d7-067786728969"). InnerVolumeSpecName "kube-api-access-mh7l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:34:05 crc kubenswrapper[4632]: I0313 10:34:05.398835 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mh7l9\" (UniqueName: \"kubernetes.io/projected/155ba738-4ba0-424a-a1d7-067786728969-kube-api-access-mh7l9\") on node \"crc\" DevicePath \"\"" Mar 13 10:34:05 crc kubenswrapper[4632]: I0313 10:34:05.813805 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556634-6n98g" event={"ID":"155ba738-4ba0-424a-a1d7-067786728969","Type":"ContainerDied","Data":"162ef96ec7c09d383ed757ecece575f1df9b067804234a90b5649e29f280fcfd"} Mar 13 10:34:05 crc kubenswrapper[4632]: I0313 10:34:05.813844 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="162ef96ec7c09d383ed757ecece575f1df9b067804234a90b5649e29f280fcfd" Mar 13 10:34:05 crc kubenswrapper[4632]: I0313 10:34:05.813884 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556634-6n98g" Mar 13 10:34:06 crc kubenswrapper[4632]: I0313 10:34:06.250293 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556628-479rr"] Mar 13 10:34:06 crc kubenswrapper[4632]: I0313 10:34:06.258396 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556628-479rr"] Mar 13 10:34:08 crc kubenswrapper[4632]: I0313 10:34:08.066833 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="658f9ba3-69b7-4d2d-8258-bb7bdf272398" path="/var/lib/kubelet/pods/658f9ba3-69b7-4d2d-8258-bb7bdf272398/volumes" Mar 13 10:34:18 crc kubenswrapper[4632]: I0313 10:34:18.051989 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:34:18 crc kubenswrapper[4632]: E0313 10:34:18.052812 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:34:20 crc kubenswrapper[4632]: I0313 10:34:20.062102 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-bfb6b"] Mar 13 10:34:20 crc kubenswrapper[4632]: I0313 10:34:20.067833 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-bfb6b"] Mar 13 10:34:21 crc kubenswrapper[4632]: I0313 10:34:21.054113 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-64xvf"] Mar 13 10:34:21 crc kubenswrapper[4632]: I0313 10:34:21.070000 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a750-account-create-update-7wk26"] Mar 13 10:34:21 crc kubenswrapper[4632]: I0313 10:34:21.089241 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-7hqpw"] Mar 13 10:34:21 crc kubenswrapper[4632]: I0313 10:34:21.103252 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-ab0c-account-create-update-tr7hx"] Mar 13 10:34:21 crc kubenswrapper[4632]: I0313 10:34:21.112977 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-9698-account-create-update-9kfhv"] Mar 13 10:34:21 crc kubenswrapper[4632]: I0313 10:34:21.123070 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-64xvf"] Mar 13 10:34:21 crc kubenswrapper[4632]: I0313 10:34:21.132440 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-9698-account-create-update-9kfhv"] Mar 13 10:34:21 crc kubenswrapper[4632]: I0313 10:34:21.140619 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-7hqpw"] Mar 13 10:34:21 crc kubenswrapper[4632]: I0313 10:34:21.149013 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-a750-account-create-update-7wk26"] Mar 13 10:34:21 crc kubenswrapper[4632]: I0313 10:34:21.159174 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-ab0c-account-create-update-tr7hx"] Mar 13 10:34:22 crc kubenswrapper[4632]: I0313 10:34:22.060892 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2610abab-1da4-4912-9ca7-f2aa2d7c0486" path="/var/lib/kubelet/pods/2610abab-1da4-4912-9ca7-f2aa2d7c0486/volumes" Mar 13 10:34:22 crc kubenswrapper[4632]: I0313 10:34:22.063549 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="584d2818-4b22-468f-b296-bd1850c7915b" path="/var/lib/kubelet/pods/584d2818-4b22-468f-b296-bd1850c7915b/volumes" Mar 13 10:34:22 crc kubenswrapper[4632]: I0313 10:34:22.067712 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f09e2f4-4f82-4388-9b5a-a9e890d3a950" path="/var/lib/kubelet/pods/5f09e2f4-4f82-4388-9b5a-a9e890d3a950/volumes" Mar 13 10:34:22 crc kubenswrapper[4632]: I0313 10:34:22.072917 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c84aa49-2900-4a14-b81b-bb03e925d1b7" path="/var/lib/kubelet/pods/6c84aa49-2900-4a14-b81b-bb03e925d1b7/volumes" Mar 13 10:34:22 crc kubenswrapper[4632]: I0313 10:34:22.075510 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e353045-e09b-4cd2-b659-1954485ec8db" path="/var/lib/kubelet/pods/8e353045-e09b-4cd2-b659-1954485ec8db/volumes" Mar 13 10:34:22 crc kubenswrapper[4632]: I0313 10:34:22.077457 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4f6b362-7670-4867-b8fa-1f4c6170389f" path="/var/lib/kubelet/pods/c4f6b362-7670-4867-b8fa-1f4c6170389f/volumes" Mar 13 10:34:25 crc kubenswrapper[4632]: I0313 10:34:25.052374 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-lrjmj"] Mar 13 10:34:25 crc kubenswrapper[4632]: I0313 10:34:25.066991 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-lrjmj"] Mar 13 10:34:26 crc kubenswrapper[4632]: I0313 10:34:26.060215 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d670715-74f3-46a6-974c-b6953af9fdb7" path="/var/lib/kubelet/pods/2d670715-74f3-46a6-974c-b6953af9fdb7/volumes" Mar 13 10:34:29 crc kubenswrapper[4632]: I0313 10:34:29.044804 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:34:29 crc kubenswrapper[4632]: E0313 10:34:29.045714 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:34:40 crc kubenswrapper[4632]: I0313 10:34:40.044827 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:34:40 crc kubenswrapper[4632]: E0313 10:34:40.045700 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:34:49 crc kubenswrapper[4632]: I0313 10:34:49.042953 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-21a0-account-create-update-4clr7"] Mar 13 10:34:49 crc kubenswrapper[4632]: I0313 10:34:49.051613 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-21a0-account-create-update-4clr7"] Mar 13 10:34:50 crc kubenswrapper[4632]: I0313 10:34:50.032871 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-dwf4b"] Mar 13 10:34:50 crc kubenswrapper[4632]: I0313 10:34:50.057499 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a03b92ea-cd2c-455d-a88e-1d57b958b138" path="/var/lib/kubelet/pods/a03b92ea-cd2c-455d-a88e-1d57b958b138/volumes" Mar 13 10:34:50 crc kubenswrapper[4632]: I0313 10:34:50.059854 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-kp87n"] Mar 13 10:34:50 crc kubenswrapper[4632]: I0313 10:34:50.059893 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-kp87n"] Mar 13 10:34:50 crc kubenswrapper[4632]: I0313 10:34:50.064695 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-dwf4b"] Mar 13 10:34:52 crc kubenswrapper[4632]: I0313 10:34:52.063631 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="239c554e-360d-4f04-86f0-b2b98974bad3" path="/var/lib/kubelet/pods/239c554e-360d-4f04-86f0-b2b98974bad3/volumes" Mar 13 10:34:52 crc kubenswrapper[4632]: I0313 10:34:52.069955 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcdcfad1-d735-4b55-ae65-0ce16bdbc79d" path="/var/lib/kubelet/pods/bcdcfad1-d735-4b55-ae65-0ce16bdbc79d/volumes" Mar 13 10:34:54 crc kubenswrapper[4632]: I0313 10:34:54.044315 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:34:54 crc kubenswrapper[4632]: E0313 10:34:54.045033 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:34:55 crc kubenswrapper[4632]: I0313 10:34:55.050536 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-da66-account-create-update-tk8pd"] Mar 13 10:34:55 crc kubenswrapper[4632]: I0313 10:34:55.062801 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-pnvjb"] Mar 13 10:34:55 crc kubenswrapper[4632]: I0313 10:34:55.074415 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-b742-account-create-update-gfdkg"] Mar 13 10:34:55 crc kubenswrapper[4632]: I0313 10:34:55.079349 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-g7pfc"] Mar 13 10:34:55 crc kubenswrapper[4632]: I0313 10:34:55.088531 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-pnvjb"] Mar 13 10:34:55 crc kubenswrapper[4632]: I0313 10:34:55.098128 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-g7pfc"] Mar 13 10:34:55 crc kubenswrapper[4632]: I0313 10:34:55.105081 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-da66-account-create-update-tk8pd"] Mar 13 10:34:55 crc kubenswrapper[4632]: I0313 10:34:55.113818 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-4dec-account-create-update-hfnth"] Mar 13 10:34:55 crc kubenswrapper[4632]: I0313 10:34:55.126366 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-b742-account-create-update-gfdkg"] Mar 13 10:34:55 crc kubenswrapper[4632]: I0313 10:34:55.135993 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-4dec-account-create-update-hfnth"] Mar 13 10:34:56 crc kubenswrapper[4632]: I0313 10:34:56.059219 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d045bc7-38b2-46f5-8cd8-cf634003bedf" path="/var/lib/kubelet/pods/0d045bc7-38b2-46f5-8cd8-cf634003bedf/volumes" Mar 13 10:34:56 crc kubenswrapper[4632]: I0313 10:34:56.062630 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47870992-2db9-46f4-84d9-fd50fb9851eb" path="/var/lib/kubelet/pods/47870992-2db9-46f4-84d9-fd50fb9851eb/volumes" Mar 13 10:34:56 crc kubenswrapper[4632]: I0313 10:34:56.064747 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa0000da-8f11-4e97-8ab5-1bcfea0ac894" path="/var/lib/kubelet/pods/aa0000da-8f11-4e97-8ab5-1bcfea0ac894/volumes" Mar 13 10:34:56 crc kubenswrapper[4632]: I0313 10:34:56.066707 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb216b07-9809-4b8b-857b-ac1192747b9c" path="/var/lib/kubelet/pods/cb216b07-9809-4b8b-857b-ac1192747b9c/volumes" Mar 13 10:34:56 crc kubenswrapper[4632]: I0313 10:34:56.069720 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee94a050-f905-44f1-a5da-16536b8cdfa7" path="/var/lib/kubelet/pods/ee94a050-f905-44f1-a5da-16536b8cdfa7/volumes" Mar 13 10:35:03 crc kubenswrapper[4632]: I0313 10:35:03.034851 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-mq9np"] Mar 13 10:35:03 crc kubenswrapper[4632]: I0313 10:35:03.047228 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-mq9np"] Mar 13 10:35:04 crc kubenswrapper[4632]: I0313 10:35:04.059447 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e824ae7d-dbbd-496b-b8b0-8b5c59a4d419" path="/var/lib/kubelet/pods/e824ae7d-dbbd-496b-b8b0-8b5c59a4d419/volumes" Mar 13 10:35:06 crc kubenswrapper[4632]: I0313 10:35:06.045526 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:35:06 crc kubenswrapper[4632]: E0313 10:35:06.046274 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:35:06 crc kubenswrapper[4632]: I0313 10:35:06.857028 4632 scope.go:117] "RemoveContainer" containerID="05358506b7b8a5602da80aa6b4985f897c7b0818d4a2f70ed84421563493ee78" Mar 13 10:35:06 crc kubenswrapper[4632]: I0313 10:35:06.885099 4632 scope.go:117] "RemoveContainer" containerID="53c212eae0f18baff6fdcd0d88db82f3271a3997b68292e7fdae508ea7808719" Mar 13 10:35:06 crc kubenswrapper[4632]: I0313 10:35:06.934853 4632 scope.go:117] "RemoveContainer" containerID="c0ed44d952b9a10d8f17f6b274d11ae8079f72b678bca2ec969eb44a14c0f18e" Mar 13 10:35:06 crc kubenswrapper[4632]: I0313 10:35:06.979225 4632 scope.go:117] "RemoveContainer" containerID="24cb5f7263654577bea6ec83ce575dcb325e9b55c8adac840790cd7a29363013" Mar 13 10:35:07 crc kubenswrapper[4632]: I0313 10:35:07.033866 4632 scope.go:117] "RemoveContainer" containerID="10bcedf0effae05b832e3793407fcf2703d9df4f7136a8211c78de6b0a99c17b" Mar 13 10:35:07 crc kubenswrapper[4632]: I0313 10:35:07.071736 4632 scope.go:117] "RemoveContainer" containerID="2fd6ae14a44d07bfe626dada3603473befbf9326ca83648414737abd80e0ce5e" Mar 13 10:35:07 crc kubenswrapper[4632]: I0313 10:35:07.117237 4632 scope.go:117] "RemoveContainer" containerID="207587c5bdcbf92f71ab5aedfecf2486734ea587705753fb95e8790e674e977d" Mar 13 10:35:07 crc kubenswrapper[4632]: I0313 10:35:07.141124 4632 scope.go:117] "RemoveContainer" containerID="dc07b5437ef3867ede6e9debff7196fad98555045e8df8dafdb4a11a7fb9808e" Mar 13 10:35:07 crc kubenswrapper[4632]: I0313 10:35:07.166250 4632 scope.go:117] "RemoveContainer" containerID="9bfb87771985986bb5edbb713355c76b663fe8b23df1170e73c42c65479f44df" Mar 13 10:35:07 crc kubenswrapper[4632]: I0313 10:35:07.188957 4632 scope.go:117] "RemoveContainer" containerID="dccd7606dfc8be32af7f5d6d0a4bf2a63f79937bfd68d93b573f727a7eb9e402" Mar 13 10:35:07 crc kubenswrapper[4632]: I0313 10:35:07.238496 4632 scope.go:117] "RemoveContainer" containerID="209b78ccf3afd3b3582d4d4eae9056be2d6d19f860431a427d43f1899c69be92" Mar 13 10:35:07 crc kubenswrapper[4632]: I0313 10:35:07.262831 4632 scope.go:117] "RemoveContainer" containerID="f13e115025698b8daa562f4881b31bb57b43cf222144f35c644ca079c94f546c" Mar 13 10:35:07 crc kubenswrapper[4632]: I0313 10:35:07.287100 4632 scope.go:117] "RemoveContainer" containerID="1ead25cb79a035bd17ce1b8995cb1c20666089312b5c266ebcbccc7e66e7c0cc" Mar 13 10:35:07 crc kubenswrapper[4632]: I0313 10:35:07.307562 4632 scope.go:117] "RemoveContainer" containerID="d92125a86d78e277913519dc023b0643c481c49ac75357c10f1cb11e638c36a3" Mar 13 10:35:07 crc kubenswrapper[4632]: I0313 10:35:07.333434 4632 scope.go:117] "RemoveContainer" containerID="40127d251d4cb7407ae0ce8a1705cd5210171fb2a750df3289fa3b2b9a54b055" Mar 13 10:35:07 crc kubenswrapper[4632]: I0313 10:35:07.354074 4632 scope.go:117] "RemoveContainer" containerID="f79fdacee095a4d2c557179a3aeeb0eea1874c7280d8a656f2dd9779cf567f1e" Mar 13 10:35:07 crc kubenswrapper[4632]: I0313 10:35:07.377152 4632 scope.go:117] "RemoveContainer" containerID="e12bb579655132c65f7afaf171587507463b77c9b73d0902f8981397a2c342cd" Mar 13 10:35:20 crc kubenswrapper[4632]: I0313 10:35:20.045380 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:35:20 crc kubenswrapper[4632]: E0313 10:35:20.046190 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:35:25 crc kubenswrapper[4632]: I0313 10:35:25.606256 4632 generic.go:334] "Generic (PLEG): container finished" podID="684a2658-ba02-40cf-a371-ec2a8934c0d3" containerID="2a49e9be958588623abc6e501e323438380932e75041c2f3f5b099060c297811" exitCode=0 Mar 13 10:35:25 crc kubenswrapper[4632]: I0313 10:35:25.606405 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" event={"ID":"684a2658-ba02-40cf-a371-ec2a8934c0d3","Type":"ContainerDied","Data":"2a49e9be958588623abc6e501e323438380932e75041c2f3f5b099060c297811"} Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.076889 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.191191 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-inventory\") pod \"684a2658-ba02-40cf-a371-ec2a8934c0d3\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.191240 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2z5j\" (UniqueName: \"kubernetes.io/projected/684a2658-ba02-40cf-a371-ec2a8934c0d3-kube-api-access-p2z5j\") pod \"684a2658-ba02-40cf-a371-ec2a8934c0d3\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.191547 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-ssh-key-openstack-edpm-ipam\") pod \"684a2658-ba02-40cf-a371-ec2a8934c0d3\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.191613 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-bootstrap-combined-ca-bundle\") pod \"684a2658-ba02-40cf-a371-ec2a8934c0d3\" (UID: \"684a2658-ba02-40cf-a371-ec2a8934c0d3\") " Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.196858 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/684a2658-ba02-40cf-a371-ec2a8934c0d3-kube-api-access-p2z5j" (OuterVolumeSpecName: "kube-api-access-p2z5j") pod "684a2658-ba02-40cf-a371-ec2a8934c0d3" (UID: "684a2658-ba02-40cf-a371-ec2a8934c0d3"). InnerVolumeSpecName "kube-api-access-p2z5j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.199202 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "684a2658-ba02-40cf-a371-ec2a8934c0d3" (UID: "684a2658-ba02-40cf-a371-ec2a8934c0d3"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.220965 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-inventory" (OuterVolumeSpecName: "inventory") pod "684a2658-ba02-40cf-a371-ec2a8934c0d3" (UID: "684a2658-ba02-40cf-a371-ec2a8934c0d3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.221956 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "684a2658-ba02-40cf-a371-ec2a8934c0d3" (UID: "684a2658-ba02-40cf-a371-ec2a8934c0d3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.293553 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.293588 4632 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.293600 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/684a2658-ba02-40cf-a371-ec2a8934c0d3-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.293609 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2z5j\" (UniqueName: \"kubernetes.io/projected/684a2658-ba02-40cf-a371-ec2a8934c0d3-kube-api-access-p2z5j\") on node \"crc\" DevicePath \"\"" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.629226 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" event={"ID":"684a2658-ba02-40cf-a371-ec2a8934c0d3","Type":"ContainerDied","Data":"7e733b330c11a117b4bd0ac4f5dc54f5cdcd79d6738005ec9435833371e5bb1e"} Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.629270 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e733b330c11a117b4bd0ac4f5dc54f5cdcd79d6738005ec9435833371e5bb1e" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.629279 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.726808 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp"] Mar 13 10:35:27 crc kubenswrapper[4632]: E0313 10:35:27.727219 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="684a2658-ba02-40cf-a371-ec2a8934c0d3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.727237 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="684a2658-ba02-40cf-a371-ec2a8934c0d3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 13 10:35:27 crc kubenswrapper[4632]: E0313 10:35:27.727265 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="155ba738-4ba0-424a-a1d7-067786728969" containerName="oc" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.727271 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="155ba738-4ba0-424a-a1d7-067786728969" containerName="oc" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.727449 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="684a2658-ba02-40cf-a371-ec2a8934c0d3" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.727481 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="155ba738-4ba0-424a-a1d7-067786728969" containerName="oc" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.728267 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.731031 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.731078 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.731429 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.731489 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.738106 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp"] Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.802514 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qctqc\" (UniqueName: \"kubernetes.io/projected/0d75181a-4c91-485e-8bcd-02e2aedd4d45-kube-api-access-qctqc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-754cp\" (UID: \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.802639 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d75181a-4c91-485e-8bcd-02e2aedd4d45-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-754cp\" (UID: \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.802723 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d75181a-4c91-485e-8bcd-02e2aedd4d45-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-754cp\" (UID: \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.905030 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d75181a-4c91-485e-8bcd-02e2aedd4d45-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-754cp\" (UID: \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.905135 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d75181a-4c91-485e-8bcd-02e2aedd4d45-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-754cp\" (UID: \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.905236 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qctqc\" (UniqueName: \"kubernetes.io/projected/0d75181a-4c91-485e-8bcd-02e2aedd4d45-kube-api-access-qctqc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-754cp\" (UID: \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.933517 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d75181a-4c91-485e-8bcd-02e2aedd4d45-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-754cp\" (UID: \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.936687 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d75181a-4c91-485e-8bcd-02e2aedd4d45-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-754cp\" (UID: \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" Mar 13 10:35:27 crc kubenswrapper[4632]: I0313 10:35:27.942834 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qctqc\" (UniqueName: \"kubernetes.io/projected/0d75181a-4c91-485e-8bcd-02e2aedd4d45-kube-api-access-qctqc\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-754cp\" (UID: \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" Mar 13 10:35:28 crc kubenswrapper[4632]: I0313 10:35:28.043993 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" Mar 13 10:35:28 crc kubenswrapper[4632]: I0313 10:35:28.708930 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp"] Mar 13 10:35:28 crc kubenswrapper[4632]: I0313 10:35:28.719112 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:35:29 crc kubenswrapper[4632]: I0313 10:35:29.652473 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" event={"ID":"0d75181a-4c91-485e-8bcd-02e2aedd4d45","Type":"ContainerStarted","Data":"bd09862a5fc80def82e97b44b7d539caee7c696bb410023ec19cde3384abb6ae"} Mar 13 10:35:29 crc kubenswrapper[4632]: I0313 10:35:29.652825 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" event={"ID":"0d75181a-4c91-485e-8bcd-02e2aedd4d45","Type":"ContainerStarted","Data":"b2534ff7153cbd23578a95b816373371c87f8d95ec5b00bff35a4aeb9a12cb51"} Mar 13 10:35:29 crc kubenswrapper[4632]: I0313 10:35:29.683527 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" podStartSLOduration=2.196335944 podStartE2EDuration="2.683503676s" podCreationTimestamp="2026-03-13 10:35:27 +0000 UTC" firstStartedPulling="2026-03-13 10:35:28.718832011 +0000 UTC m=+1902.741362144" lastFinishedPulling="2026-03-13 10:35:29.205999743 +0000 UTC m=+1903.228529876" observedRunningTime="2026-03-13 10:35:29.672305749 +0000 UTC m=+1903.694835882" watchObservedRunningTime="2026-03-13 10:35:29.683503676 +0000 UTC m=+1903.706033819" Mar 13 10:35:34 crc kubenswrapper[4632]: I0313 10:35:34.044460 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:35:34 crc kubenswrapper[4632]: E0313 10:35:34.045392 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:35:37 crc kubenswrapper[4632]: I0313 10:35:37.042963 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-l6hpb"] Mar 13 10:35:37 crc kubenswrapper[4632]: I0313 10:35:37.051736 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-l6hpb"] Mar 13 10:35:38 crc kubenswrapper[4632]: I0313 10:35:38.055893 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f1c5663-463b-45e2-b200-64e73e6d5698" path="/var/lib/kubelet/pods/4f1c5663-463b-45e2-b200-64e73e6d5698/volumes" Mar 13 10:35:48 crc kubenswrapper[4632]: I0313 10:35:48.086912 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:35:48 crc kubenswrapper[4632]: E0313 10:35:48.089278 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:35:53 crc kubenswrapper[4632]: I0313 10:35:53.043556 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-hlsnz"] Mar 13 10:35:53 crc kubenswrapper[4632]: I0313 10:35:53.054472 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-hlsnz"] Mar 13 10:35:54 crc kubenswrapper[4632]: I0313 10:35:54.087331 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7221b50-7231-4ade-917e-b10f177cb539" path="/var/lib/kubelet/pods/b7221b50-7231-4ade-917e-b10f177cb539/volumes" Mar 13 10:35:59 crc kubenswrapper[4632]: I0313 10:35:59.045259 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:35:59 crc kubenswrapper[4632]: E0313 10:35:59.046174 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:36:00 crc kubenswrapper[4632]: I0313 10:36:00.145924 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556636-zncpw"] Mar 13 10:36:00 crc kubenswrapper[4632]: I0313 10:36:00.147376 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556636-zncpw" Mar 13 10:36:00 crc kubenswrapper[4632]: I0313 10:36:00.150057 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:36:00 crc kubenswrapper[4632]: I0313 10:36:00.150835 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:36:00 crc kubenswrapper[4632]: I0313 10:36:00.151871 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:36:00 crc kubenswrapper[4632]: I0313 10:36:00.158363 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556636-zncpw"] Mar 13 10:36:00 crc kubenswrapper[4632]: I0313 10:36:00.232872 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2ksk\" (UniqueName: \"kubernetes.io/projected/7cb201b3-b479-4877-a996-58045d0720c4-kube-api-access-r2ksk\") pod \"auto-csr-approver-29556636-zncpw\" (UID: \"7cb201b3-b479-4877-a996-58045d0720c4\") " pod="openshift-infra/auto-csr-approver-29556636-zncpw" Mar 13 10:36:00 crc kubenswrapper[4632]: I0313 10:36:00.334477 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2ksk\" (UniqueName: \"kubernetes.io/projected/7cb201b3-b479-4877-a996-58045d0720c4-kube-api-access-r2ksk\") pod \"auto-csr-approver-29556636-zncpw\" (UID: \"7cb201b3-b479-4877-a996-58045d0720c4\") " pod="openshift-infra/auto-csr-approver-29556636-zncpw" Mar 13 10:36:00 crc kubenswrapper[4632]: I0313 10:36:00.365246 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2ksk\" (UniqueName: \"kubernetes.io/projected/7cb201b3-b479-4877-a996-58045d0720c4-kube-api-access-r2ksk\") pod \"auto-csr-approver-29556636-zncpw\" (UID: \"7cb201b3-b479-4877-a996-58045d0720c4\") " pod="openshift-infra/auto-csr-approver-29556636-zncpw" Mar 13 10:36:00 crc kubenswrapper[4632]: I0313 10:36:00.482379 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556636-zncpw" Mar 13 10:36:01 crc kubenswrapper[4632]: I0313 10:36:01.038749 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556636-zncpw"] Mar 13 10:36:01 crc kubenswrapper[4632]: I0313 10:36:01.950302 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556636-zncpw" event={"ID":"7cb201b3-b479-4877-a996-58045d0720c4","Type":"ContainerStarted","Data":"5a14d1fb4e0b470a0079f1b91f89aa951453210f6aa6c7e6131e24e29d1de5b2"} Mar 13 10:36:02 crc kubenswrapper[4632]: I0313 10:36:02.961625 4632 generic.go:334] "Generic (PLEG): container finished" podID="7cb201b3-b479-4877-a996-58045d0720c4" containerID="1d5789598fed395c0d259939fb11bb98aa8eec3b7168c00349a4a3635d4bd5ce" exitCode=0 Mar 13 10:36:02 crc kubenswrapper[4632]: I0313 10:36:02.961670 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556636-zncpw" event={"ID":"7cb201b3-b479-4877-a996-58045d0720c4","Type":"ContainerDied","Data":"1d5789598fed395c0d259939fb11bb98aa8eec3b7168c00349a4a3635d4bd5ce"} Mar 13 10:36:04 crc kubenswrapper[4632]: I0313 10:36:04.339403 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556636-zncpw" Mar 13 10:36:04 crc kubenswrapper[4632]: I0313 10:36:04.529142 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2ksk\" (UniqueName: \"kubernetes.io/projected/7cb201b3-b479-4877-a996-58045d0720c4-kube-api-access-r2ksk\") pod \"7cb201b3-b479-4877-a996-58045d0720c4\" (UID: \"7cb201b3-b479-4877-a996-58045d0720c4\") " Mar 13 10:36:04 crc kubenswrapper[4632]: I0313 10:36:04.535651 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7cb201b3-b479-4877-a996-58045d0720c4-kube-api-access-r2ksk" (OuterVolumeSpecName: "kube-api-access-r2ksk") pod "7cb201b3-b479-4877-a996-58045d0720c4" (UID: "7cb201b3-b479-4877-a996-58045d0720c4"). InnerVolumeSpecName "kube-api-access-r2ksk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:36:04 crc kubenswrapper[4632]: I0313 10:36:04.631502 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2ksk\" (UniqueName: \"kubernetes.io/projected/7cb201b3-b479-4877-a996-58045d0720c4-kube-api-access-r2ksk\") on node \"crc\" DevicePath \"\"" Mar 13 10:36:04 crc kubenswrapper[4632]: I0313 10:36:04.985202 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556636-zncpw" event={"ID":"7cb201b3-b479-4877-a996-58045d0720c4","Type":"ContainerDied","Data":"5a14d1fb4e0b470a0079f1b91f89aa951453210f6aa6c7e6131e24e29d1de5b2"} Mar 13 10:36:04 crc kubenswrapper[4632]: I0313 10:36:04.985335 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a14d1fb4e0b470a0079f1b91f89aa951453210f6aa6c7e6131e24e29d1de5b2" Mar 13 10:36:04 crc kubenswrapper[4632]: I0313 10:36:04.985819 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556636-zncpw" Mar 13 10:36:05 crc kubenswrapper[4632]: I0313 10:36:05.413620 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556630-kxrkn"] Mar 13 10:36:05 crc kubenswrapper[4632]: I0313 10:36:05.422970 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556630-kxrkn"] Mar 13 10:36:06 crc kubenswrapper[4632]: I0313 10:36:06.057163 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e" path="/var/lib/kubelet/pods/b0ccb00a-40ce-4b3d-86e8-8f87354c1e8e/volumes" Mar 13 10:36:07 crc kubenswrapper[4632]: I0313 10:36:07.685542 4632 scope.go:117] "RemoveContainer" containerID="bf8d93edd68f1cf79021467ff9910419baf75397a4140fb3d25bca7f97abbf70" Mar 13 10:36:07 crc kubenswrapper[4632]: I0313 10:36:07.711928 4632 scope.go:117] "RemoveContainer" containerID="4df2156f6fe32fab45f05d256a8ec2adb23f786a2989c939b92b996a496f122f" Mar 13 10:36:07 crc kubenswrapper[4632]: I0313 10:36:07.761419 4632 scope.go:117] "RemoveContainer" containerID="e604b5ae6ce92dde6f33a140a99a7c7d5949aebd7f4821ef087f38b50a0e872b" Mar 13 10:36:07 crc kubenswrapper[4632]: I0313 10:36:07.787065 4632 scope.go:117] "RemoveContainer" containerID="a4f9bd4f877455829b998ee69c6d5f9dd7fb999a6d06fe2960e4af1bfddc1eb0" Mar 13 10:36:07 crc kubenswrapper[4632]: I0313 10:36:07.826240 4632 scope.go:117] "RemoveContainer" containerID="dd1843e80da062d2b847859e60f624eed6f5f23e9e94519edc79cfc924e74d60" Mar 13 10:36:12 crc kubenswrapper[4632]: I0313 10:36:12.045174 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:36:13 crc kubenswrapper[4632]: I0313 10:36:13.082224 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"2bb4e222f4f89a1d4e4bebc809fc60cc762d7ea9b6811f4bcc9cb78c179cd0bd"} Mar 13 10:36:14 crc kubenswrapper[4632]: I0313 10:36:14.035584 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-htnd9"] Mar 13 10:36:14 crc kubenswrapper[4632]: I0313 10:36:14.060267 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-htnd9"] Mar 13 10:36:16 crc kubenswrapper[4632]: I0313 10:36:16.056303 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e92afa62-9c75-4e0e-92f4-76e57328d7a0" path="/var/lib/kubelet/pods/e92afa62-9c75-4e0e-92f4-76e57328d7a0/volumes" Mar 13 10:36:24 crc kubenswrapper[4632]: I0313 10:36:24.072798 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-zdgpw"] Mar 13 10:36:24 crc kubenswrapper[4632]: I0313 10:36:24.085920 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-x8tq8"] Mar 13 10:36:24 crc kubenswrapper[4632]: I0313 10:36:24.096080 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-7fvlk"] Mar 13 10:36:24 crc kubenswrapper[4632]: I0313 10:36:24.109163 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-7fvlk"] Mar 13 10:36:24 crc kubenswrapper[4632]: I0313 10:36:24.119690 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-zdgpw"] Mar 13 10:36:24 crc kubenswrapper[4632]: I0313 10:36:24.129074 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-x8tq8"] Mar 13 10:36:26 crc kubenswrapper[4632]: I0313 10:36:26.059414 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="418cb883-abd1-46b4-957f-0a40f3e62297" path="/var/lib/kubelet/pods/418cb883-abd1-46b4-957f-0a40f3e62297/volumes" Mar 13 10:36:26 crc kubenswrapper[4632]: I0313 10:36:26.060653 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d722ddd7-e65d-44f7-a02d-18ddf126ccf5" path="/var/lib/kubelet/pods/d722ddd7-e65d-44f7-a02d-18ddf126ccf5/volumes" Mar 13 10:36:26 crc kubenswrapper[4632]: I0313 10:36:26.064180 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8d0f662-d180-4137-8107-e465c5fb0621" path="/var/lib/kubelet/pods/d8d0f662-d180-4137-8107-e465c5fb0621/volumes" Mar 13 10:36:33 crc kubenswrapper[4632]: I0313 10:36:33.051775 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-kq8lc"] Mar 13 10:36:33 crc kubenswrapper[4632]: I0313 10:36:33.065411 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-kq8lc"] Mar 13 10:36:34 crc kubenswrapper[4632]: I0313 10:36:34.062897 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f916c05-f172-42b6-9b13-0c8d2058bfb1" path="/var/lib/kubelet/pods/8f916c05-f172-42b6-9b13-0c8d2058bfb1/volumes" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.399889 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-cjx7m"] Mar 13 10:36:50 crc kubenswrapper[4632]: E0313 10:36:50.401011 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7cb201b3-b479-4877-a996-58045d0720c4" containerName="oc" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.401029 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7cb201b3-b479-4877-a996-58045d0720c4" containerName="oc" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.401289 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="7cb201b3-b479-4877-a996-58045d0720c4" containerName="oc" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.403064 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.420310 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cjx7m"] Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.424617 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a273fb-d2f3-477f-9c9b-807b65124f71-catalog-content\") pod \"certified-operators-cjx7m\" (UID: \"88a273fb-d2f3-477f-9c9b-807b65124f71\") " pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.424683 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jw2k4\" (UniqueName: \"kubernetes.io/projected/88a273fb-d2f3-477f-9c9b-807b65124f71-kube-api-access-jw2k4\") pod \"certified-operators-cjx7m\" (UID: \"88a273fb-d2f3-477f-9c9b-807b65124f71\") " pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.424716 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a273fb-d2f3-477f-9c9b-807b65124f71-utilities\") pod \"certified-operators-cjx7m\" (UID: \"88a273fb-d2f3-477f-9c9b-807b65124f71\") " pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.526230 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a273fb-d2f3-477f-9c9b-807b65124f71-catalog-content\") pod \"certified-operators-cjx7m\" (UID: \"88a273fb-d2f3-477f-9c9b-807b65124f71\") " pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.526541 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jw2k4\" (UniqueName: \"kubernetes.io/projected/88a273fb-d2f3-477f-9c9b-807b65124f71-kube-api-access-jw2k4\") pod \"certified-operators-cjx7m\" (UID: \"88a273fb-d2f3-477f-9c9b-807b65124f71\") " pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.526723 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a273fb-d2f3-477f-9c9b-807b65124f71-utilities\") pod \"certified-operators-cjx7m\" (UID: \"88a273fb-d2f3-477f-9c9b-807b65124f71\") " pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.526802 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a273fb-d2f3-477f-9c9b-807b65124f71-catalog-content\") pod \"certified-operators-cjx7m\" (UID: \"88a273fb-d2f3-477f-9c9b-807b65124f71\") " pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.527219 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a273fb-d2f3-477f-9c9b-807b65124f71-utilities\") pod \"certified-operators-cjx7m\" (UID: \"88a273fb-d2f3-477f-9c9b-807b65124f71\") " pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.555877 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jw2k4\" (UniqueName: \"kubernetes.io/projected/88a273fb-d2f3-477f-9c9b-807b65124f71-kube-api-access-jw2k4\") pod \"certified-operators-cjx7m\" (UID: \"88a273fb-d2f3-477f-9c9b-807b65124f71\") " pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.595709 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-rpx2w"] Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.598178 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.616723 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rpx2w"] Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.628490 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f03f5b-b62a-4142-9594-79c6ea30f9e2-utilities\") pod \"redhat-marketplace-rpx2w\" (UID: \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\") " pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.628838 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ps6l\" (UniqueName: \"kubernetes.io/projected/27f03f5b-b62a-4142-9594-79c6ea30f9e2-kube-api-access-4ps6l\") pod \"redhat-marketplace-rpx2w\" (UID: \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\") " pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.629016 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f03f5b-b62a-4142-9594-79c6ea30f9e2-catalog-content\") pod \"redhat-marketplace-rpx2w\" (UID: \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\") " pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.730115 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.745308 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ps6l\" (UniqueName: \"kubernetes.io/projected/27f03f5b-b62a-4142-9594-79c6ea30f9e2-kube-api-access-4ps6l\") pod \"redhat-marketplace-rpx2w\" (UID: \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\") " pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.745497 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f03f5b-b62a-4142-9594-79c6ea30f9e2-catalog-content\") pod \"redhat-marketplace-rpx2w\" (UID: \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\") " pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.745605 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f03f5b-b62a-4142-9594-79c6ea30f9e2-utilities\") pod \"redhat-marketplace-rpx2w\" (UID: \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\") " pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.746066 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f03f5b-b62a-4142-9594-79c6ea30f9e2-catalog-content\") pod \"redhat-marketplace-rpx2w\" (UID: \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\") " pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.746214 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f03f5b-b62a-4142-9594-79c6ea30f9e2-utilities\") pod \"redhat-marketplace-rpx2w\" (UID: \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\") " pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.780620 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ps6l\" (UniqueName: \"kubernetes.io/projected/27f03f5b-b62a-4142-9594-79c6ea30f9e2-kube-api-access-4ps6l\") pod \"redhat-marketplace-rpx2w\" (UID: \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\") " pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:36:50 crc kubenswrapper[4632]: I0313 10:36:50.952120 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:36:51 crc kubenswrapper[4632]: I0313 10:36:51.237075 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-cjx7m"] Mar 13 10:36:51 crc kubenswrapper[4632]: I0313 10:36:51.447760 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjx7m" event={"ID":"88a273fb-d2f3-477f-9c9b-807b65124f71","Type":"ContainerStarted","Data":"5ad059c559e099a30b0bde8f6cfb84fc66c8ba7e893d883c696d37b72a5f0e91"} Mar 13 10:36:51 crc kubenswrapper[4632]: I0313 10:36:51.487369 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-rpx2w"] Mar 13 10:36:51 crc kubenswrapper[4632]: W0313 10:36:51.538828 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27f03f5b_b62a_4142_9594_79c6ea30f9e2.slice/crio-bd872c1ca4d35e0f9e248bc64b5943094d58c8ac175f58d64936220c31510187 WatchSource:0}: Error finding container bd872c1ca4d35e0f9e248bc64b5943094d58c8ac175f58d64936220c31510187: Status 404 returned error can't find the container with id bd872c1ca4d35e0f9e248bc64b5943094d58c8ac175f58d64936220c31510187 Mar 13 10:36:52 crc kubenswrapper[4632]: I0313 10:36:52.472576 4632 generic.go:334] "Generic (PLEG): container finished" podID="27f03f5b-b62a-4142-9594-79c6ea30f9e2" containerID="8c073b0c9c22380ad30c5fc7961ef51acdc644797f56ab38d2b96fb0fcea4cdf" exitCode=0 Mar 13 10:36:52 crc kubenswrapper[4632]: I0313 10:36:52.472903 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rpx2w" event={"ID":"27f03f5b-b62a-4142-9594-79c6ea30f9e2","Type":"ContainerDied","Data":"8c073b0c9c22380ad30c5fc7961ef51acdc644797f56ab38d2b96fb0fcea4cdf"} Mar 13 10:36:52 crc kubenswrapper[4632]: I0313 10:36:52.472931 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rpx2w" event={"ID":"27f03f5b-b62a-4142-9594-79c6ea30f9e2","Type":"ContainerStarted","Data":"bd872c1ca4d35e0f9e248bc64b5943094d58c8ac175f58d64936220c31510187"} Mar 13 10:36:52 crc kubenswrapper[4632]: I0313 10:36:52.476177 4632 generic.go:334] "Generic (PLEG): container finished" podID="88a273fb-d2f3-477f-9c9b-807b65124f71" containerID="7eb7750a4f5485b1996c42446b4311cb486ceb4d805041caa4edb556018bf415" exitCode=0 Mar 13 10:36:52 crc kubenswrapper[4632]: I0313 10:36:52.476218 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjx7m" event={"ID":"88a273fb-d2f3-477f-9c9b-807b65124f71","Type":"ContainerDied","Data":"7eb7750a4f5485b1996c42446b4311cb486ceb4d805041caa4edb556018bf415"} Mar 13 10:36:53 crc kubenswrapper[4632]: I0313 10:36:53.487836 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjx7m" event={"ID":"88a273fb-d2f3-477f-9c9b-807b65124f71","Type":"ContainerStarted","Data":"c0feac2c5cf2c2fe35bb95c87df88b965616af6218f4c079105180daf90200ee"} Mar 13 10:36:53 crc kubenswrapper[4632]: I0313 10:36:53.492563 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rpx2w" event={"ID":"27f03f5b-b62a-4142-9594-79c6ea30f9e2","Type":"ContainerStarted","Data":"6db39b0a3dfdd27faaf913cbea3cd500e63cfb09859815521c37cd4347e70f19"} Mar 13 10:36:55 crc kubenswrapper[4632]: I0313 10:36:55.513408 4632 generic.go:334] "Generic (PLEG): container finished" podID="27f03f5b-b62a-4142-9594-79c6ea30f9e2" containerID="6db39b0a3dfdd27faaf913cbea3cd500e63cfb09859815521c37cd4347e70f19" exitCode=0 Mar 13 10:36:55 crc kubenswrapper[4632]: I0313 10:36:55.513532 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rpx2w" event={"ID":"27f03f5b-b62a-4142-9594-79c6ea30f9e2","Type":"ContainerDied","Data":"6db39b0a3dfdd27faaf913cbea3cd500e63cfb09859815521c37cd4347e70f19"} Mar 13 10:36:56 crc kubenswrapper[4632]: I0313 10:36:56.525663 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rpx2w" event={"ID":"27f03f5b-b62a-4142-9594-79c6ea30f9e2","Type":"ContainerStarted","Data":"e02267f2e4b4bd2ba62fd1e078a850349c8f601cd41e32ffd2eca3037d604627"} Mar 13 10:36:57 crc kubenswrapper[4632]: I0313 10:36:57.556310 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-rpx2w" podStartSLOduration=3.907886684 podStartE2EDuration="7.556293815s" podCreationTimestamp="2026-03-13 10:36:50 +0000 UTC" firstStartedPulling="2026-03-13 10:36:52.474838153 +0000 UTC m=+1986.497368286" lastFinishedPulling="2026-03-13 10:36:56.123245284 +0000 UTC m=+1990.145775417" observedRunningTime="2026-03-13 10:36:57.550470872 +0000 UTC m=+1991.573001005" watchObservedRunningTime="2026-03-13 10:36:57.556293815 +0000 UTC m=+1991.578823948" Mar 13 10:36:58 crc kubenswrapper[4632]: I0313 10:36:58.548675 4632 generic.go:334] "Generic (PLEG): container finished" podID="88a273fb-d2f3-477f-9c9b-807b65124f71" containerID="c0feac2c5cf2c2fe35bb95c87df88b965616af6218f4c079105180daf90200ee" exitCode=0 Mar 13 10:36:58 crc kubenswrapper[4632]: I0313 10:36:58.548743 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjx7m" event={"ID":"88a273fb-d2f3-477f-9c9b-807b65124f71","Type":"ContainerDied","Data":"c0feac2c5cf2c2fe35bb95c87df88b965616af6218f4c079105180daf90200ee"} Mar 13 10:36:59 crc kubenswrapper[4632]: I0313 10:36:59.564703 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjx7m" event={"ID":"88a273fb-d2f3-477f-9c9b-807b65124f71","Type":"ContainerStarted","Data":"d70e1caa95eb14a2494c9054b381b3b987b202fd909d5550b73c7aa627b50a73"} Mar 13 10:36:59 crc kubenswrapper[4632]: I0313 10:36:59.606754 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-cjx7m" podStartSLOduration=3.051961518 podStartE2EDuration="9.606713857s" podCreationTimestamp="2026-03-13 10:36:50 +0000 UTC" firstStartedPulling="2026-03-13 10:36:52.478239017 +0000 UTC m=+1986.500769150" lastFinishedPulling="2026-03-13 10:36:59.032991356 +0000 UTC m=+1993.055521489" observedRunningTime="2026-03-13 10:36:59.594524846 +0000 UTC m=+1993.617054979" watchObservedRunningTime="2026-03-13 10:36:59.606713857 +0000 UTC m=+1993.629243990" Mar 13 10:37:00 crc kubenswrapper[4632]: I0313 10:37:00.730514 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:37:00 crc kubenswrapper[4632]: I0313 10:37:00.730573 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:37:00 crc kubenswrapper[4632]: I0313 10:37:00.952953 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:37:00 crc kubenswrapper[4632]: I0313 10:37:00.953015 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:37:01 crc kubenswrapper[4632]: I0313 10:37:01.789197 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-cjx7m" podUID="88a273fb-d2f3-477f-9c9b-807b65124f71" containerName="registry-server" probeResult="failure" output=< Mar 13 10:37:01 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:37:01 crc kubenswrapper[4632]: > Mar 13 10:37:02 crc kubenswrapper[4632]: I0313 10:37:02.014748 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-rpx2w" podUID="27f03f5b-b62a-4142-9594-79c6ea30f9e2" containerName="registry-server" probeResult="failure" output=< Mar 13 10:37:02 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:37:02 crc kubenswrapper[4632]: > Mar 13 10:37:08 crc kubenswrapper[4632]: I0313 10:37:07.999655 4632 scope.go:117] "RemoveContainer" containerID="68a82ec143a93c9f66b6d5e73e70ead182bba11acadf06a0bc0700ee8971357d" Mar 13 10:37:08 crc kubenswrapper[4632]: I0313 10:37:08.056127 4632 scope.go:117] "RemoveContainer" containerID="90dfbecc999c31c0a51b0624874627a8f3c0659cb11e205820b8e9aab659a4a1" Mar 13 10:37:08 crc kubenswrapper[4632]: I0313 10:37:08.098732 4632 scope.go:117] "RemoveContainer" containerID="3b5385b113397b9418c59a941d2a27f232c7b0df4b245db65886e55380c57297" Mar 13 10:37:08 crc kubenswrapper[4632]: I0313 10:37:08.146262 4632 scope.go:117] "RemoveContainer" containerID="3672f721f5cc963fe48f19a0fe26275ae0f1cbd82fd44ed2d6b14dcbb240be1d" Mar 13 10:37:08 crc kubenswrapper[4632]: I0313 10:37:08.222322 4632 scope.go:117] "RemoveContainer" containerID="3ef3ce34ce4d2a0d8d000d31874aca20b10c953ddde87f68a0b04979e69b8bae" Mar 13 10:37:08 crc kubenswrapper[4632]: I0313 10:37:08.281309 4632 scope.go:117] "RemoveContainer" containerID="6d5ac5d7a6aab5517e4300c2e14808710d4f8cfa4977c9841f6552b262144012" Mar 13 10:37:08 crc kubenswrapper[4632]: I0313 10:37:08.332609 4632 scope.go:117] "RemoveContainer" containerID="1e8d2b5aecd08236cabb2c50425d69df7147e32b58dae758550f96994f27f434" Mar 13 10:37:10 crc kubenswrapper[4632]: I0313 10:37:10.786440 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:37:10 crc kubenswrapper[4632]: I0313 10:37:10.846535 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:37:11 crc kubenswrapper[4632]: I0313 10:37:11.003543 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:37:11 crc kubenswrapper[4632]: I0313 10:37:11.025400 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cjx7m"] Mar 13 10:37:11 crc kubenswrapper[4632]: I0313 10:37:11.060418 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:37:12 crc kubenswrapper[4632]: I0313 10:37:12.680021 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-cjx7m" podUID="88a273fb-d2f3-477f-9c9b-807b65124f71" containerName="registry-server" containerID="cri-o://d70e1caa95eb14a2494c9054b381b3b987b202fd909d5550b73c7aa627b50a73" gracePeriod=2 Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.174422 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.243609 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a273fb-d2f3-477f-9c9b-807b65124f71-utilities\") pod \"88a273fb-d2f3-477f-9c9b-807b65124f71\" (UID: \"88a273fb-d2f3-477f-9c9b-807b65124f71\") " Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.243856 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a273fb-d2f3-477f-9c9b-807b65124f71-catalog-content\") pod \"88a273fb-d2f3-477f-9c9b-807b65124f71\" (UID: \"88a273fb-d2f3-477f-9c9b-807b65124f71\") " Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.243934 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jw2k4\" (UniqueName: \"kubernetes.io/projected/88a273fb-d2f3-477f-9c9b-807b65124f71-kube-api-access-jw2k4\") pod \"88a273fb-d2f3-477f-9c9b-807b65124f71\" (UID: \"88a273fb-d2f3-477f-9c9b-807b65124f71\") " Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.244730 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88a273fb-d2f3-477f-9c9b-807b65124f71-utilities" (OuterVolumeSpecName: "utilities") pod "88a273fb-d2f3-477f-9c9b-807b65124f71" (UID: "88a273fb-d2f3-477f-9c9b-807b65124f71"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.251363 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88a273fb-d2f3-477f-9c9b-807b65124f71-kube-api-access-jw2k4" (OuterVolumeSpecName: "kube-api-access-jw2k4") pod "88a273fb-d2f3-477f-9c9b-807b65124f71" (UID: "88a273fb-d2f3-477f-9c9b-807b65124f71"). InnerVolumeSpecName "kube-api-access-jw2k4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.303683 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/88a273fb-d2f3-477f-9c9b-807b65124f71-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "88a273fb-d2f3-477f-9c9b-807b65124f71" (UID: "88a273fb-d2f3-477f-9c9b-807b65124f71"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.346401 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/88a273fb-d2f3-477f-9c9b-807b65124f71-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.346648 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/88a273fb-d2f3-477f-9c9b-807b65124f71-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.346744 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jw2k4\" (UniqueName: \"kubernetes.io/projected/88a273fb-d2f3-477f-9c9b-807b65124f71-kube-api-access-jw2k4\") on node \"crc\" DevicePath \"\"" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.425234 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rpx2w"] Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.425810 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-rpx2w" podUID="27f03f5b-b62a-4142-9594-79c6ea30f9e2" containerName="registry-server" containerID="cri-o://e02267f2e4b4bd2ba62fd1e078a850349c8f601cd41e32ffd2eca3037d604627" gracePeriod=2 Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.712427 4632 generic.go:334] "Generic (PLEG): container finished" podID="88a273fb-d2f3-477f-9c9b-807b65124f71" containerID="d70e1caa95eb14a2494c9054b381b3b987b202fd909d5550b73c7aa627b50a73" exitCode=0 Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.712470 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjx7m" event={"ID":"88a273fb-d2f3-477f-9c9b-807b65124f71","Type":"ContainerDied","Data":"d70e1caa95eb14a2494c9054b381b3b987b202fd909d5550b73c7aa627b50a73"} Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.712503 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-cjx7m" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.712520 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-cjx7m" event={"ID":"88a273fb-d2f3-477f-9c9b-807b65124f71","Type":"ContainerDied","Data":"5ad059c559e099a30b0bde8f6cfb84fc66c8ba7e893d883c696d37b72a5f0e91"} Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.712543 4632 scope.go:117] "RemoveContainer" containerID="d70e1caa95eb14a2494c9054b381b3b987b202fd909d5550b73c7aa627b50a73" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.734196 4632 generic.go:334] "Generic (PLEG): container finished" podID="27f03f5b-b62a-4142-9594-79c6ea30f9e2" containerID="e02267f2e4b4bd2ba62fd1e078a850349c8f601cd41e32ffd2eca3037d604627" exitCode=0 Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.734236 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rpx2w" event={"ID":"27f03f5b-b62a-4142-9594-79c6ea30f9e2","Type":"ContainerDied","Data":"e02267f2e4b4bd2ba62fd1e078a850349c8f601cd41e32ffd2eca3037d604627"} Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.770902 4632 scope.go:117] "RemoveContainer" containerID="c0feac2c5cf2c2fe35bb95c87df88b965616af6218f4c079105180daf90200ee" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.778203 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-cjx7m"] Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.791761 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-cjx7m"] Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.809370 4632 scope.go:117] "RemoveContainer" containerID="7eb7750a4f5485b1996c42446b4311cb486ceb4d805041caa4edb556018bf415" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.911575 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.921565 4632 scope.go:117] "RemoveContainer" containerID="d70e1caa95eb14a2494c9054b381b3b987b202fd909d5550b73c7aa627b50a73" Mar 13 10:37:13 crc kubenswrapper[4632]: E0313 10:37:13.922411 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d70e1caa95eb14a2494c9054b381b3b987b202fd909d5550b73c7aa627b50a73\": container with ID starting with d70e1caa95eb14a2494c9054b381b3b987b202fd909d5550b73c7aa627b50a73 not found: ID does not exist" containerID="d70e1caa95eb14a2494c9054b381b3b987b202fd909d5550b73c7aa627b50a73" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.922451 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d70e1caa95eb14a2494c9054b381b3b987b202fd909d5550b73c7aa627b50a73"} err="failed to get container status \"d70e1caa95eb14a2494c9054b381b3b987b202fd909d5550b73c7aa627b50a73\": rpc error: code = NotFound desc = could not find container \"d70e1caa95eb14a2494c9054b381b3b987b202fd909d5550b73c7aa627b50a73\": container with ID starting with d70e1caa95eb14a2494c9054b381b3b987b202fd909d5550b73c7aa627b50a73 not found: ID does not exist" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.922614 4632 scope.go:117] "RemoveContainer" containerID="c0feac2c5cf2c2fe35bb95c87df88b965616af6218f4c079105180daf90200ee" Mar 13 10:37:13 crc kubenswrapper[4632]: E0313 10:37:13.923138 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0feac2c5cf2c2fe35bb95c87df88b965616af6218f4c079105180daf90200ee\": container with ID starting with c0feac2c5cf2c2fe35bb95c87df88b965616af6218f4c079105180daf90200ee not found: ID does not exist" containerID="c0feac2c5cf2c2fe35bb95c87df88b965616af6218f4c079105180daf90200ee" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.923194 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0feac2c5cf2c2fe35bb95c87df88b965616af6218f4c079105180daf90200ee"} err="failed to get container status \"c0feac2c5cf2c2fe35bb95c87df88b965616af6218f4c079105180daf90200ee\": rpc error: code = NotFound desc = could not find container \"c0feac2c5cf2c2fe35bb95c87df88b965616af6218f4c079105180daf90200ee\": container with ID starting with c0feac2c5cf2c2fe35bb95c87df88b965616af6218f4c079105180daf90200ee not found: ID does not exist" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.923228 4632 scope.go:117] "RemoveContainer" containerID="7eb7750a4f5485b1996c42446b4311cb486ceb4d805041caa4edb556018bf415" Mar 13 10:37:13 crc kubenswrapper[4632]: E0313 10:37:13.923565 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7eb7750a4f5485b1996c42446b4311cb486ceb4d805041caa4edb556018bf415\": container with ID starting with 7eb7750a4f5485b1996c42446b4311cb486ceb4d805041caa4edb556018bf415 not found: ID does not exist" containerID="7eb7750a4f5485b1996c42446b4311cb486ceb4d805041caa4edb556018bf415" Mar 13 10:37:13 crc kubenswrapper[4632]: I0313 10:37:13.923595 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7eb7750a4f5485b1996c42446b4311cb486ceb4d805041caa4edb556018bf415"} err="failed to get container status \"7eb7750a4f5485b1996c42446b4311cb486ceb4d805041caa4edb556018bf415\": rpc error: code = NotFound desc = could not find container \"7eb7750a4f5485b1996c42446b4311cb486ceb4d805041caa4edb556018bf415\": container with ID starting with 7eb7750a4f5485b1996c42446b4311cb486ceb4d805041caa4edb556018bf415 not found: ID does not exist" Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.053991 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88a273fb-d2f3-477f-9c9b-807b65124f71" path="/var/lib/kubelet/pods/88a273fb-d2f3-477f-9c9b-807b65124f71/volumes" Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.065508 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f03f5b-b62a-4142-9594-79c6ea30f9e2-catalog-content\") pod \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\" (UID: \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\") " Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.065576 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ps6l\" (UniqueName: \"kubernetes.io/projected/27f03f5b-b62a-4142-9594-79c6ea30f9e2-kube-api-access-4ps6l\") pod \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\" (UID: \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\") " Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.065810 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f03f5b-b62a-4142-9594-79c6ea30f9e2-utilities\") pod \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\" (UID: \"27f03f5b-b62a-4142-9594-79c6ea30f9e2\") " Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.066811 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27f03f5b-b62a-4142-9594-79c6ea30f9e2-utilities" (OuterVolumeSpecName: "utilities") pod "27f03f5b-b62a-4142-9594-79c6ea30f9e2" (UID: "27f03f5b-b62a-4142-9594-79c6ea30f9e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.071899 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27f03f5b-b62a-4142-9594-79c6ea30f9e2-kube-api-access-4ps6l" (OuterVolumeSpecName: "kube-api-access-4ps6l") pod "27f03f5b-b62a-4142-9594-79c6ea30f9e2" (UID: "27f03f5b-b62a-4142-9594-79c6ea30f9e2"). InnerVolumeSpecName "kube-api-access-4ps6l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.094926 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27f03f5b-b62a-4142-9594-79c6ea30f9e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27f03f5b-b62a-4142-9594-79c6ea30f9e2" (UID: "27f03f5b-b62a-4142-9594-79c6ea30f9e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.167835 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27f03f5b-b62a-4142-9594-79c6ea30f9e2-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.168131 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27f03f5b-b62a-4142-9594-79c6ea30f9e2-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.168205 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ps6l\" (UniqueName: \"kubernetes.io/projected/27f03f5b-b62a-4142-9594-79c6ea30f9e2-kube-api-access-4ps6l\") on node \"crc\" DevicePath \"\"" Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.745493 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-rpx2w" Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.745506 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-rpx2w" event={"ID":"27f03f5b-b62a-4142-9594-79c6ea30f9e2","Type":"ContainerDied","Data":"bd872c1ca4d35e0f9e248bc64b5943094d58c8ac175f58d64936220c31510187"} Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.746782 4632 scope.go:117] "RemoveContainer" containerID="e02267f2e4b4bd2ba62fd1e078a850349c8f601cd41e32ffd2eca3037d604627" Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.782233 4632 scope.go:117] "RemoveContainer" containerID="6db39b0a3dfdd27faaf913cbea3cd500e63cfb09859815521c37cd4347e70f19" Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.784642 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-rpx2w"] Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.802637 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-rpx2w"] Mar 13 10:37:14 crc kubenswrapper[4632]: I0313 10:37:14.807488 4632 scope.go:117] "RemoveContainer" containerID="8c073b0c9c22380ad30c5fc7961ef51acdc644797f56ab38d2b96fb0fcea4cdf" Mar 13 10:37:16 crc kubenswrapper[4632]: I0313 10:37:16.058036 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27f03f5b-b62a-4142-9594-79c6ea30f9e2" path="/var/lib/kubelet/pods/27f03f5b-b62a-4142-9594-79c6ea30f9e2/volumes" Mar 13 10:37:28 crc kubenswrapper[4632]: I0313 10:37:28.059267 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-fshjb"] Mar 13 10:37:28 crc kubenswrapper[4632]: I0313 10:37:28.073996 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-kswhw"] Mar 13 10:37:28 crc kubenswrapper[4632]: I0313 10:37:28.082432 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-fshjb"] Mar 13 10:37:28 crc kubenswrapper[4632]: I0313 10:37:28.090961 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-kswhw"] Mar 13 10:37:30 crc kubenswrapper[4632]: I0313 10:37:30.038048 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-2f8c-account-create-update-g4b8g"] Mar 13 10:37:30 crc kubenswrapper[4632]: I0313 10:37:30.057282 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09bd98be-9d10-4a53-8ef6-c4718b05c3f6" path="/var/lib/kubelet/pods/09bd98be-9d10-4a53-8ef6-c4718b05c3f6/volumes" Mar 13 10:37:30 crc kubenswrapper[4632]: I0313 10:37:30.059280 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e0fb1fc-c94a-44f0-a269-e7211c6fcfba" path="/var/lib/kubelet/pods/8e0fb1fc-c94a-44f0-a269-e7211c6fcfba/volumes" Mar 13 10:37:30 crc kubenswrapper[4632]: I0313 10:37:30.060188 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-wgv42"] Mar 13 10:37:30 crc kubenswrapper[4632]: I0313 10:37:30.064615 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-f3f1-account-create-update-29g8s"] Mar 13 10:37:30 crc kubenswrapper[4632]: I0313 10:37:30.070741 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-2f8c-account-create-update-g4b8g"] Mar 13 10:37:30 crc kubenswrapper[4632]: I0313 10:37:30.078485 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-f3f1-account-create-update-29g8s"] Mar 13 10:37:30 crc kubenswrapper[4632]: I0313 10:37:30.085799 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-wgv42"] Mar 13 10:37:31 crc kubenswrapper[4632]: I0313 10:37:31.036759 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-86d4-account-create-update-5c7rj"] Mar 13 10:37:31 crc kubenswrapper[4632]: I0313 10:37:31.045981 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-86d4-account-create-update-5c7rj"] Mar 13 10:37:32 crc kubenswrapper[4632]: I0313 10:37:32.055656 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="234a900d-887b-448c-8336-010107726c1e" path="/var/lib/kubelet/pods/234a900d-887b-448c-8336-010107726c1e/volumes" Mar 13 10:37:32 crc kubenswrapper[4632]: I0313 10:37:32.057574 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8462be25-a577-476d-b54a-73790a8aa189" path="/var/lib/kubelet/pods/8462be25-a577-476d-b54a-73790a8aa189/volumes" Mar 13 10:37:32 crc kubenswrapper[4632]: I0313 10:37:32.058512 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bbaf5a79-1c34-4518-afb9-19703fe6c45b" path="/var/lib/kubelet/pods/bbaf5a79-1c34-4518-afb9-19703fe6c45b/volumes" Mar 13 10:37:32 crc kubenswrapper[4632]: I0313 10:37:32.059321 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0c32ed5-c3b0-45ea-99de-87c45cb1ba77" path="/var/lib/kubelet/pods/f0c32ed5-c3b0-45ea-99de-87c45cb1ba77/volumes" Mar 13 10:37:33 crc kubenswrapper[4632]: I0313 10:37:33.920301 4632 generic.go:334] "Generic (PLEG): container finished" podID="0d75181a-4c91-485e-8bcd-02e2aedd4d45" containerID="bd09862a5fc80def82e97b44b7d539caee7c696bb410023ec19cde3384abb6ae" exitCode=0 Mar 13 10:37:33 crc kubenswrapper[4632]: I0313 10:37:33.920381 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" event={"ID":"0d75181a-4c91-485e-8bcd-02e2aedd4d45","Type":"ContainerDied","Data":"bd09862a5fc80def82e97b44b7d539caee7c696bb410023ec19cde3384abb6ae"} Mar 13 10:37:35 crc kubenswrapper[4632]: I0313 10:37:35.379388 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" Mar 13 10:37:35 crc kubenswrapper[4632]: I0313 10:37:35.567665 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d75181a-4c91-485e-8bcd-02e2aedd4d45-ssh-key-openstack-edpm-ipam\") pod \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\" (UID: \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\") " Mar 13 10:37:35 crc kubenswrapper[4632]: I0313 10:37:35.569066 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qctqc\" (UniqueName: \"kubernetes.io/projected/0d75181a-4c91-485e-8bcd-02e2aedd4d45-kube-api-access-qctqc\") pod \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\" (UID: \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\") " Mar 13 10:37:35 crc kubenswrapper[4632]: I0313 10:37:35.569232 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d75181a-4c91-485e-8bcd-02e2aedd4d45-inventory\") pod \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\" (UID: \"0d75181a-4c91-485e-8bcd-02e2aedd4d45\") " Mar 13 10:37:35 crc kubenswrapper[4632]: I0313 10:37:35.576019 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d75181a-4c91-485e-8bcd-02e2aedd4d45-kube-api-access-qctqc" (OuterVolumeSpecName: "kube-api-access-qctqc") pod "0d75181a-4c91-485e-8bcd-02e2aedd4d45" (UID: "0d75181a-4c91-485e-8bcd-02e2aedd4d45"). InnerVolumeSpecName "kube-api-access-qctqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:37:35 crc kubenswrapper[4632]: I0313 10:37:35.601767 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d75181a-4c91-485e-8bcd-02e2aedd4d45-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0d75181a-4c91-485e-8bcd-02e2aedd4d45" (UID: "0d75181a-4c91-485e-8bcd-02e2aedd4d45"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:37:35 crc kubenswrapper[4632]: I0313 10:37:35.612627 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d75181a-4c91-485e-8bcd-02e2aedd4d45-inventory" (OuterVolumeSpecName: "inventory") pod "0d75181a-4c91-485e-8bcd-02e2aedd4d45" (UID: "0d75181a-4c91-485e-8bcd-02e2aedd4d45"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:37:35 crc kubenswrapper[4632]: I0313 10:37:35.671386 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d75181a-4c91-485e-8bcd-02e2aedd4d45-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:37:35 crc kubenswrapper[4632]: I0313 10:37:35.671626 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d75181a-4c91-485e-8bcd-02e2aedd4d45-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:37:35 crc kubenswrapper[4632]: I0313 10:37:35.671709 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qctqc\" (UniqueName: \"kubernetes.io/projected/0d75181a-4c91-485e-8bcd-02e2aedd4d45-kube-api-access-qctqc\") on node \"crc\" DevicePath \"\"" Mar 13 10:37:35 crc kubenswrapper[4632]: I0313 10:37:35.942653 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" event={"ID":"0d75181a-4c91-485e-8bcd-02e2aedd4d45","Type":"ContainerDied","Data":"b2534ff7153cbd23578a95b816373371c87f8d95ec5b00bff35a4aeb9a12cb51"} Mar 13 10:37:35 crc kubenswrapper[4632]: I0313 10:37:35.942694 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b2534ff7153cbd23578a95b816373371c87f8d95ec5b00bff35a4aeb9a12cb51" Mar 13 10:37:35 crc kubenswrapper[4632]: I0313 10:37:35.942700 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-754cp" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.042352 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84"] Mar 13 10:37:36 crc kubenswrapper[4632]: E0313 10:37:36.043056 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f03f5b-b62a-4142-9594-79c6ea30f9e2" containerName="extract-content" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.043074 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f03f5b-b62a-4142-9594-79c6ea30f9e2" containerName="extract-content" Mar 13 10:37:36 crc kubenswrapper[4632]: E0313 10:37:36.043094 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f03f5b-b62a-4142-9594-79c6ea30f9e2" containerName="extract-utilities" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.043101 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f03f5b-b62a-4142-9594-79c6ea30f9e2" containerName="extract-utilities" Mar 13 10:37:36 crc kubenswrapper[4632]: E0313 10:37:36.043112 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a273fb-d2f3-477f-9c9b-807b65124f71" containerName="extract-utilities" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.043119 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a273fb-d2f3-477f-9c9b-807b65124f71" containerName="extract-utilities" Mar 13 10:37:36 crc kubenswrapper[4632]: E0313 10:37:36.043137 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a273fb-d2f3-477f-9c9b-807b65124f71" containerName="registry-server" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.043143 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a273fb-d2f3-477f-9c9b-807b65124f71" containerName="registry-server" Mar 13 10:37:36 crc kubenswrapper[4632]: E0313 10:37:36.043159 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d75181a-4c91-485e-8bcd-02e2aedd4d45" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.043166 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d75181a-4c91-485e-8bcd-02e2aedd4d45" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 13 10:37:36 crc kubenswrapper[4632]: E0313 10:37:36.043176 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a273fb-d2f3-477f-9c9b-807b65124f71" containerName="extract-content" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.043182 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a273fb-d2f3-477f-9c9b-807b65124f71" containerName="extract-content" Mar 13 10:37:36 crc kubenswrapper[4632]: E0313 10:37:36.043194 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27f03f5b-b62a-4142-9594-79c6ea30f9e2" containerName="registry-server" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.043200 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="27f03f5b-b62a-4142-9594-79c6ea30f9e2" containerName="registry-server" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.043375 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="88a273fb-d2f3-477f-9c9b-807b65124f71" containerName="registry-server" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.043400 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="27f03f5b-b62a-4142-9594-79c6ea30f9e2" containerName="registry-server" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.043411 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d75181a-4c91-485e-8bcd-02e2aedd4d45" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.044182 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.047242 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.047528 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.047686 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.047828 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.075794 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84"] Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.181439 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bcd0e6df-81c2-4541-b0b5-d5c539f03451-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tpk84\" (UID: \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.181556 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6hxn\" (UniqueName: \"kubernetes.io/projected/bcd0e6df-81c2-4541-b0b5-d5c539f03451-kube-api-access-x6hxn\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tpk84\" (UID: \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.181649 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcd0e6df-81c2-4541-b0b5-d5c539f03451-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tpk84\" (UID: \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.282853 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6hxn\" (UniqueName: \"kubernetes.io/projected/bcd0e6df-81c2-4541-b0b5-d5c539f03451-kube-api-access-x6hxn\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tpk84\" (UID: \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.282992 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcd0e6df-81c2-4541-b0b5-d5c539f03451-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tpk84\" (UID: \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.283068 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bcd0e6df-81c2-4541-b0b5-d5c539f03451-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tpk84\" (UID: \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.288758 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bcd0e6df-81c2-4541-b0b5-d5c539f03451-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tpk84\" (UID: \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.289554 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcd0e6df-81c2-4541-b0b5-d5c539f03451-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tpk84\" (UID: \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.303163 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6hxn\" (UniqueName: \"kubernetes.io/projected/bcd0e6df-81c2-4541-b0b5-d5c539f03451-kube-api-access-x6hxn\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-tpk84\" (UID: \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.371791 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" Mar 13 10:37:36 crc kubenswrapper[4632]: I0313 10:37:36.978376 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84"] Mar 13 10:37:37 crc kubenswrapper[4632]: I0313 10:37:37.958831 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" event={"ID":"bcd0e6df-81c2-4541-b0b5-d5c539f03451","Type":"ContainerStarted","Data":"da621c5c44f364510de28883c64cc52b63ab77c54d882aedd7a7119edf3055a8"} Mar 13 10:37:37 crc kubenswrapper[4632]: I0313 10:37:37.959164 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" event={"ID":"bcd0e6df-81c2-4541-b0b5-d5c539f03451","Type":"ContainerStarted","Data":"d035c3103945937690807050ad841d4bce39b39f6d04cfc72ad0d28a843a4add"} Mar 13 10:37:37 crc kubenswrapper[4632]: I0313 10:37:37.976762 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" podStartSLOduration=1.553680848 podStartE2EDuration="1.976746375s" podCreationTimestamp="2026-03-13 10:37:36 +0000 UTC" firstStartedPulling="2026-03-13 10:37:36.992029105 +0000 UTC m=+2031.014559238" lastFinishedPulling="2026-03-13 10:37:37.415094632 +0000 UTC m=+2031.437624765" observedRunningTime="2026-03-13 10:37:37.973281499 +0000 UTC m=+2031.995811652" watchObservedRunningTime="2026-03-13 10:37:37.976746375 +0000 UTC m=+2031.999276508" Mar 13 10:38:00 crc kubenswrapper[4632]: I0313 10:38:00.150904 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556638-p7mdh"] Mar 13 10:38:00 crc kubenswrapper[4632]: I0313 10:38:00.156809 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556638-p7mdh" Mar 13 10:38:00 crc kubenswrapper[4632]: I0313 10:38:00.161984 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:38:00 crc kubenswrapper[4632]: I0313 10:38:00.162076 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:38:00 crc kubenswrapper[4632]: I0313 10:38:00.162254 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:38:00 crc kubenswrapper[4632]: I0313 10:38:00.180139 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556638-p7mdh"] Mar 13 10:38:00 crc kubenswrapper[4632]: I0313 10:38:00.345845 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzq8p\" (UniqueName: \"kubernetes.io/projected/346e767a-d9dd-40e1-9ab3-2e4ec9184667-kube-api-access-tzq8p\") pod \"auto-csr-approver-29556638-p7mdh\" (UID: \"346e767a-d9dd-40e1-9ab3-2e4ec9184667\") " pod="openshift-infra/auto-csr-approver-29556638-p7mdh" Mar 13 10:38:00 crc kubenswrapper[4632]: I0313 10:38:00.449302 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzq8p\" (UniqueName: \"kubernetes.io/projected/346e767a-d9dd-40e1-9ab3-2e4ec9184667-kube-api-access-tzq8p\") pod \"auto-csr-approver-29556638-p7mdh\" (UID: \"346e767a-d9dd-40e1-9ab3-2e4ec9184667\") " pod="openshift-infra/auto-csr-approver-29556638-p7mdh" Mar 13 10:38:00 crc kubenswrapper[4632]: I0313 10:38:00.489732 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzq8p\" (UniqueName: \"kubernetes.io/projected/346e767a-d9dd-40e1-9ab3-2e4ec9184667-kube-api-access-tzq8p\") pod \"auto-csr-approver-29556638-p7mdh\" (UID: \"346e767a-d9dd-40e1-9ab3-2e4ec9184667\") " pod="openshift-infra/auto-csr-approver-29556638-p7mdh" Mar 13 10:38:00 crc kubenswrapper[4632]: I0313 10:38:00.782907 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556638-p7mdh" Mar 13 10:38:01 crc kubenswrapper[4632]: I0313 10:38:01.272743 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556638-p7mdh"] Mar 13 10:38:02 crc kubenswrapper[4632]: I0313 10:38:02.206813 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556638-p7mdh" event={"ID":"346e767a-d9dd-40e1-9ab3-2e4ec9184667","Type":"ContainerStarted","Data":"d8c187bfe15bf5c2773923c01881f05cc533e70265c96c2fe50269b7b59d185c"} Mar 13 10:38:03 crc kubenswrapper[4632]: I0313 10:38:03.217597 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556638-p7mdh" event={"ID":"346e767a-d9dd-40e1-9ab3-2e4ec9184667","Type":"ContainerStarted","Data":"c166f0a830c16b65f03aba2171bb98a995fe4121f1b92036d629fce2afd52c26"} Mar 13 10:38:03 crc kubenswrapper[4632]: I0313 10:38:03.244140 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556638-p7mdh" podStartSLOduration=2.099335833 podStartE2EDuration="3.244092849s" podCreationTimestamp="2026-03-13 10:38:00 +0000 UTC" firstStartedPulling="2026-03-13 10:38:01.279095088 +0000 UTC m=+2055.301625221" lastFinishedPulling="2026-03-13 10:38:02.423852104 +0000 UTC m=+2056.446382237" observedRunningTime="2026-03-13 10:38:03.231781835 +0000 UTC m=+2057.254311968" watchObservedRunningTime="2026-03-13 10:38:03.244092849 +0000 UTC m=+2057.266622982" Mar 13 10:38:04 crc kubenswrapper[4632]: I0313 10:38:04.228658 4632 generic.go:334] "Generic (PLEG): container finished" podID="346e767a-d9dd-40e1-9ab3-2e4ec9184667" containerID="c166f0a830c16b65f03aba2171bb98a995fe4121f1b92036d629fce2afd52c26" exitCode=0 Mar 13 10:38:04 crc kubenswrapper[4632]: I0313 10:38:04.228713 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556638-p7mdh" event={"ID":"346e767a-d9dd-40e1-9ab3-2e4ec9184667","Type":"ContainerDied","Data":"c166f0a830c16b65f03aba2171bb98a995fe4121f1b92036d629fce2afd52c26"} Mar 13 10:38:05 crc kubenswrapper[4632]: I0313 10:38:05.552924 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556638-p7mdh" Mar 13 10:38:05 crc kubenswrapper[4632]: I0313 10:38:05.585325 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzq8p\" (UniqueName: \"kubernetes.io/projected/346e767a-d9dd-40e1-9ab3-2e4ec9184667-kube-api-access-tzq8p\") pod \"346e767a-d9dd-40e1-9ab3-2e4ec9184667\" (UID: \"346e767a-d9dd-40e1-9ab3-2e4ec9184667\") " Mar 13 10:38:05 crc kubenswrapper[4632]: I0313 10:38:05.598795 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/346e767a-d9dd-40e1-9ab3-2e4ec9184667-kube-api-access-tzq8p" (OuterVolumeSpecName: "kube-api-access-tzq8p") pod "346e767a-d9dd-40e1-9ab3-2e4ec9184667" (UID: "346e767a-d9dd-40e1-9ab3-2e4ec9184667"). InnerVolumeSpecName "kube-api-access-tzq8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:38:05 crc kubenswrapper[4632]: I0313 10:38:05.687806 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzq8p\" (UniqueName: \"kubernetes.io/projected/346e767a-d9dd-40e1-9ab3-2e4ec9184667-kube-api-access-tzq8p\") on node \"crc\" DevicePath \"\"" Mar 13 10:38:06 crc kubenswrapper[4632]: I0313 10:38:06.255431 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556638-p7mdh" event={"ID":"346e767a-d9dd-40e1-9ab3-2e4ec9184667","Type":"ContainerDied","Data":"d8c187bfe15bf5c2773923c01881f05cc533e70265c96c2fe50269b7b59d185c"} Mar 13 10:38:06 crc kubenswrapper[4632]: I0313 10:38:06.255506 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d8c187bfe15bf5c2773923c01881f05cc533e70265c96c2fe50269b7b59d185c" Mar 13 10:38:06 crc kubenswrapper[4632]: I0313 10:38:06.256164 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556638-p7mdh" Mar 13 10:38:06 crc kubenswrapper[4632]: I0313 10:38:06.319417 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556632-sr4l5"] Mar 13 10:38:06 crc kubenswrapper[4632]: I0313 10:38:06.328658 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556632-sr4l5"] Mar 13 10:38:08 crc kubenswrapper[4632]: I0313 10:38:08.059131 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="009f055c-d442-4b23-8f55-52a43362bbb2" path="/var/lib/kubelet/pods/009f055c-d442-4b23-8f55-52a43362bbb2/volumes" Mar 13 10:38:08 crc kubenswrapper[4632]: I0313 10:38:08.490202 4632 scope.go:117] "RemoveContainer" containerID="f531fb1c9798e5386771f799aeaf5ec81a37e70faa215029f1e44845844c0b7a" Mar 13 10:38:08 crc kubenswrapper[4632]: I0313 10:38:08.522183 4632 scope.go:117] "RemoveContainer" containerID="62c66b71b16f2cd37ff478080f4c30eed65f51b807f687725f8ec89f5dd9d0dc" Mar 13 10:38:08 crc kubenswrapper[4632]: I0313 10:38:08.562684 4632 scope.go:117] "RemoveContainer" containerID="d9f2ab5e1a5be1d4939b9fe05ba3a5cdbc725953ea1e78a027cf1f61d4444ba0" Mar 13 10:38:08 crc kubenswrapper[4632]: I0313 10:38:08.614954 4632 scope.go:117] "RemoveContainer" containerID="9cee7abc6c76d73494106b5582f85b871d225f179b8f40700ad2248a8daa7c60" Mar 13 10:38:08 crc kubenswrapper[4632]: I0313 10:38:08.659235 4632 scope.go:117] "RemoveContainer" containerID="4ccfb76824c418f1c761a43ca7732c6a7a69b7b1944ea2ee35bd45c569e7d7c6" Mar 13 10:38:08 crc kubenswrapper[4632]: I0313 10:38:08.731738 4632 scope.go:117] "RemoveContainer" containerID="baa73e1779483e615256cb324392bd7ff43cccd507e79b501108b7a61007ed58" Mar 13 10:38:08 crc kubenswrapper[4632]: I0313 10:38:08.767399 4632 scope.go:117] "RemoveContainer" containerID="a73d11226d1411728675707324588174ab20222ac0a86a31f153adf5c08496b7" Mar 13 10:38:40 crc kubenswrapper[4632]: I0313 10:38:40.460804 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:38:40 crc kubenswrapper[4632]: I0313 10:38:40.461513 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:38:41 crc kubenswrapper[4632]: I0313 10:38:41.064199 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5mlm2"] Mar 13 10:38:41 crc kubenswrapper[4632]: I0313 10:38:41.082577 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-5mlm2"] Mar 13 10:38:42 crc kubenswrapper[4632]: I0313 10:38:42.056354 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5de81924-9bfc-484e-8276-0216f0bbf72c" path="/var/lib/kubelet/pods/5de81924-9bfc-484e-8276-0216f0bbf72c/volumes" Mar 13 10:38:54 crc kubenswrapper[4632]: I0313 10:38:54.005708 4632 generic.go:334] "Generic (PLEG): container finished" podID="bcd0e6df-81c2-4541-b0b5-d5c539f03451" containerID="da621c5c44f364510de28883c64cc52b63ab77c54d882aedd7a7119edf3055a8" exitCode=0 Mar 13 10:38:54 crc kubenswrapper[4632]: I0313 10:38:54.005786 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" event={"ID":"bcd0e6df-81c2-4541-b0b5-d5c539f03451","Type":"ContainerDied","Data":"da621c5c44f364510de28883c64cc52b63ab77c54d882aedd7a7119edf3055a8"} Mar 13 10:38:55 crc kubenswrapper[4632]: I0313 10:38:55.562177 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" Mar 13 10:38:55 crc kubenswrapper[4632]: I0313 10:38:55.728123 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bcd0e6df-81c2-4541-b0b5-d5c539f03451-ssh-key-openstack-edpm-ipam\") pod \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\" (UID: \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\") " Mar 13 10:38:55 crc kubenswrapper[4632]: I0313 10:38:55.728256 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6hxn\" (UniqueName: \"kubernetes.io/projected/bcd0e6df-81c2-4541-b0b5-d5c539f03451-kube-api-access-x6hxn\") pod \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\" (UID: \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\") " Mar 13 10:38:55 crc kubenswrapper[4632]: I0313 10:38:55.728431 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcd0e6df-81c2-4541-b0b5-d5c539f03451-inventory\") pod \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\" (UID: \"bcd0e6df-81c2-4541-b0b5-d5c539f03451\") " Mar 13 10:38:55 crc kubenswrapper[4632]: I0313 10:38:55.764252 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcd0e6df-81c2-4541-b0b5-d5c539f03451-kube-api-access-x6hxn" (OuterVolumeSpecName: "kube-api-access-x6hxn") pod "bcd0e6df-81c2-4541-b0b5-d5c539f03451" (UID: "bcd0e6df-81c2-4541-b0b5-d5c539f03451"). InnerVolumeSpecName "kube-api-access-x6hxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:38:55 crc kubenswrapper[4632]: I0313 10:38:55.810170 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcd0e6df-81c2-4541-b0b5-d5c539f03451-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "bcd0e6df-81c2-4541-b0b5-d5c539f03451" (UID: "bcd0e6df-81c2-4541-b0b5-d5c539f03451"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:38:55 crc kubenswrapper[4632]: I0313 10:38:55.817131 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bcd0e6df-81c2-4541-b0b5-d5c539f03451-inventory" (OuterVolumeSpecName: "inventory") pod "bcd0e6df-81c2-4541-b0b5-d5c539f03451" (UID: "bcd0e6df-81c2-4541-b0b5-d5c539f03451"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:38:55 crc kubenswrapper[4632]: I0313 10:38:55.833224 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bcd0e6df-81c2-4541-b0b5-d5c539f03451-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:38:55 crc kubenswrapper[4632]: I0313 10:38:55.833575 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bcd0e6df-81c2-4541-b0b5-d5c539f03451-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:38:55 crc kubenswrapper[4632]: I0313 10:38:55.833587 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6hxn\" (UniqueName: \"kubernetes.io/projected/bcd0e6df-81c2-4541-b0b5-d5c539f03451-kube-api-access-x6hxn\") on node \"crc\" DevicePath \"\"" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.023456 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" event={"ID":"bcd0e6df-81c2-4541-b0b5-d5c539f03451","Type":"ContainerDied","Data":"d035c3103945937690807050ad841d4bce39b39f6d04cfc72ad0d28a843a4add"} Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.023521 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d035c3103945937690807050ad841d4bce39b39f6d04cfc72ad0d28a843a4add" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.023535 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-tpk84" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.190851 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg"] Mar 13 10:38:56 crc kubenswrapper[4632]: E0313 10:38:56.192406 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcd0e6df-81c2-4541-b0b5-d5c539f03451" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.192455 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcd0e6df-81c2-4541-b0b5-d5c539f03451" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 13 10:38:56 crc kubenswrapper[4632]: E0313 10:38:56.192508 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="346e767a-d9dd-40e1-9ab3-2e4ec9184667" containerName="oc" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.192530 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="346e767a-d9dd-40e1-9ab3-2e4ec9184667" containerName="oc" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.193263 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcd0e6df-81c2-4541-b0b5-d5c539f03451" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.193357 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="346e767a-d9dd-40e1-9ab3-2e4ec9184667" containerName="oc" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.195836 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.203167 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.203300 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.203182 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.204778 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.252070 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddrbk\" (UniqueName: \"kubernetes.io/projected/a1c30ff2-4a23-4fb1-b689-59318014bf57-kube-api-access-ddrbk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg\" (UID: \"a1c30ff2-4a23-4fb1-b689-59318014bf57\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.252561 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1c30ff2-4a23-4fb1-b689-59318014bf57-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg\" (UID: \"a1c30ff2-4a23-4fb1-b689-59318014bf57\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.252971 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1c30ff2-4a23-4fb1-b689-59318014bf57-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg\" (UID: \"a1c30ff2-4a23-4fb1-b689-59318014bf57\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.263956 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg"] Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.355291 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddrbk\" (UniqueName: \"kubernetes.io/projected/a1c30ff2-4a23-4fb1-b689-59318014bf57-kube-api-access-ddrbk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg\" (UID: \"a1c30ff2-4a23-4fb1-b689-59318014bf57\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.355502 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1c30ff2-4a23-4fb1-b689-59318014bf57-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg\" (UID: \"a1c30ff2-4a23-4fb1-b689-59318014bf57\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.355621 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1c30ff2-4a23-4fb1-b689-59318014bf57-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg\" (UID: \"a1c30ff2-4a23-4fb1-b689-59318014bf57\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.361869 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1c30ff2-4a23-4fb1-b689-59318014bf57-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg\" (UID: \"a1c30ff2-4a23-4fb1-b689-59318014bf57\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.362390 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1c30ff2-4a23-4fb1-b689-59318014bf57-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg\" (UID: \"a1c30ff2-4a23-4fb1-b689-59318014bf57\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.386813 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddrbk\" (UniqueName: \"kubernetes.io/projected/a1c30ff2-4a23-4fb1-b689-59318014bf57-kube-api-access-ddrbk\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg\" (UID: \"a1c30ff2-4a23-4fb1-b689-59318014bf57\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" Mar 13 10:38:56 crc kubenswrapper[4632]: I0313 10:38:56.566123 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" Mar 13 10:38:57 crc kubenswrapper[4632]: I0313 10:38:57.130098 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg"] Mar 13 10:38:58 crc kubenswrapper[4632]: I0313 10:38:58.076467 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" event={"ID":"a1c30ff2-4a23-4fb1-b689-59318014bf57","Type":"ContainerStarted","Data":"d7a973a4828687613ca13c1cd91ec5768f0e35db944e95cc7f52edd24c762464"} Mar 13 10:38:58 crc kubenswrapper[4632]: I0313 10:38:58.077122 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" event={"ID":"a1c30ff2-4a23-4fb1-b689-59318014bf57","Type":"ContainerStarted","Data":"f435c3660d1b5d90d5cebceffa800b6b01daace4f9fed586d1ee0eae3bfc0830"} Mar 13 10:38:58 crc kubenswrapper[4632]: I0313 10:38:58.092293 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" podStartSLOduration=1.601623713 podStartE2EDuration="2.092271709s" podCreationTimestamp="2026-03-13 10:38:56 +0000 UTC" firstStartedPulling="2026-03-13 10:38:57.156815408 +0000 UTC m=+2111.179345541" lastFinishedPulling="2026-03-13 10:38:57.647463404 +0000 UTC m=+2111.669993537" observedRunningTime="2026-03-13 10:38:58.086979809 +0000 UTC m=+2112.109509962" watchObservedRunningTime="2026-03-13 10:38:58.092271709 +0000 UTC m=+2112.114801852" Mar 13 10:39:03 crc kubenswrapper[4632]: I0313 10:39:03.096782 4632 generic.go:334] "Generic (PLEG): container finished" podID="a1c30ff2-4a23-4fb1-b689-59318014bf57" containerID="d7a973a4828687613ca13c1cd91ec5768f0e35db944e95cc7f52edd24c762464" exitCode=0 Mar 13 10:39:03 crc kubenswrapper[4632]: I0313 10:39:03.096844 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" event={"ID":"a1c30ff2-4a23-4fb1-b689-59318014bf57","Type":"ContainerDied","Data":"d7a973a4828687613ca13c1cd91ec5768f0e35db944e95cc7f52edd24c762464"} Mar 13 10:39:04 crc kubenswrapper[4632]: I0313 10:39:04.547337 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" Mar 13 10:39:04 crc kubenswrapper[4632]: I0313 10:39:04.634608 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddrbk\" (UniqueName: \"kubernetes.io/projected/a1c30ff2-4a23-4fb1-b689-59318014bf57-kube-api-access-ddrbk\") pod \"a1c30ff2-4a23-4fb1-b689-59318014bf57\" (UID: \"a1c30ff2-4a23-4fb1-b689-59318014bf57\") " Mar 13 10:39:04 crc kubenswrapper[4632]: I0313 10:39:04.635081 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1c30ff2-4a23-4fb1-b689-59318014bf57-inventory\") pod \"a1c30ff2-4a23-4fb1-b689-59318014bf57\" (UID: \"a1c30ff2-4a23-4fb1-b689-59318014bf57\") " Mar 13 10:39:04 crc kubenswrapper[4632]: I0313 10:39:04.635180 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1c30ff2-4a23-4fb1-b689-59318014bf57-ssh-key-openstack-edpm-ipam\") pod \"a1c30ff2-4a23-4fb1-b689-59318014bf57\" (UID: \"a1c30ff2-4a23-4fb1-b689-59318014bf57\") " Mar 13 10:39:04 crc kubenswrapper[4632]: I0313 10:39:04.647741 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1c30ff2-4a23-4fb1-b689-59318014bf57-kube-api-access-ddrbk" (OuterVolumeSpecName: "kube-api-access-ddrbk") pod "a1c30ff2-4a23-4fb1-b689-59318014bf57" (UID: "a1c30ff2-4a23-4fb1-b689-59318014bf57"). InnerVolumeSpecName "kube-api-access-ddrbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:39:04 crc kubenswrapper[4632]: I0313 10:39:04.672612 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1c30ff2-4a23-4fb1-b689-59318014bf57-inventory" (OuterVolumeSpecName: "inventory") pod "a1c30ff2-4a23-4fb1-b689-59318014bf57" (UID: "a1c30ff2-4a23-4fb1-b689-59318014bf57"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:39:04 crc kubenswrapper[4632]: I0313 10:39:04.673115 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1c30ff2-4a23-4fb1-b689-59318014bf57-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "a1c30ff2-4a23-4fb1-b689-59318014bf57" (UID: "a1c30ff2-4a23-4fb1-b689-59318014bf57"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:39:04 crc kubenswrapper[4632]: I0313 10:39:04.737651 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/a1c30ff2-4a23-4fb1-b689-59318014bf57-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:39:04 crc kubenswrapper[4632]: I0313 10:39:04.737707 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/a1c30ff2-4a23-4fb1-b689-59318014bf57-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:39:04 crc kubenswrapper[4632]: I0313 10:39:04.737724 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddrbk\" (UniqueName: \"kubernetes.io/projected/a1c30ff2-4a23-4fb1-b689-59318014bf57-kube-api-access-ddrbk\") on node \"crc\" DevicePath \"\"" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.121627 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" event={"ID":"a1c30ff2-4a23-4fb1-b689-59318014bf57","Type":"ContainerDied","Data":"f435c3660d1b5d90d5cebceffa800b6b01daace4f9fed586d1ee0eae3bfc0830"} Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.122204 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f435c3660d1b5d90d5cebceffa800b6b01daace4f9fed586d1ee0eae3bfc0830" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.121704 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.214618 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg"] Mar 13 10:39:05 crc kubenswrapper[4632]: E0313 10:39:05.215517 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c30ff2-4a23-4fb1-b689-59318014bf57" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.215677 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c30ff2-4a23-4fb1-b689-59318014bf57" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.216055 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c30ff2-4a23-4fb1-b689-59318014bf57" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.216962 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.221891 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.222663 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.223225 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.229599 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.242803 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg"] Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.350180 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qfd27\" (UniqueName: \"kubernetes.io/projected/78de7f45-2a11-4cbe-84bf-46c4307a1459-kube-api-access-qfd27\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-srdvg\" (UID: \"78de7f45-2a11-4cbe-84bf-46c4307a1459\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.350257 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78de7f45-2a11-4cbe-84bf-46c4307a1459-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-srdvg\" (UID: \"78de7f45-2a11-4cbe-84bf-46c4307a1459\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.350309 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78de7f45-2a11-4cbe-84bf-46c4307a1459-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-srdvg\" (UID: \"78de7f45-2a11-4cbe-84bf-46c4307a1459\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.452333 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qfd27\" (UniqueName: \"kubernetes.io/projected/78de7f45-2a11-4cbe-84bf-46c4307a1459-kube-api-access-qfd27\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-srdvg\" (UID: \"78de7f45-2a11-4cbe-84bf-46c4307a1459\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.452778 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78de7f45-2a11-4cbe-84bf-46c4307a1459-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-srdvg\" (UID: \"78de7f45-2a11-4cbe-84bf-46c4307a1459\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.452988 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78de7f45-2a11-4cbe-84bf-46c4307a1459-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-srdvg\" (UID: \"78de7f45-2a11-4cbe-84bf-46c4307a1459\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.457903 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78de7f45-2a11-4cbe-84bf-46c4307a1459-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-srdvg\" (UID: \"78de7f45-2a11-4cbe-84bf-46c4307a1459\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.460414 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78de7f45-2a11-4cbe-84bf-46c4307a1459-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-srdvg\" (UID: \"78de7f45-2a11-4cbe-84bf-46c4307a1459\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.472542 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qfd27\" (UniqueName: \"kubernetes.io/projected/78de7f45-2a11-4cbe-84bf-46c4307a1459-kube-api-access-qfd27\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-srdvg\" (UID: \"78de7f45-2a11-4cbe-84bf-46c4307a1459\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" Mar 13 10:39:05 crc kubenswrapper[4632]: I0313 10:39:05.554721 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" Mar 13 10:39:06 crc kubenswrapper[4632]: I0313 10:39:06.126732 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg"] Mar 13 10:39:07 crc kubenswrapper[4632]: I0313 10:39:07.145031 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" event={"ID":"78de7f45-2a11-4cbe-84bf-46c4307a1459","Type":"ContainerStarted","Data":"273653ee710557decd4c33917e21a560bb09b0670eb74d27ef3d7ffc75810416"} Mar 13 10:39:07 crc kubenswrapper[4632]: I0313 10:39:07.145657 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" event={"ID":"78de7f45-2a11-4cbe-84bf-46c4307a1459","Type":"ContainerStarted","Data":"3b3cfd0adf831ab35b815f5b2519338f27aa52d69c55fcd9c9ad26553b698b72"} Mar 13 10:39:07 crc kubenswrapper[4632]: I0313 10:39:07.178496 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" podStartSLOduration=1.749609712 podStartE2EDuration="2.178473064s" podCreationTimestamp="2026-03-13 10:39:05 +0000 UTC" firstStartedPulling="2026-03-13 10:39:06.14189683 +0000 UTC m=+2120.164426963" lastFinishedPulling="2026-03-13 10:39:06.570760182 +0000 UTC m=+2120.593290315" observedRunningTime="2026-03-13 10:39:07.168235612 +0000 UTC m=+2121.190765745" watchObservedRunningTime="2026-03-13 10:39:07.178473064 +0000 UTC m=+2121.201003197" Mar 13 10:39:08 crc kubenswrapper[4632]: I0313 10:39:08.986290 4632 scope.go:117] "RemoveContainer" containerID="afb05bb00debb2ea4a81d169362ff2bd38d824053184e249dbe02cc1cb10e945" Mar 13 10:39:10 crc kubenswrapper[4632]: I0313 10:39:10.460737 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:39:10 crc kubenswrapper[4632]: I0313 10:39:10.461157 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:39:18 crc kubenswrapper[4632]: I0313 10:39:18.073169 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-gwj5n"] Mar 13 10:39:18 crc kubenswrapper[4632]: I0313 10:39:18.084206 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-gwj5n"] Mar 13 10:39:19 crc kubenswrapper[4632]: I0313 10:39:19.033023 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9n7gj"] Mar 13 10:39:19 crc kubenswrapper[4632]: I0313 10:39:19.041551 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-9n7gj"] Mar 13 10:39:20 crc kubenswrapper[4632]: I0313 10:39:20.058233 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bcce9343-52a3-4e6d-98fd-8e66390020ac" path="/var/lib/kubelet/pods/bcce9343-52a3-4e6d-98fd-8e66390020ac/volumes" Mar 13 10:39:20 crc kubenswrapper[4632]: I0313 10:39:20.060997 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf19672e-3284-49bc-a460-f2e629881d9b" path="/var/lib/kubelet/pods/cf19672e-3284-49bc-a460-f2e629881d9b/volumes" Mar 13 10:39:40 crc kubenswrapper[4632]: I0313 10:39:40.460798 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:39:40 crc kubenswrapper[4632]: I0313 10:39:40.461478 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:39:40 crc kubenswrapper[4632]: I0313 10:39:40.461535 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:39:40 crc kubenswrapper[4632]: I0313 10:39:40.462456 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2bb4e222f4f89a1d4e4bebc809fc60cc762d7ea9b6811f4bcc9cb78c179cd0bd"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:39:40 crc kubenswrapper[4632]: I0313 10:39:40.462514 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://2bb4e222f4f89a1d4e4bebc809fc60cc762d7ea9b6811f4bcc9cb78c179cd0bd" gracePeriod=600 Mar 13 10:39:41 crc kubenswrapper[4632]: I0313 10:39:41.462741 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="2bb4e222f4f89a1d4e4bebc809fc60cc762d7ea9b6811f4bcc9cb78c179cd0bd" exitCode=0 Mar 13 10:39:41 crc kubenswrapper[4632]: I0313 10:39:41.462841 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"2bb4e222f4f89a1d4e4bebc809fc60cc762d7ea9b6811f4bcc9cb78c179cd0bd"} Mar 13 10:39:41 crc kubenswrapper[4632]: I0313 10:39:41.463199 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20"} Mar 13 10:39:41 crc kubenswrapper[4632]: I0313 10:39:41.463232 4632 scope.go:117] "RemoveContainer" containerID="8c0ae371a519eb9db1c6eb843b2bd1981f031101e4086e8bec3fc57f1e905a6f" Mar 13 10:39:48 crc kubenswrapper[4632]: I0313 10:39:48.527801 4632 generic.go:334] "Generic (PLEG): container finished" podID="78de7f45-2a11-4cbe-84bf-46c4307a1459" containerID="273653ee710557decd4c33917e21a560bb09b0670eb74d27ef3d7ffc75810416" exitCode=0 Mar 13 10:39:48 crc kubenswrapper[4632]: I0313 10:39:48.527908 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" event={"ID":"78de7f45-2a11-4cbe-84bf-46c4307a1459","Type":"ContainerDied","Data":"273653ee710557decd4c33917e21a560bb09b0670eb74d27ef3d7ffc75810416"} Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.228161 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.310362 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qfd27\" (UniqueName: \"kubernetes.io/projected/78de7f45-2a11-4cbe-84bf-46c4307a1459-kube-api-access-qfd27\") pod \"78de7f45-2a11-4cbe-84bf-46c4307a1459\" (UID: \"78de7f45-2a11-4cbe-84bf-46c4307a1459\") " Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.310432 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78de7f45-2a11-4cbe-84bf-46c4307a1459-ssh-key-openstack-edpm-ipam\") pod \"78de7f45-2a11-4cbe-84bf-46c4307a1459\" (UID: \"78de7f45-2a11-4cbe-84bf-46c4307a1459\") " Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.310672 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78de7f45-2a11-4cbe-84bf-46c4307a1459-inventory\") pod \"78de7f45-2a11-4cbe-84bf-46c4307a1459\" (UID: \"78de7f45-2a11-4cbe-84bf-46c4307a1459\") " Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.317163 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78de7f45-2a11-4cbe-84bf-46c4307a1459-kube-api-access-qfd27" (OuterVolumeSpecName: "kube-api-access-qfd27") pod "78de7f45-2a11-4cbe-84bf-46c4307a1459" (UID: "78de7f45-2a11-4cbe-84bf-46c4307a1459"). InnerVolumeSpecName "kube-api-access-qfd27". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.342827 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78de7f45-2a11-4cbe-84bf-46c4307a1459-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "78de7f45-2a11-4cbe-84bf-46c4307a1459" (UID: "78de7f45-2a11-4cbe-84bf-46c4307a1459"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.343068 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/78de7f45-2a11-4cbe-84bf-46c4307a1459-inventory" (OuterVolumeSpecName: "inventory") pod "78de7f45-2a11-4cbe-84bf-46c4307a1459" (UID: "78de7f45-2a11-4cbe-84bf-46c4307a1459"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.414650 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/78de7f45-2a11-4cbe-84bf-46c4307a1459-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.414680 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qfd27\" (UniqueName: \"kubernetes.io/projected/78de7f45-2a11-4cbe-84bf-46c4307a1459-kube-api-access-qfd27\") on node \"crc\" DevicePath \"\"" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.414693 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/78de7f45-2a11-4cbe-84bf-46c4307a1459-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.550401 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" event={"ID":"78de7f45-2a11-4cbe-84bf-46c4307a1459","Type":"ContainerDied","Data":"3b3cfd0adf831ab35b815f5b2519338f27aa52d69c55fcd9c9ad26553b698b72"} Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.550448 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-srdvg" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.550461 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b3cfd0adf831ab35b815f5b2519338f27aa52d69c55fcd9c9ad26553b698b72" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.664085 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw"] Mar 13 10:39:50 crc kubenswrapper[4632]: E0313 10:39:50.664595 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78de7f45-2a11-4cbe-84bf-46c4307a1459" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.664624 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="78de7f45-2a11-4cbe-84bf-46c4307a1459" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.664958 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="78de7f45-2a11-4cbe-84bf-46c4307a1459" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.665970 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.669648 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.670087 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.670292 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.670308 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.681255 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw"] Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.822363 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4931647b-bba4-489f-b5c1-cbe714834388-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw\" (UID: \"4931647b-bba4-489f-b5c1-cbe714834388\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.822882 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4931647b-bba4-489f-b5c1-cbe714834388-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw\" (UID: \"4931647b-bba4-489f-b5c1-cbe714834388\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.822985 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr759\" (UniqueName: \"kubernetes.io/projected/4931647b-bba4-489f-b5c1-cbe714834388-kube-api-access-hr759\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw\" (UID: \"4931647b-bba4-489f-b5c1-cbe714834388\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.925781 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4931647b-bba4-489f-b5c1-cbe714834388-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw\" (UID: \"4931647b-bba4-489f-b5c1-cbe714834388\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.925876 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4931647b-bba4-489f-b5c1-cbe714834388-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw\" (UID: \"4931647b-bba4-489f-b5c1-cbe714834388\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.925956 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr759\" (UniqueName: \"kubernetes.io/projected/4931647b-bba4-489f-b5c1-cbe714834388-kube-api-access-hr759\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw\" (UID: \"4931647b-bba4-489f-b5c1-cbe714834388\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.931225 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4931647b-bba4-489f-b5c1-cbe714834388-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw\" (UID: \"4931647b-bba4-489f-b5c1-cbe714834388\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.931592 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4931647b-bba4-489f-b5c1-cbe714834388-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw\" (UID: \"4931647b-bba4-489f-b5c1-cbe714834388\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.944879 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr759\" (UniqueName: \"kubernetes.io/projected/4931647b-bba4-489f-b5c1-cbe714834388-kube-api-access-hr759\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw\" (UID: \"4931647b-bba4-489f-b5c1-cbe714834388\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" Mar 13 10:39:50 crc kubenswrapper[4632]: I0313 10:39:50.990881 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" Mar 13 10:39:51 crc kubenswrapper[4632]: I0313 10:39:51.582520 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw"] Mar 13 10:39:52 crc kubenswrapper[4632]: I0313 10:39:52.573173 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" event={"ID":"4931647b-bba4-489f-b5c1-cbe714834388","Type":"ContainerStarted","Data":"66ae927115cf223ae80a7f1174de3d22e1d8235d0f7cb5cfd88ba65e0b69c69d"} Mar 13 10:39:52 crc kubenswrapper[4632]: I0313 10:39:52.573510 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" event={"ID":"4931647b-bba4-489f-b5c1-cbe714834388","Type":"ContainerStarted","Data":"23b2acab45805714175c6485dd856ee0e45e8f30118161e4579b9b2fc662cd4d"} Mar 13 10:39:52 crc kubenswrapper[4632]: I0313 10:39:52.595583 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" podStartSLOduration=2.197014011 podStartE2EDuration="2.595567647s" podCreationTimestamp="2026-03-13 10:39:50 +0000 UTC" firstStartedPulling="2026-03-13 10:39:51.602812703 +0000 UTC m=+2165.625342836" lastFinishedPulling="2026-03-13 10:39:52.001366339 +0000 UTC m=+2166.023896472" observedRunningTime="2026-03-13 10:39:52.594582353 +0000 UTC m=+2166.617112496" watchObservedRunningTime="2026-03-13 10:39:52.595567647 +0000 UTC m=+2166.618097780" Mar 13 10:40:00 crc kubenswrapper[4632]: I0313 10:40:00.056134 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-ngzsx"] Mar 13 10:40:00 crc kubenswrapper[4632]: I0313 10:40:00.064680 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-ngzsx"] Mar 13 10:40:00 crc kubenswrapper[4632]: I0313 10:40:00.149698 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556640-zlhsq"] Mar 13 10:40:00 crc kubenswrapper[4632]: I0313 10:40:00.151260 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556640-zlhsq" Mar 13 10:40:00 crc kubenswrapper[4632]: I0313 10:40:00.153625 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:40:00 crc kubenswrapper[4632]: I0313 10:40:00.154198 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:40:00 crc kubenswrapper[4632]: I0313 10:40:00.154262 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:40:00 crc kubenswrapper[4632]: I0313 10:40:00.169596 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556640-zlhsq"] Mar 13 10:40:00 crc kubenswrapper[4632]: I0313 10:40:00.213985 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj8j9\" (UniqueName: \"kubernetes.io/projected/4bcf0de2-27ca-4278-80a3-080ce237e6df-kube-api-access-rj8j9\") pod \"auto-csr-approver-29556640-zlhsq\" (UID: \"4bcf0de2-27ca-4278-80a3-080ce237e6df\") " pod="openshift-infra/auto-csr-approver-29556640-zlhsq" Mar 13 10:40:00 crc kubenswrapper[4632]: I0313 10:40:00.316071 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj8j9\" (UniqueName: \"kubernetes.io/projected/4bcf0de2-27ca-4278-80a3-080ce237e6df-kube-api-access-rj8j9\") pod \"auto-csr-approver-29556640-zlhsq\" (UID: \"4bcf0de2-27ca-4278-80a3-080ce237e6df\") " pod="openshift-infra/auto-csr-approver-29556640-zlhsq" Mar 13 10:40:00 crc kubenswrapper[4632]: I0313 10:40:00.335609 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj8j9\" (UniqueName: \"kubernetes.io/projected/4bcf0de2-27ca-4278-80a3-080ce237e6df-kube-api-access-rj8j9\") pod \"auto-csr-approver-29556640-zlhsq\" (UID: \"4bcf0de2-27ca-4278-80a3-080ce237e6df\") " pod="openshift-infra/auto-csr-approver-29556640-zlhsq" Mar 13 10:40:00 crc kubenswrapper[4632]: I0313 10:40:00.480617 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556640-zlhsq" Mar 13 10:40:01 crc kubenswrapper[4632]: I0313 10:40:01.018730 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556640-zlhsq"] Mar 13 10:40:01 crc kubenswrapper[4632]: I0313 10:40:01.668920 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556640-zlhsq" event={"ID":"4bcf0de2-27ca-4278-80a3-080ce237e6df","Type":"ContainerStarted","Data":"dc6b97651a5d32b9be163f9c3152748b4ed8b19f6748652486195b679826838d"} Mar 13 10:40:02 crc kubenswrapper[4632]: I0313 10:40:02.057739 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="601f3615-5015-486a-bbb5-04c683da6990" path="/var/lib/kubelet/pods/601f3615-5015-486a-bbb5-04c683da6990/volumes" Mar 13 10:40:02 crc kubenswrapper[4632]: I0313 10:40:02.678524 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556640-zlhsq" event={"ID":"4bcf0de2-27ca-4278-80a3-080ce237e6df","Type":"ContainerStarted","Data":"84030d1b6c9dd12b070ed748955e52fe36ed2cac9f9bdddb744ca14dc6fbfa0a"} Mar 13 10:40:02 crc kubenswrapper[4632]: I0313 10:40:02.696642 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556640-zlhsq" podStartSLOduration=1.48696992 podStartE2EDuration="2.696617s" podCreationTimestamp="2026-03-13 10:40:00 +0000 UTC" firstStartedPulling="2026-03-13 10:40:01.032449784 +0000 UTC m=+2175.054979917" lastFinishedPulling="2026-03-13 10:40:02.242096864 +0000 UTC m=+2176.264626997" observedRunningTime="2026-03-13 10:40:02.692860477 +0000 UTC m=+2176.715390610" watchObservedRunningTime="2026-03-13 10:40:02.696617 +0000 UTC m=+2176.719147133" Mar 13 10:40:03 crc kubenswrapper[4632]: I0313 10:40:03.690022 4632 generic.go:334] "Generic (PLEG): container finished" podID="4bcf0de2-27ca-4278-80a3-080ce237e6df" containerID="84030d1b6c9dd12b070ed748955e52fe36ed2cac9f9bdddb744ca14dc6fbfa0a" exitCode=0 Mar 13 10:40:03 crc kubenswrapper[4632]: I0313 10:40:03.690334 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556640-zlhsq" event={"ID":"4bcf0de2-27ca-4278-80a3-080ce237e6df","Type":"ContainerDied","Data":"84030d1b6c9dd12b070ed748955e52fe36ed2cac9f9bdddb744ca14dc6fbfa0a"} Mar 13 10:40:05 crc kubenswrapper[4632]: I0313 10:40:05.053755 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556640-zlhsq" Mar 13 10:40:05 crc kubenswrapper[4632]: I0313 10:40:05.106837 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rj8j9\" (UniqueName: \"kubernetes.io/projected/4bcf0de2-27ca-4278-80a3-080ce237e6df-kube-api-access-rj8j9\") pod \"4bcf0de2-27ca-4278-80a3-080ce237e6df\" (UID: \"4bcf0de2-27ca-4278-80a3-080ce237e6df\") " Mar 13 10:40:05 crc kubenswrapper[4632]: I0313 10:40:05.114308 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bcf0de2-27ca-4278-80a3-080ce237e6df-kube-api-access-rj8j9" (OuterVolumeSpecName: "kube-api-access-rj8j9") pod "4bcf0de2-27ca-4278-80a3-080ce237e6df" (UID: "4bcf0de2-27ca-4278-80a3-080ce237e6df"). InnerVolumeSpecName "kube-api-access-rj8j9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:40:05 crc kubenswrapper[4632]: I0313 10:40:05.211522 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rj8j9\" (UniqueName: \"kubernetes.io/projected/4bcf0de2-27ca-4278-80a3-080ce237e6df-kube-api-access-rj8j9\") on node \"crc\" DevicePath \"\"" Mar 13 10:40:05 crc kubenswrapper[4632]: I0313 10:40:05.710209 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556640-zlhsq" event={"ID":"4bcf0de2-27ca-4278-80a3-080ce237e6df","Type":"ContainerDied","Data":"dc6b97651a5d32b9be163f9c3152748b4ed8b19f6748652486195b679826838d"} Mar 13 10:40:05 crc kubenswrapper[4632]: I0313 10:40:05.710248 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dc6b97651a5d32b9be163f9c3152748b4ed8b19f6748652486195b679826838d" Mar 13 10:40:05 crc kubenswrapper[4632]: I0313 10:40:05.710489 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556640-zlhsq" Mar 13 10:40:05 crc kubenswrapper[4632]: I0313 10:40:05.762839 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556634-6n98g"] Mar 13 10:40:05 crc kubenswrapper[4632]: I0313 10:40:05.769870 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556634-6n98g"] Mar 13 10:40:06 crc kubenswrapper[4632]: I0313 10:40:06.058504 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="155ba738-4ba0-424a-a1d7-067786728969" path="/var/lib/kubelet/pods/155ba738-4ba0-424a-a1d7-067786728969/volumes" Mar 13 10:40:09 crc kubenswrapper[4632]: I0313 10:40:09.042268 4632 scope.go:117] "RemoveContainer" containerID="a8f98d9cfd7da7677c0fe463edd081d6aa2858ecb1027917673862b2700f1545" Mar 13 10:40:09 crc kubenswrapper[4632]: I0313 10:40:09.084227 4632 scope.go:117] "RemoveContainer" containerID="f95b291d052d44a477db7fca5558efb7e90f20270d66ae208043b37111d582be" Mar 13 10:40:09 crc kubenswrapper[4632]: I0313 10:40:09.160262 4632 scope.go:117] "RemoveContainer" containerID="379356ecac878a5f4776d015be267e8c7eec62c977ce924abd53ff44455ce8e4" Mar 13 10:40:09 crc kubenswrapper[4632]: I0313 10:40:09.205542 4632 scope.go:117] "RemoveContainer" containerID="e181311595cfc3a50154df8d12fbc0793d907a3185d962d8a64fc357e0b6ee4f" Mar 13 10:40:45 crc kubenswrapper[4632]: I0313 10:40:45.065022 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" event={"ID":"4931647b-bba4-489f-b5c1-cbe714834388","Type":"ContainerDied","Data":"66ae927115cf223ae80a7f1174de3d22e1d8235d0f7cb5cfd88ba65e0b69c69d"} Mar 13 10:40:45 crc kubenswrapper[4632]: I0313 10:40:45.065033 4632 generic.go:334] "Generic (PLEG): container finished" podID="4931647b-bba4-489f-b5c1-cbe714834388" containerID="66ae927115cf223ae80a7f1174de3d22e1d8235d0f7cb5cfd88ba65e0b69c69d" exitCode=0 Mar 13 10:40:46 crc kubenswrapper[4632]: I0313 10:40:46.631240 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" Mar 13 10:40:46 crc kubenswrapper[4632]: I0313 10:40:46.730680 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr759\" (UniqueName: \"kubernetes.io/projected/4931647b-bba4-489f-b5c1-cbe714834388-kube-api-access-hr759\") pod \"4931647b-bba4-489f-b5c1-cbe714834388\" (UID: \"4931647b-bba4-489f-b5c1-cbe714834388\") " Mar 13 10:40:46 crc kubenswrapper[4632]: I0313 10:40:46.731310 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4931647b-bba4-489f-b5c1-cbe714834388-ssh-key-openstack-edpm-ipam\") pod \"4931647b-bba4-489f-b5c1-cbe714834388\" (UID: \"4931647b-bba4-489f-b5c1-cbe714834388\") " Mar 13 10:40:46 crc kubenswrapper[4632]: I0313 10:40:46.731461 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4931647b-bba4-489f-b5c1-cbe714834388-inventory\") pod \"4931647b-bba4-489f-b5c1-cbe714834388\" (UID: \"4931647b-bba4-489f-b5c1-cbe714834388\") " Mar 13 10:40:46 crc kubenswrapper[4632]: I0313 10:40:46.753696 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4931647b-bba4-489f-b5c1-cbe714834388-kube-api-access-hr759" (OuterVolumeSpecName: "kube-api-access-hr759") pod "4931647b-bba4-489f-b5c1-cbe714834388" (UID: "4931647b-bba4-489f-b5c1-cbe714834388"). InnerVolumeSpecName "kube-api-access-hr759". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:40:46 crc kubenswrapper[4632]: I0313 10:40:46.770485 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4931647b-bba4-489f-b5c1-cbe714834388-inventory" (OuterVolumeSpecName: "inventory") pod "4931647b-bba4-489f-b5c1-cbe714834388" (UID: "4931647b-bba4-489f-b5c1-cbe714834388"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:40:46 crc kubenswrapper[4632]: I0313 10:40:46.778521 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4931647b-bba4-489f-b5c1-cbe714834388-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4931647b-bba4-489f-b5c1-cbe714834388" (UID: "4931647b-bba4-489f-b5c1-cbe714834388"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:40:46 crc kubenswrapper[4632]: I0313 10:40:46.836011 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr759\" (UniqueName: \"kubernetes.io/projected/4931647b-bba4-489f-b5c1-cbe714834388-kube-api-access-hr759\") on node \"crc\" DevicePath \"\"" Mar 13 10:40:46 crc kubenswrapper[4632]: I0313 10:40:46.836061 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4931647b-bba4-489f-b5c1-cbe714834388-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:40:46 crc kubenswrapper[4632]: I0313 10:40:46.836078 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4931647b-bba4-489f-b5c1-cbe714834388-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.090230 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" event={"ID":"4931647b-bba4-489f-b5c1-cbe714834388","Type":"ContainerDied","Data":"23b2acab45805714175c6485dd856ee0e45e8f30118161e4579b9b2fc662cd4d"} Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.090283 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23b2acab45805714175c6485dd856ee0e45e8f30118161e4579b9b2fc662cd4d" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.090370 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.166666 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-55n7g"] Mar 13 10:40:47 crc kubenswrapper[4632]: E0313 10:40:47.167068 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4931647b-bba4-489f-b5c1-cbe714834388" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.167088 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="4931647b-bba4-489f-b5c1-cbe714834388" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 13 10:40:47 crc kubenswrapper[4632]: E0313 10:40:47.167110 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bcf0de2-27ca-4278-80a3-080ce237e6df" containerName="oc" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.167116 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bcf0de2-27ca-4278-80a3-080ce237e6df" containerName="oc" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.167319 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="4931647b-bba4-489f-b5c1-cbe714834388" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.167353 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bcf0de2-27ca-4278-80a3-080ce237e6df" containerName="oc" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.167973 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.170090 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.170561 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.170802 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.171058 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.189925 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-55n7g"] Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.345134 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-55n7g\" (UID: \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\") " pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.345291 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nzt5\" (UniqueName: \"kubernetes.io/projected/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-kube-api-access-7nzt5\") pod \"ssh-known-hosts-edpm-deployment-55n7g\" (UID: \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\") " pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.345385 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-55n7g\" (UID: \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\") " pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.447738 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nzt5\" (UniqueName: \"kubernetes.io/projected/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-kube-api-access-7nzt5\") pod \"ssh-known-hosts-edpm-deployment-55n7g\" (UID: \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\") " pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.447878 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-55n7g\" (UID: \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\") " pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.447976 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-55n7g\" (UID: \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\") " pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.452617 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-55n7g\" (UID: \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\") " pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.471645 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-55n7g\" (UID: \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\") " pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.482140 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nzt5\" (UniqueName: \"kubernetes.io/projected/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-kube-api-access-7nzt5\") pod \"ssh-known-hosts-edpm-deployment-55n7g\" (UID: \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\") " pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" Mar 13 10:40:47 crc kubenswrapper[4632]: I0313 10:40:47.485028 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" Mar 13 10:40:48 crc kubenswrapper[4632]: I0313 10:40:48.357562 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-55n7g"] Mar 13 10:40:48 crc kubenswrapper[4632]: I0313 10:40:48.376148 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:40:48 crc kubenswrapper[4632]: I0313 10:40:48.833195 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:40:49 crc kubenswrapper[4632]: I0313 10:40:49.112085 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" event={"ID":"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1","Type":"ContainerStarted","Data":"67d1c660a605cb7f487da3c88f2f7b771a995c235aa1d21c9c9772f3c660b828"} Mar 13 10:40:49 crc kubenswrapper[4632]: I0313 10:40:49.113077 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" event={"ID":"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1","Type":"ContainerStarted","Data":"8cb8154bbb80936b80e55c8289b4e5686e6030c27b5b59809debc612b023074a"} Mar 13 10:40:49 crc kubenswrapper[4632]: I0313 10:40:49.141102 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" podStartSLOduration=1.685400061 podStartE2EDuration="2.141072403s" podCreationTimestamp="2026-03-13 10:40:47 +0000 UTC" firstStartedPulling="2026-03-13 10:40:48.375116381 +0000 UTC m=+2222.397646514" lastFinishedPulling="2026-03-13 10:40:48.830788723 +0000 UTC m=+2222.853318856" observedRunningTime="2026-03-13 10:40:49.131739482 +0000 UTC m=+2223.154269625" watchObservedRunningTime="2026-03-13 10:40:49.141072403 +0000 UTC m=+2223.163602536" Mar 13 10:40:53 crc kubenswrapper[4632]: I0313 10:40:53.862993 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-c6vkp"] Mar 13 10:40:53 crc kubenswrapper[4632]: I0313 10:40:53.865853 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:40:53 crc kubenswrapper[4632]: I0313 10:40:53.894311 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c6vkp"] Mar 13 10:40:54 crc kubenswrapper[4632]: I0313 10:40:54.031862 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94fd5d57-5fb3-4b34-a545-6bdc3f219354-catalog-content\") pod \"redhat-operators-c6vkp\" (UID: \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\") " pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:40:54 crc kubenswrapper[4632]: I0313 10:40:54.032079 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94fd5d57-5fb3-4b34-a545-6bdc3f219354-utilities\") pod \"redhat-operators-c6vkp\" (UID: \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\") " pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:40:54 crc kubenswrapper[4632]: I0313 10:40:54.032272 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-srngn\" (UniqueName: \"kubernetes.io/projected/94fd5d57-5fb3-4b34-a545-6bdc3f219354-kube-api-access-srngn\") pod \"redhat-operators-c6vkp\" (UID: \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\") " pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:40:54 crc kubenswrapper[4632]: I0313 10:40:54.134396 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94fd5d57-5fb3-4b34-a545-6bdc3f219354-catalog-content\") pod \"redhat-operators-c6vkp\" (UID: \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\") " pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:40:54 crc kubenswrapper[4632]: I0313 10:40:54.134504 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94fd5d57-5fb3-4b34-a545-6bdc3f219354-utilities\") pod \"redhat-operators-c6vkp\" (UID: \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\") " pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:40:54 crc kubenswrapper[4632]: I0313 10:40:54.134581 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-srngn\" (UniqueName: \"kubernetes.io/projected/94fd5d57-5fb3-4b34-a545-6bdc3f219354-kube-api-access-srngn\") pod \"redhat-operators-c6vkp\" (UID: \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\") " pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:40:54 crc kubenswrapper[4632]: I0313 10:40:54.135276 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94fd5d57-5fb3-4b34-a545-6bdc3f219354-utilities\") pod \"redhat-operators-c6vkp\" (UID: \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\") " pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:40:54 crc kubenswrapper[4632]: I0313 10:40:54.135360 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94fd5d57-5fb3-4b34-a545-6bdc3f219354-catalog-content\") pod \"redhat-operators-c6vkp\" (UID: \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\") " pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:40:54 crc kubenswrapper[4632]: I0313 10:40:54.157840 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-srngn\" (UniqueName: \"kubernetes.io/projected/94fd5d57-5fb3-4b34-a545-6bdc3f219354-kube-api-access-srngn\") pod \"redhat-operators-c6vkp\" (UID: \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\") " pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:40:54 crc kubenswrapper[4632]: I0313 10:40:54.197777 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:40:54 crc kubenswrapper[4632]: I0313 10:40:54.732670 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-c6vkp"] Mar 13 10:40:55 crc kubenswrapper[4632]: I0313 10:40:55.179825 4632 generic.go:334] "Generic (PLEG): container finished" podID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerID="d6c5924afdfbc6243bd936794f31e2092a18f09ea985a6206e818699a4eab9bc" exitCode=0 Mar 13 10:40:55 crc kubenswrapper[4632]: I0313 10:40:55.180233 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6vkp" event={"ID":"94fd5d57-5fb3-4b34-a545-6bdc3f219354","Type":"ContainerDied","Data":"d6c5924afdfbc6243bd936794f31e2092a18f09ea985a6206e818699a4eab9bc"} Mar 13 10:40:55 crc kubenswrapper[4632]: I0313 10:40:55.180262 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6vkp" event={"ID":"94fd5d57-5fb3-4b34-a545-6bdc3f219354","Type":"ContainerStarted","Data":"cfe1bf74852782b72ab02590cc17b2a400ee246e92a62ec6365c71796dd98bad"} Mar 13 10:40:57 crc kubenswrapper[4632]: I0313 10:40:57.219356 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6vkp" event={"ID":"94fd5d57-5fb3-4b34-a545-6bdc3f219354","Type":"ContainerStarted","Data":"e5b275bdebfc8c0e84bc1511d86355567607af8b44b7d9fd25e952a957898f8e"} Mar 13 10:40:58 crc kubenswrapper[4632]: I0313 10:40:58.229154 4632 generic.go:334] "Generic (PLEG): container finished" podID="9ff4122d-b9f1-4dd0-80dc-deb9d84760e1" containerID="67d1c660a605cb7f487da3c88f2f7b771a995c235aa1d21c9c9772f3c660b828" exitCode=0 Mar 13 10:40:58 crc kubenswrapper[4632]: I0313 10:40:58.229237 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" event={"ID":"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1","Type":"ContainerDied","Data":"67d1c660a605cb7f487da3c88f2f7b771a995c235aa1d21c9c9772f3c660b828"} Mar 13 10:40:59 crc kubenswrapper[4632]: I0313 10:40:59.834548 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" Mar 13 10:40:59 crc kubenswrapper[4632]: I0313 10:40:59.975466 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nzt5\" (UniqueName: \"kubernetes.io/projected/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-kube-api-access-7nzt5\") pod \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\" (UID: \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\") " Mar 13 10:40:59 crc kubenswrapper[4632]: I0313 10:40:59.975851 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-ssh-key-openstack-edpm-ipam\") pod \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\" (UID: \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\") " Mar 13 10:40:59 crc kubenswrapper[4632]: I0313 10:40:59.975985 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-inventory-0\") pod \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\" (UID: \"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1\") " Mar 13 10:40:59 crc kubenswrapper[4632]: I0313 10:40:59.994660 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-kube-api-access-7nzt5" (OuterVolumeSpecName: "kube-api-access-7nzt5") pod "9ff4122d-b9f1-4dd0-80dc-deb9d84760e1" (UID: "9ff4122d-b9f1-4dd0-80dc-deb9d84760e1"). InnerVolumeSpecName "kube-api-access-7nzt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.015107 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9ff4122d-b9f1-4dd0-80dc-deb9d84760e1" (UID: "9ff4122d-b9f1-4dd0-80dc-deb9d84760e1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.020278 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "9ff4122d-b9f1-4dd0-80dc-deb9d84760e1" (UID: "9ff4122d-b9f1-4dd0-80dc-deb9d84760e1"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.078591 4632 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-inventory-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.078629 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nzt5\" (UniqueName: \"kubernetes.io/projected/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-kube-api-access-7nzt5\") on node \"crc\" DevicePath \"\"" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.078661 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9ff4122d-b9f1-4dd0-80dc-deb9d84760e1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.248826 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" event={"ID":"9ff4122d-b9f1-4dd0-80dc-deb9d84760e1","Type":"ContainerDied","Data":"8cb8154bbb80936b80e55c8289b4e5686e6030c27b5b59809debc612b023074a"} Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.248868 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-55n7g" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.248869 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cb8154bbb80936b80e55c8289b4e5686e6030c27b5b59809debc612b023074a" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.336719 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk"] Mar 13 10:41:00 crc kubenswrapper[4632]: E0313 10:41:00.337751 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ff4122d-b9f1-4dd0-80dc-deb9d84760e1" containerName="ssh-known-hosts-edpm-deployment" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.337781 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ff4122d-b9f1-4dd0-80dc-deb9d84760e1" containerName="ssh-known-hosts-edpm-deployment" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.338343 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ff4122d-b9f1-4dd0-80dc-deb9d84760e1" containerName="ssh-known-hosts-edpm-deployment" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.343626 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.348451 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.360660 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.360844 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.361184 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.407929 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk"] Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.488316 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-f9rbk\" (UID: \"f69a3b21-eb1c-4300-91dc-55766900da95\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.488411 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-f9rbk\" (UID: \"f69a3b21-eb1c-4300-91dc-55766900da95\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.488483 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6d8hr\" (UniqueName: \"kubernetes.io/projected/f69a3b21-eb1c-4300-91dc-55766900da95-kube-api-access-6d8hr\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-f9rbk\" (UID: \"f69a3b21-eb1c-4300-91dc-55766900da95\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.590328 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-f9rbk\" (UID: \"f69a3b21-eb1c-4300-91dc-55766900da95\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.590421 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-f9rbk\" (UID: \"f69a3b21-eb1c-4300-91dc-55766900da95\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.590491 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6d8hr\" (UniqueName: \"kubernetes.io/projected/f69a3b21-eb1c-4300-91dc-55766900da95-kube-api-access-6d8hr\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-f9rbk\" (UID: \"f69a3b21-eb1c-4300-91dc-55766900da95\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.596219 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-f9rbk\" (UID: \"f69a3b21-eb1c-4300-91dc-55766900da95\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.597204 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-f9rbk\" (UID: \"f69a3b21-eb1c-4300-91dc-55766900da95\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.617258 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6d8hr\" (UniqueName: \"kubernetes.io/projected/f69a3b21-eb1c-4300-91dc-55766900da95-kube-api-access-6d8hr\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-f9rbk\" (UID: \"f69a3b21-eb1c-4300-91dc-55766900da95\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" Mar 13 10:41:00 crc kubenswrapper[4632]: I0313 10:41:00.688915 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" Mar 13 10:41:01 crc kubenswrapper[4632]: I0313 10:41:01.316696 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk"] Mar 13 10:41:01 crc kubenswrapper[4632]: W0313 10:41:01.369160 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf69a3b21_eb1c_4300_91dc_55766900da95.slice/crio-c652193575851bc88eb9652edd9cc1fa8aa0f642f4ed0ff9f52d97763733bab8 WatchSource:0}: Error finding container c652193575851bc88eb9652edd9cc1fa8aa0f642f4ed0ff9f52d97763733bab8: Status 404 returned error can't find the container with id c652193575851bc88eb9652edd9cc1fa8aa0f642f4ed0ff9f52d97763733bab8 Mar 13 10:41:02 crc kubenswrapper[4632]: I0313 10:41:02.270722 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" event={"ID":"f69a3b21-eb1c-4300-91dc-55766900da95","Type":"ContainerStarted","Data":"d84c74e44c6bd120cba0edb788c43ce395addd117643e58c370d3d9a16355c26"} Mar 13 10:41:02 crc kubenswrapper[4632]: I0313 10:41:02.271055 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" event={"ID":"f69a3b21-eb1c-4300-91dc-55766900da95","Type":"ContainerStarted","Data":"c652193575851bc88eb9652edd9cc1fa8aa0f642f4ed0ff9f52d97763733bab8"} Mar 13 10:41:02 crc kubenswrapper[4632]: I0313 10:41:02.273205 4632 generic.go:334] "Generic (PLEG): container finished" podID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerID="e5b275bdebfc8c0e84bc1511d86355567607af8b44b7d9fd25e952a957898f8e" exitCode=0 Mar 13 10:41:02 crc kubenswrapper[4632]: I0313 10:41:02.273236 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6vkp" event={"ID":"94fd5d57-5fb3-4b34-a545-6bdc3f219354","Type":"ContainerDied","Data":"e5b275bdebfc8c0e84bc1511d86355567607af8b44b7d9fd25e952a957898f8e"} Mar 13 10:41:02 crc kubenswrapper[4632]: I0313 10:41:02.323957 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" podStartSLOduration=1.84543192 podStartE2EDuration="2.323925517s" podCreationTimestamp="2026-03-13 10:41:00 +0000 UTC" firstStartedPulling="2026-03-13 10:41:01.372558564 +0000 UTC m=+2235.395088697" lastFinishedPulling="2026-03-13 10:41:01.851052161 +0000 UTC m=+2235.873582294" observedRunningTime="2026-03-13 10:41:02.296086602 +0000 UTC m=+2236.318616735" watchObservedRunningTime="2026-03-13 10:41:02.323925517 +0000 UTC m=+2236.346455650" Mar 13 10:41:03 crc kubenswrapper[4632]: I0313 10:41:03.283645 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6vkp" event={"ID":"94fd5d57-5fb3-4b34-a545-6bdc3f219354","Type":"ContainerStarted","Data":"c7a9c2f67c2bc9ec6015b2a50c3cfdd53b590fc7fda70e8b949fe66cb4e0aff2"} Mar 13 10:41:03 crc kubenswrapper[4632]: I0313 10:41:03.306726 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-c6vkp" podStartSLOduration=2.655884976 podStartE2EDuration="10.306708495s" podCreationTimestamp="2026-03-13 10:40:53 +0000 UTC" firstStartedPulling="2026-03-13 10:40:55.182312252 +0000 UTC m=+2229.204842385" lastFinishedPulling="2026-03-13 10:41:02.833135771 +0000 UTC m=+2236.855665904" observedRunningTime="2026-03-13 10:41:03.302905842 +0000 UTC m=+2237.325435985" watchObservedRunningTime="2026-03-13 10:41:03.306708495 +0000 UTC m=+2237.329238628" Mar 13 10:41:04 crc kubenswrapper[4632]: I0313 10:41:04.198453 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:41:04 crc kubenswrapper[4632]: I0313 10:41:04.199960 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:41:05 crc kubenswrapper[4632]: I0313 10:41:05.252110 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c6vkp" podUID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerName="registry-server" probeResult="failure" output=< Mar 13 10:41:05 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:41:05 crc kubenswrapper[4632]: > Mar 13 10:41:12 crc kubenswrapper[4632]: I0313 10:41:12.374133 4632 generic.go:334] "Generic (PLEG): container finished" podID="f69a3b21-eb1c-4300-91dc-55766900da95" containerID="d84c74e44c6bd120cba0edb788c43ce395addd117643e58c370d3d9a16355c26" exitCode=0 Mar 13 10:41:12 crc kubenswrapper[4632]: I0313 10:41:12.374221 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" event={"ID":"f69a3b21-eb1c-4300-91dc-55766900da95","Type":"ContainerDied","Data":"d84c74e44c6bd120cba0edb788c43ce395addd117643e58c370d3d9a16355c26"} Mar 13 10:41:13 crc kubenswrapper[4632]: I0313 10:41:13.861363 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" Mar 13 10:41:13 crc kubenswrapper[4632]: I0313 10:41:13.960970 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-inventory\") pod \"f69a3b21-eb1c-4300-91dc-55766900da95\" (UID: \"f69a3b21-eb1c-4300-91dc-55766900da95\") " Mar 13 10:41:13 crc kubenswrapper[4632]: I0313 10:41:13.961123 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-ssh-key-openstack-edpm-ipam\") pod \"f69a3b21-eb1c-4300-91dc-55766900da95\" (UID: \"f69a3b21-eb1c-4300-91dc-55766900da95\") " Mar 13 10:41:13 crc kubenswrapper[4632]: I0313 10:41:13.961993 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6d8hr\" (UniqueName: \"kubernetes.io/projected/f69a3b21-eb1c-4300-91dc-55766900da95-kube-api-access-6d8hr\") pod \"f69a3b21-eb1c-4300-91dc-55766900da95\" (UID: \"f69a3b21-eb1c-4300-91dc-55766900da95\") " Mar 13 10:41:13 crc kubenswrapper[4632]: I0313 10:41:13.967256 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f69a3b21-eb1c-4300-91dc-55766900da95-kube-api-access-6d8hr" (OuterVolumeSpecName: "kube-api-access-6d8hr") pod "f69a3b21-eb1c-4300-91dc-55766900da95" (UID: "f69a3b21-eb1c-4300-91dc-55766900da95"). InnerVolumeSpecName "kube-api-access-6d8hr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:41:13 crc kubenswrapper[4632]: E0313 10:41:13.990015 4632 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-inventory podName:f69a3b21-eb1c-4300-91dc-55766900da95 nodeName:}" failed. No retries permitted until 2026-03-13 10:41:14.489435037 +0000 UTC m=+2248.511965170 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-inventory") pod "f69a3b21-eb1c-4300-91dc-55766900da95" (UID: "f69a3b21-eb1c-4300-91dc-55766900da95") : error deleting /var/lib/kubelet/pods/f69a3b21-eb1c-4300-91dc-55766900da95/volume-subpaths: remove /var/lib/kubelet/pods/f69a3b21-eb1c-4300-91dc-55766900da95/volume-subpaths: no such file or directory Mar 13 10:41:13 crc kubenswrapper[4632]: I0313 10:41:13.992048 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f69a3b21-eb1c-4300-91dc-55766900da95" (UID: "f69a3b21-eb1c-4300-91dc-55766900da95"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.067155 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.067528 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6d8hr\" (UniqueName: \"kubernetes.io/projected/f69a3b21-eb1c-4300-91dc-55766900da95-kube-api-access-6d8hr\") on node \"crc\" DevicePath \"\"" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.393007 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" event={"ID":"f69a3b21-eb1c-4300-91dc-55766900da95","Type":"ContainerDied","Data":"c652193575851bc88eb9652edd9cc1fa8aa0f642f4ed0ff9f52d97763733bab8"} Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.393055 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c652193575851bc88eb9652edd9cc1fa8aa0f642f4ed0ff9f52d97763733bab8" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.393085 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-f9rbk" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.471617 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz"] Mar 13 10:41:14 crc kubenswrapper[4632]: E0313 10:41:14.472025 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f69a3b21-eb1c-4300-91dc-55766900da95" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.472042 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f69a3b21-eb1c-4300-91dc-55766900da95" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.472255 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f69a3b21-eb1c-4300-91dc-55766900da95" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.472974 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.494303 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz"] Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.576424 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-inventory\") pod \"f69a3b21-eb1c-4300-91dc-55766900da95\" (UID: \"f69a3b21-eb1c-4300-91dc-55766900da95\") " Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.576962 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cztz\" (UniqueName: \"kubernetes.io/projected/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-kube-api-access-6cztz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz\" (UID: \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.577020 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz\" (UID: \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.577090 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz\" (UID: \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.592632 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-inventory" (OuterVolumeSpecName: "inventory") pod "f69a3b21-eb1c-4300-91dc-55766900da95" (UID: "f69a3b21-eb1c-4300-91dc-55766900da95"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.678925 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz\" (UID: \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.679052 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz\" (UID: \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.679175 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cztz\" (UniqueName: \"kubernetes.io/projected/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-kube-api-access-6cztz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz\" (UID: \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.679225 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f69a3b21-eb1c-4300-91dc-55766900da95-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.682750 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz\" (UID: \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.683038 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz\" (UID: \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.696109 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cztz\" (UniqueName: \"kubernetes.io/projected/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-kube-api-access-6cztz\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz\" (UID: \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" Mar 13 10:41:14 crc kubenswrapper[4632]: I0313 10:41:14.794879 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" Mar 13 10:41:15 crc kubenswrapper[4632]: I0313 10:41:15.245605 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c6vkp" podUID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerName="registry-server" probeResult="failure" output=< Mar 13 10:41:15 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:41:15 crc kubenswrapper[4632]: > Mar 13 10:41:15 crc kubenswrapper[4632]: I0313 10:41:15.383736 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz"] Mar 13 10:41:15 crc kubenswrapper[4632]: I0313 10:41:15.406821 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" event={"ID":"4eaeef27-fa4c-41d9-a197-a780a6a6cebd","Type":"ContainerStarted","Data":"ced96d05e4eec27d78e68d54f820e0f8c837211b9f85da7592e625e2f0f9fcaa"} Mar 13 10:41:16 crc kubenswrapper[4632]: I0313 10:41:16.519062 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" event={"ID":"4eaeef27-fa4c-41d9-a197-a780a6a6cebd","Type":"ContainerStarted","Data":"cd71d4c7541e14dbca5f760d64290b31a098a75a4bb28c7734560d76d0b91bde"} Mar 13 10:41:16 crc kubenswrapper[4632]: I0313 10:41:16.537764 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" podStartSLOduration=2.060925844 podStartE2EDuration="2.537742448s" podCreationTimestamp="2026-03-13 10:41:14 +0000 UTC" firstStartedPulling="2026-03-13 10:41:15.383346301 +0000 UTC m=+2249.405876444" lastFinishedPulling="2026-03-13 10:41:15.860162915 +0000 UTC m=+2249.882693048" observedRunningTime="2026-03-13 10:41:16.535239856 +0000 UTC m=+2250.557769989" watchObservedRunningTime="2026-03-13 10:41:16.537742448 +0000 UTC m=+2250.560272601" Mar 13 10:41:25 crc kubenswrapper[4632]: I0313 10:41:25.247957 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c6vkp" podUID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerName="registry-server" probeResult="failure" output=< Mar 13 10:41:25 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:41:25 crc kubenswrapper[4632]: > Mar 13 10:41:27 crc kubenswrapper[4632]: I0313 10:41:27.621164 4632 generic.go:334] "Generic (PLEG): container finished" podID="4eaeef27-fa4c-41d9-a197-a780a6a6cebd" containerID="cd71d4c7541e14dbca5f760d64290b31a098a75a4bb28c7734560d76d0b91bde" exitCode=0 Mar 13 10:41:27 crc kubenswrapper[4632]: I0313 10:41:27.621233 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" event={"ID":"4eaeef27-fa4c-41d9-a197-a780a6a6cebd","Type":"ContainerDied","Data":"cd71d4c7541e14dbca5f760d64290b31a098a75a4bb28c7734560d76d0b91bde"} Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.076676 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.111421 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cztz\" (UniqueName: \"kubernetes.io/projected/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-kube-api-access-6cztz\") pod \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\" (UID: \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\") " Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.111589 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-ssh-key-openstack-edpm-ipam\") pod \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\" (UID: \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\") " Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.111618 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-inventory\") pod \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\" (UID: \"4eaeef27-fa4c-41d9-a197-a780a6a6cebd\") " Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.118389 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-kube-api-access-6cztz" (OuterVolumeSpecName: "kube-api-access-6cztz") pod "4eaeef27-fa4c-41d9-a197-a780a6a6cebd" (UID: "4eaeef27-fa4c-41d9-a197-a780a6a6cebd"). InnerVolumeSpecName "kube-api-access-6cztz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.146229 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-inventory" (OuterVolumeSpecName: "inventory") pod "4eaeef27-fa4c-41d9-a197-a780a6a6cebd" (UID: "4eaeef27-fa4c-41d9-a197-a780a6a6cebd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.148873 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4eaeef27-fa4c-41d9-a197-a780a6a6cebd" (UID: "4eaeef27-fa4c-41d9-a197-a780a6a6cebd"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.213870 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cztz\" (UniqueName: \"kubernetes.io/projected/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-kube-api-access-6cztz\") on node \"crc\" DevicePath \"\"" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.213912 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.213954 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4eaeef27-fa4c-41d9-a197-a780a6a6cebd-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.639549 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" event={"ID":"4eaeef27-fa4c-41d9-a197-a780a6a6cebd","Type":"ContainerDied","Data":"ced96d05e4eec27d78e68d54f820e0f8c837211b9f85da7592e625e2f0f9fcaa"} Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.639590 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ced96d05e4eec27d78e68d54f820e0f8c837211b9f85da7592e625e2f0f9fcaa" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.639641 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.741735 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn"] Mar 13 10:41:29 crc kubenswrapper[4632]: E0313 10:41:29.742259 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4eaeef27-fa4c-41d9-a197-a780a6a6cebd" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.742287 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="4eaeef27-fa4c-41d9-a197-a780a6a6cebd" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.742569 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="4eaeef27-fa4c-41d9-a197-a780a6a6cebd" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.743732 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.746012 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.746105 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.746756 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.748347 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.748850 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.750959 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.751148 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.751460 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.781554 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn"] Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.825445 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.825503 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzlkb\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-kube-api-access-rzlkb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.825538 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.825561 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.825587 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.825609 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.825645 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.825683 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.825747 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.825766 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.825799 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.825862 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.825907 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.825949 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.927549 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.927616 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzlkb\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-kube-api-access-rzlkb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.927655 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.927686 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.927753 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.927786 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.927834 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.927884 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.927996 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.928024 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.928067 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.928144 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.928197 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.928236 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.935749 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.938235 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.939281 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.940043 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.940749 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.942701 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.943150 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.945364 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.948704 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.949870 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.950860 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzlkb\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-kube-api-access-rzlkb\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.951198 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.952056 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:29 crc kubenswrapper[4632]: I0313 10:41:29.960322 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:30 crc kubenswrapper[4632]: I0313 10:41:30.066635 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:41:30 crc kubenswrapper[4632]: I0313 10:41:30.602432 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn"] Mar 13 10:41:30 crc kubenswrapper[4632]: I0313 10:41:30.650957 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" event={"ID":"41861d23-3e34-4f91-bafc-1b7eeee125db","Type":"ContainerStarted","Data":"aa943b891e21fa2ddf52c1882c6b609ca996747cf346a718083b0fc64cf76d58"} Mar 13 10:41:31 crc kubenswrapper[4632]: I0313 10:41:31.667811 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" event={"ID":"41861d23-3e34-4f91-bafc-1b7eeee125db","Type":"ContainerStarted","Data":"8bcf160168d0bd44e7bdc4cd090dfd9a1207f38af53c4c136868bdc7acd63fcf"} Mar 13 10:41:31 crc kubenswrapper[4632]: I0313 10:41:31.690027 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" podStartSLOduration=2.279689938 podStartE2EDuration="2.690008032s" podCreationTimestamp="2026-03-13 10:41:29 +0000 UTC" firstStartedPulling="2026-03-13 10:41:30.611613949 +0000 UTC m=+2264.634144072" lastFinishedPulling="2026-03-13 10:41:31.021932043 +0000 UTC m=+2265.044462166" observedRunningTime="2026-03-13 10:41:31.687954142 +0000 UTC m=+2265.710484275" watchObservedRunningTime="2026-03-13 10:41:31.690008032 +0000 UTC m=+2265.712538165" Mar 13 10:41:35 crc kubenswrapper[4632]: I0313 10:41:35.243029 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-c6vkp" podUID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerName="registry-server" probeResult="failure" output=< Mar 13 10:41:35 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:41:35 crc kubenswrapper[4632]: > Mar 13 10:41:40 crc kubenswrapper[4632]: I0313 10:41:40.460856 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:41:40 crc kubenswrapper[4632]: I0313 10:41:40.461374 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:41:44 crc kubenswrapper[4632]: I0313 10:41:44.248638 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:41:44 crc kubenswrapper[4632]: I0313 10:41:44.302160 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:41:44 crc kubenswrapper[4632]: I0313 10:41:44.491991 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c6vkp"] Mar 13 10:41:45 crc kubenswrapper[4632]: I0313 10:41:45.786511 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-c6vkp" podUID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerName="registry-server" containerID="cri-o://c7a9c2f67c2bc9ec6015b2a50c3cfdd53b590fc7fda70e8b949fe66cb4e0aff2" gracePeriod=2 Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.384483 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.480684 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94fd5d57-5fb3-4b34-a545-6bdc3f219354-utilities\") pod \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\" (UID: \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\") " Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.480721 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94fd5d57-5fb3-4b34-a545-6bdc3f219354-catalog-content\") pod \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\" (UID: \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\") " Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.480934 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-srngn\" (UniqueName: \"kubernetes.io/projected/94fd5d57-5fb3-4b34-a545-6bdc3f219354-kube-api-access-srngn\") pod \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\" (UID: \"94fd5d57-5fb3-4b34-a545-6bdc3f219354\") " Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.490460 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94fd5d57-5fb3-4b34-a545-6bdc3f219354-kube-api-access-srngn" (OuterVolumeSpecName: "kube-api-access-srngn") pod "94fd5d57-5fb3-4b34-a545-6bdc3f219354" (UID: "94fd5d57-5fb3-4b34-a545-6bdc3f219354"). InnerVolumeSpecName "kube-api-access-srngn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.498714 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94fd5d57-5fb3-4b34-a545-6bdc3f219354-utilities" (OuterVolumeSpecName: "utilities") pod "94fd5d57-5fb3-4b34-a545-6bdc3f219354" (UID: "94fd5d57-5fb3-4b34-a545-6bdc3f219354"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.583563 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94fd5d57-5fb3-4b34-a545-6bdc3f219354-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.583600 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-srngn\" (UniqueName: \"kubernetes.io/projected/94fd5d57-5fb3-4b34-a545-6bdc3f219354-kube-api-access-srngn\") on node \"crc\" DevicePath \"\"" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.669140 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94fd5d57-5fb3-4b34-a545-6bdc3f219354-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94fd5d57-5fb3-4b34-a545-6bdc3f219354" (UID: "94fd5d57-5fb3-4b34-a545-6bdc3f219354"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.685175 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94fd5d57-5fb3-4b34-a545-6bdc3f219354-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.800720 4632 generic.go:334] "Generic (PLEG): container finished" podID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerID="c7a9c2f67c2bc9ec6015b2a50c3cfdd53b590fc7fda70e8b949fe66cb4e0aff2" exitCode=0 Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.800798 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6vkp" event={"ID":"94fd5d57-5fb3-4b34-a545-6bdc3f219354","Type":"ContainerDied","Data":"c7a9c2f67c2bc9ec6015b2a50c3cfdd53b590fc7fda70e8b949fe66cb4e0aff2"} Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.800846 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-c6vkp" event={"ID":"94fd5d57-5fb3-4b34-a545-6bdc3f219354","Type":"ContainerDied","Data":"cfe1bf74852782b72ab02590cc17b2a400ee246e92a62ec6365c71796dd98bad"} Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.800849 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-c6vkp" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.800872 4632 scope.go:117] "RemoveContainer" containerID="c7a9c2f67c2bc9ec6015b2a50c3cfdd53b590fc7fda70e8b949fe66cb4e0aff2" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.853406 4632 scope.go:117] "RemoveContainer" containerID="e5b275bdebfc8c0e84bc1511d86355567607af8b44b7d9fd25e952a957898f8e" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.854064 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-c6vkp"] Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.863143 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-c6vkp"] Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.892927 4632 scope.go:117] "RemoveContainer" containerID="d6c5924afdfbc6243bd936794f31e2092a18f09ea985a6206e818699a4eab9bc" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.930784 4632 scope.go:117] "RemoveContainer" containerID="c7a9c2f67c2bc9ec6015b2a50c3cfdd53b590fc7fda70e8b949fe66cb4e0aff2" Mar 13 10:41:46 crc kubenswrapper[4632]: E0313 10:41:46.931440 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7a9c2f67c2bc9ec6015b2a50c3cfdd53b590fc7fda70e8b949fe66cb4e0aff2\": container with ID starting with c7a9c2f67c2bc9ec6015b2a50c3cfdd53b590fc7fda70e8b949fe66cb4e0aff2 not found: ID does not exist" containerID="c7a9c2f67c2bc9ec6015b2a50c3cfdd53b590fc7fda70e8b949fe66cb4e0aff2" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.931489 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7a9c2f67c2bc9ec6015b2a50c3cfdd53b590fc7fda70e8b949fe66cb4e0aff2"} err="failed to get container status \"c7a9c2f67c2bc9ec6015b2a50c3cfdd53b590fc7fda70e8b949fe66cb4e0aff2\": rpc error: code = NotFound desc = could not find container \"c7a9c2f67c2bc9ec6015b2a50c3cfdd53b590fc7fda70e8b949fe66cb4e0aff2\": container with ID starting with c7a9c2f67c2bc9ec6015b2a50c3cfdd53b590fc7fda70e8b949fe66cb4e0aff2 not found: ID does not exist" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.931518 4632 scope.go:117] "RemoveContainer" containerID="e5b275bdebfc8c0e84bc1511d86355567607af8b44b7d9fd25e952a957898f8e" Mar 13 10:41:46 crc kubenswrapper[4632]: E0313 10:41:46.931806 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e5b275bdebfc8c0e84bc1511d86355567607af8b44b7d9fd25e952a957898f8e\": container with ID starting with e5b275bdebfc8c0e84bc1511d86355567607af8b44b7d9fd25e952a957898f8e not found: ID does not exist" containerID="e5b275bdebfc8c0e84bc1511d86355567607af8b44b7d9fd25e952a957898f8e" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.931830 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e5b275bdebfc8c0e84bc1511d86355567607af8b44b7d9fd25e952a957898f8e"} err="failed to get container status \"e5b275bdebfc8c0e84bc1511d86355567607af8b44b7d9fd25e952a957898f8e\": rpc error: code = NotFound desc = could not find container \"e5b275bdebfc8c0e84bc1511d86355567607af8b44b7d9fd25e952a957898f8e\": container with ID starting with e5b275bdebfc8c0e84bc1511d86355567607af8b44b7d9fd25e952a957898f8e not found: ID does not exist" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.931844 4632 scope.go:117] "RemoveContainer" containerID="d6c5924afdfbc6243bd936794f31e2092a18f09ea985a6206e818699a4eab9bc" Mar 13 10:41:46 crc kubenswrapper[4632]: E0313 10:41:46.932154 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6c5924afdfbc6243bd936794f31e2092a18f09ea985a6206e818699a4eab9bc\": container with ID starting with d6c5924afdfbc6243bd936794f31e2092a18f09ea985a6206e818699a4eab9bc not found: ID does not exist" containerID="d6c5924afdfbc6243bd936794f31e2092a18f09ea985a6206e818699a4eab9bc" Mar 13 10:41:46 crc kubenswrapper[4632]: I0313 10:41:46.932175 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6c5924afdfbc6243bd936794f31e2092a18f09ea985a6206e818699a4eab9bc"} err="failed to get container status \"d6c5924afdfbc6243bd936794f31e2092a18f09ea985a6206e818699a4eab9bc\": rpc error: code = NotFound desc = could not find container \"d6c5924afdfbc6243bd936794f31e2092a18f09ea985a6206e818699a4eab9bc\": container with ID starting with d6c5924afdfbc6243bd936794f31e2092a18f09ea985a6206e818699a4eab9bc not found: ID does not exist" Mar 13 10:41:48 crc kubenswrapper[4632]: I0313 10:41:48.055729 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" path="/var/lib/kubelet/pods/94fd5d57-5fb3-4b34-a545-6bdc3f219354/volumes" Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.148178 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556642-f6lzb"] Mar 13 10:42:00 crc kubenswrapper[4632]: E0313 10:42:00.150135 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerName="extract-utilities" Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.150245 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerName="extract-utilities" Mar 13 10:42:00 crc kubenswrapper[4632]: E0313 10:42:00.150339 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerName="extract-content" Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.150405 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerName="extract-content" Mar 13 10:42:00 crc kubenswrapper[4632]: E0313 10:42:00.150473 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerName="registry-server" Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.150536 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerName="registry-server" Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.150813 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="94fd5d57-5fb3-4b34-a545-6bdc3f219354" containerName="registry-server" Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.152693 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556642-f6lzb" Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.154592 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.155472 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.155782 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.161094 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556642-f6lzb"] Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.279236 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8lqks\" (UniqueName: \"kubernetes.io/projected/a6df4a28-3b7b-4904-aa41-62caa26889a8-kube-api-access-8lqks\") pod \"auto-csr-approver-29556642-f6lzb\" (UID: \"a6df4a28-3b7b-4904-aa41-62caa26889a8\") " pod="openshift-infra/auto-csr-approver-29556642-f6lzb" Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.381073 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8lqks\" (UniqueName: \"kubernetes.io/projected/a6df4a28-3b7b-4904-aa41-62caa26889a8-kube-api-access-8lqks\") pod \"auto-csr-approver-29556642-f6lzb\" (UID: \"a6df4a28-3b7b-4904-aa41-62caa26889a8\") " pod="openshift-infra/auto-csr-approver-29556642-f6lzb" Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.405638 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8lqks\" (UniqueName: \"kubernetes.io/projected/a6df4a28-3b7b-4904-aa41-62caa26889a8-kube-api-access-8lqks\") pod \"auto-csr-approver-29556642-f6lzb\" (UID: \"a6df4a28-3b7b-4904-aa41-62caa26889a8\") " pod="openshift-infra/auto-csr-approver-29556642-f6lzb" Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.476633 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556642-f6lzb" Mar 13 10:42:00 crc kubenswrapper[4632]: I0313 10:42:00.977322 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556642-f6lzb"] Mar 13 10:42:01 crc kubenswrapper[4632]: I0313 10:42:01.941129 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556642-f6lzb" event={"ID":"a6df4a28-3b7b-4904-aa41-62caa26889a8","Type":"ContainerStarted","Data":"f9e840b0d2d0ed339be1bb29f4a60a1df28737f0945b06042560ed3f762bf5b1"} Mar 13 10:42:02 crc kubenswrapper[4632]: I0313 10:42:02.952693 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556642-f6lzb" event={"ID":"a6df4a28-3b7b-4904-aa41-62caa26889a8","Type":"ContainerStarted","Data":"2360a4309504ac747bcde26fcceae28cb04d811f34c5f4e463b65c45b06c70f5"} Mar 13 10:42:02 crc kubenswrapper[4632]: I0313 10:42:02.975079 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556642-f6lzb" podStartSLOduration=1.969066635 podStartE2EDuration="2.975052425s" podCreationTimestamp="2026-03-13 10:42:00 +0000 UTC" firstStartedPulling="2026-03-13 10:42:00.998567251 +0000 UTC m=+2295.021097374" lastFinishedPulling="2026-03-13 10:42:02.004553031 +0000 UTC m=+2296.027083164" observedRunningTime="2026-03-13 10:42:02.965693555 +0000 UTC m=+2296.988223688" watchObservedRunningTime="2026-03-13 10:42:02.975052425 +0000 UTC m=+2296.997582558" Mar 13 10:42:03 crc kubenswrapper[4632]: I0313 10:42:03.965851 4632 generic.go:334] "Generic (PLEG): container finished" podID="a6df4a28-3b7b-4904-aa41-62caa26889a8" containerID="2360a4309504ac747bcde26fcceae28cb04d811f34c5f4e463b65c45b06c70f5" exitCode=0 Mar 13 10:42:03 crc kubenswrapper[4632]: I0313 10:42:03.965897 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556642-f6lzb" event={"ID":"a6df4a28-3b7b-4904-aa41-62caa26889a8","Type":"ContainerDied","Data":"2360a4309504ac747bcde26fcceae28cb04d811f34c5f4e463b65c45b06c70f5"} Mar 13 10:42:05 crc kubenswrapper[4632]: I0313 10:42:05.412003 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556642-f6lzb" Mar 13 10:42:05 crc kubenswrapper[4632]: I0313 10:42:05.492163 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8lqks\" (UniqueName: \"kubernetes.io/projected/a6df4a28-3b7b-4904-aa41-62caa26889a8-kube-api-access-8lqks\") pod \"a6df4a28-3b7b-4904-aa41-62caa26889a8\" (UID: \"a6df4a28-3b7b-4904-aa41-62caa26889a8\") " Mar 13 10:42:05 crc kubenswrapper[4632]: I0313 10:42:05.499574 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6df4a28-3b7b-4904-aa41-62caa26889a8-kube-api-access-8lqks" (OuterVolumeSpecName: "kube-api-access-8lqks") pod "a6df4a28-3b7b-4904-aa41-62caa26889a8" (UID: "a6df4a28-3b7b-4904-aa41-62caa26889a8"). InnerVolumeSpecName "kube-api-access-8lqks". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:42:05 crc kubenswrapper[4632]: I0313 10:42:05.595522 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8lqks\" (UniqueName: \"kubernetes.io/projected/a6df4a28-3b7b-4904-aa41-62caa26889a8-kube-api-access-8lqks\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:05 crc kubenswrapper[4632]: I0313 10:42:05.982519 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556642-f6lzb" event={"ID":"a6df4a28-3b7b-4904-aa41-62caa26889a8","Type":"ContainerDied","Data":"f9e840b0d2d0ed339be1bb29f4a60a1df28737f0945b06042560ed3f762bf5b1"} Mar 13 10:42:05 crc kubenswrapper[4632]: I0313 10:42:05.982556 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f9e840b0d2d0ed339be1bb29f4a60a1df28737f0945b06042560ed3f762bf5b1" Mar 13 10:42:05 crc kubenswrapper[4632]: I0313 10:42:05.982562 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556642-f6lzb" Mar 13 10:42:06 crc kubenswrapper[4632]: I0313 10:42:06.093709 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556636-zncpw"] Mar 13 10:42:06 crc kubenswrapper[4632]: I0313 10:42:06.112311 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556636-zncpw"] Mar 13 10:42:08 crc kubenswrapper[4632]: I0313 10:42:08.064318 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7cb201b3-b479-4877-a996-58045d0720c4" path="/var/lib/kubelet/pods/7cb201b3-b479-4877-a996-58045d0720c4/volumes" Mar 13 10:42:09 crc kubenswrapper[4632]: I0313 10:42:09.008996 4632 generic.go:334] "Generic (PLEG): container finished" podID="41861d23-3e34-4f91-bafc-1b7eeee125db" containerID="8bcf160168d0bd44e7bdc4cd090dfd9a1207f38af53c4c136868bdc7acd63fcf" exitCode=0 Mar 13 10:42:09 crc kubenswrapper[4632]: I0313 10:42:09.009119 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" event={"ID":"41861d23-3e34-4f91-bafc-1b7eeee125db","Type":"ContainerDied","Data":"8bcf160168d0bd44e7bdc4cd090dfd9a1207f38af53c4c136868bdc7acd63fcf"} Mar 13 10:42:09 crc kubenswrapper[4632]: I0313 10:42:09.379651 4632 scope.go:117] "RemoveContainer" containerID="1d5789598fed395c0d259939fb11bb98aa8eec3b7168c00349a4a3635d4bd5ce" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.433385 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.464635 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.464686 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.519967 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-telemetry-combined-ca-bundle\") pod \"41861d23-3e34-4f91-bafc-1b7eeee125db\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.520051 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"41861d23-3e34-4f91-bafc-1b7eeee125db\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.520121 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-inventory\") pod \"41861d23-3e34-4f91-bafc-1b7eeee125db\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.520157 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzlkb\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-kube-api-access-rzlkb\") pod \"41861d23-3e34-4f91-bafc-1b7eeee125db\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.520192 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-neutron-metadata-combined-ca-bundle\") pod \"41861d23-3e34-4f91-bafc-1b7eeee125db\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.520332 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"41861d23-3e34-4f91-bafc-1b7eeee125db\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.520372 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-repo-setup-combined-ca-bundle\") pod \"41861d23-3e34-4f91-bafc-1b7eeee125db\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.520434 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-libvirt-combined-ca-bundle\") pod \"41861d23-3e34-4f91-bafc-1b7eeee125db\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.520470 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-ovn-combined-ca-bundle\") pod \"41861d23-3e34-4f91-bafc-1b7eeee125db\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.520539 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-bootstrap-combined-ca-bundle\") pod \"41861d23-3e34-4f91-bafc-1b7eeee125db\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.520578 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-ovn-default-certs-0\") pod \"41861d23-3e34-4f91-bafc-1b7eeee125db\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.520634 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-ssh-key-openstack-edpm-ipam\") pod \"41861d23-3e34-4f91-bafc-1b7eeee125db\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.520664 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"41861d23-3e34-4f91-bafc-1b7eeee125db\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.520696 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-nova-combined-ca-bundle\") pod \"41861d23-3e34-4f91-bafc-1b7eeee125db\" (UID: \"41861d23-3e34-4f91-bafc-1b7eeee125db\") " Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.527679 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "41861d23-3e34-4f91-bafc-1b7eeee125db" (UID: "41861d23-3e34-4f91-bafc-1b7eeee125db"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.529107 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-kube-api-access-rzlkb" (OuterVolumeSpecName: "kube-api-access-rzlkb") pod "41861d23-3e34-4f91-bafc-1b7eeee125db" (UID: "41861d23-3e34-4f91-bafc-1b7eeee125db"). InnerVolumeSpecName "kube-api-access-rzlkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.530201 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "41861d23-3e34-4f91-bafc-1b7eeee125db" (UID: "41861d23-3e34-4f91-bafc-1b7eeee125db"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.530571 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "41861d23-3e34-4f91-bafc-1b7eeee125db" (UID: "41861d23-3e34-4f91-bafc-1b7eeee125db"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.531065 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "41861d23-3e34-4f91-bafc-1b7eeee125db" (UID: "41861d23-3e34-4f91-bafc-1b7eeee125db"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.532337 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "41861d23-3e34-4f91-bafc-1b7eeee125db" (UID: "41861d23-3e34-4f91-bafc-1b7eeee125db"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.532810 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "41861d23-3e34-4f91-bafc-1b7eeee125db" (UID: "41861d23-3e34-4f91-bafc-1b7eeee125db"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.533597 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "41861d23-3e34-4f91-bafc-1b7eeee125db" (UID: "41861d23-3e34-4f91-bafc-1b7eeee125db"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.534096 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "41861d23-3e34-4f91-bafc-1b7eeee125db" (UID: "41861d23-3e34-4f91-bafc-1b7eeee125db"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.535222 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "41861d23-3e34-4f91-bafc-1b7eeee125db" (UID: "41861d23-3e34-4f91-bafc-1b7eeee125db"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.549111 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "41861d23-3e34-4f91-bafc-1b7eeee125db" (UID: "41861d23-3e34-4f91-bafc-1b7eeee125db"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.549663 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "41861d23-3e34-4f91-bafc-1b7eeee125db" (UID: "41861d23-3e34-4f91-bafc-1b7eeee125db"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.561607 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-inventory" (OuterVolumeSpecName: "inventory") pod "41861d23-3e34-4f91-bafc-1b7eeee125db" (UID: "41861d23-3e34-4f91-bafc-1b7eeee125db"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.567014 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "41861d23-3e34-4f91-bafc-1b7eeee125db" (UID: "41861d23-3e34-4f91-bafc-1b7eeee125db"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.624856 4632 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.625096 4632 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.625162 4632 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.625265 4632 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.625360 4632 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.625479 4632 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.625617 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.625698 4632 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.625770 4632 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.625834 4632 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.625898 4632 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.625976 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.626093 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzlkb\" (UniqueName: \"kubernetes.io/projected/41861d23-3e34-4f91-bafc-1b7eeee125db-kube-api-access-rzlkb\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:10 crc kubenswrapper[4632]: I0313 10:42:10.626166 4632 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41861d23-3e34-4f91-bafc-1b7eeee125db-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.030108 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" event={"ID":"41861d23-3e34-4f91-bafc-1b7eeee125db","Type":"ContainerDied","Data":"aa943b891e21fa2ddf52c1882c6b609ca996747cf346a718083b0fc64cf76d58"} Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.030153 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa943b891e21fa2ddf52c1882c6b609ca996747cf346a718083b0fc64cf76d58" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.030235 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.124424 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6"] Mar 13 10:42:11 crc kubenswrapper[4632]: E0313 10:42:11.124871 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6df4a28-3b7b-4904-aa41-62caa26889a8" containerName="oc" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.124896 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6df4a28-3b7b-4904-aa41-62caa26889a8" containerName="oc" Mar 13 10:42:11 crc kubenswrapper[4632]: E0313 10:42:11.124916 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41861d23-3e34-4f91-bafc-1b7eeee125db" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.124927 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="41861d23-3e34-4f91-bafc-1b7eeee125db" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.125180 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6df4a28-3b7b-4904-aa41-62caa26889a8" containerName="oc" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.125225 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="41861d23-3e34-4f91-bafc-1b7eeee125db" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.125992 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.128708 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.138055 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.138133 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.138365 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.144591 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.150789 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6"] Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.237771 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.238218 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.239237 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/96ca1247-6625-4b08-b155-34c56f02ec04-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.239439 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdp7x\" (UniqueName: \"kubernetes.io/projected/96ca1247-6625-4b08-b155-34c56f02ec04-kube-api-access-wdp7x\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.240014 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.342077 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.342222 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.342258 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.342338 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/96ca1247-6625-4b08-b155-34c56f02ec04-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.342369 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdp7x\" (UniqueName: \"kubernetes.io/projected/96ca1247-6625-4b08-b155-34c56f02ec04-kube-api-access-wdp7x\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.344043 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/96ca1247-6625-4b08-b155-34c56f02ec04-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.348557 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.348742 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.358764 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.373845 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdp7x\" (UniqueName: \"kubernetes.io/projected/96ca1247-6625-4b08-b155-34c56f02ec04-kube-api-access-wdp7x\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-6t9b6\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.453874 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:42:11 crc kubenswrapper[4632]: I0313 10:42:11.997356 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6"] Mar 13 10:42:12 crc kubenswrapper[4632]: I0313 10:42:12.040471 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" event={"ID":"96ca1247-6625-4b08-b155-34c56f02ec04","Type":"ContainerStarted","Data":"7cff807ce30996fc61ff5a53f3c16703d17afe3357255159f2daf0f8d5320802"} Mar 13 10:42:13 crc kubenswrapper[4632]: I0313 10:42:13.055054 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" event={"ID":"96ca1247-6625-4b08-b155-34c56f02ec04","Type":"ContainerStarted","Data":"b9c8440031ae4a9073037d4c844efc1d48630f961db4b97500123a9880df8925"} Mar 13 10:42:13 crc kubenswrapper[4632]: I0313 10:42:13.084079 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" podStartSLOduration=1.352222222 podStartE2EDuration="2.084059384s" podCreationTimestamp="2026-03-13 10:42:11 +0000 UTC" firstStartedPulling="2026-03-13 10:42:12.002032139 +0000 UTC m=+2306.024562272" lastFinishedPulling="2026-03-13 10:42:12.733869301 +0000 UTC m=+2306.756399434" observedRunningTime="2026-03-13 10:42:13.073523275 +0000 UTC m=+2307.096053438" watchObservedRunningTime="2026-03-13 10:42:13.084059384 +0000 UTC m=+2307.106589517" Mar 13 10:42:34 crc kubenswrapper[4632]: I0313 10:42:34.246827 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gnw2p"] Mar 13 10:42:34 crc kubenswrapper[4632]: I0313 10:42:34.249207 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:34 crc kubenswrapper[4632]: I0313 10:42:34.266446 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gnw2p"] Mar 13 10:42:34 crc kubenswrapper[4632]: I0313 10:42:34.277585 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-utilities\") pod \"community-operators-gnw2p\" (UID: \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\") " pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:34 crc kubenswrapper[4632]: I0313 10:42:34.277735 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-catalog-content\") pod \"community-operators-gnw2p\" (UID: \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\") " pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:34 crc kubenswrapper[4632]: I0313 10:42:34.277827 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlqlv\" (UniqueName: \"kubernetes.io/projected/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-kube-api-access-vlqlv\") pod \"community-operators-gnw2p\" (UID: \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\") " pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:34 crc kubenswrapper[4632]: I0313 10:42:34.379778 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-utilities\") pod \"community-operators-gnw2p\" (UID: \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\") " pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:34 crc kubenswrapper[4632]: I0313 10:42:34.379852 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-catalog-content\") pod \"community-operators-gnw2p\" (UID: \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\") " pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:34 crc kubenswrapper[4632]: I0313 10:42:34.379929 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlqlv\" (UniqueName: \"kubernetes.io/projected/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-kube-api-access-vlqlv\") pod \"community-operators-gnw2p\" (UID: \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\") " pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:34 crc kubenswrapper[4632]: I0313 10:42:34.380407 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-utilities\") pod \"community-operators-gnw2p\" (UID: \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\") " pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:34 crc kubenswrapper[4632]: I0313 10:42:34.380432 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-catalog-content\") pod \"community-operators-gnw2p\" (UID: \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\") " pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:34 crc kubenswrapper[4632]: I0313 10:42:34.402390 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlqlv\" (UniqueName: \"kubernetes.io/projected/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-kube-api-access-vlqlv\") pod \"community-operators-gnw2p\" (UID: \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\") " pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:34 crc kubenswrapper[4632]: I0313 10:42:34.576733 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:35 crc kubenswrapper[4632]: I0313 10:42:35.217025 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gnw2p"] Mar 13 10:42:35 crc kubenswrapper[4632]: I0313 10:42:35.294568 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnw2p" event={"ID":"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d","Type":"ContainerStarted","Data":"5d8fa1de309b7087e5b0807d70e759ae4655eeeff831ca9798400cbf8bccd783"} Mar 13 10:42:36 crc kubenswrapper[4632]: I0313 10:42:36.306926 4632 generic.go:334] "Generic (PLEG): container finished" podID="cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" containerID="3dd84647257d4d428f925670105498da793e6d9758ec9cb89c02c2e2a54e558f" exitCode=0 Mar 13 10:42:36 crc kubenswrapper[4632]: I0313 10:42:36.307139 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnw2p" event={"ID":"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d","Type":"ContainerDied","Data":"3dd84647257d4d428f925670105498da793e6d9758ec9cb89c02c2e2a54e558f"} Mar 13 10:42:37 crc kubenswrapper[4632]: I0313 10:42:37.318302 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnw2p" event={"ID":"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d","Type":"ContainerStarted","Data":"8702aa0a52c8778ee09c744c907f2b88b34171b80b2c62359a090363a7ecc8ac"} Mar 13 10:42:38 crc kubenswrapper[4632]: I0313 10:42:38.329981 4632 generic.go:334] "Generic (PLEG): container finished" podID="cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" containerID="8702aa0a52c8778ee09c744c907f2b88b34171b80b2c62359a090363a7ecc8ac" exitCode=0 Mar 13 10:42:38 crc kubenswrapper[4632]: I0313 10:42:38.330341 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnw2p" event={"ID":"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d","Type":"ContainerDied","Data":"8702aa0a52c8778ee09c744c907f2b88b34171b80b2c62359a090363a7ecc8ac"} Mar 13 10:42:39 crc kubenswrapper[4632]: I0313 10:42:39.342426 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnw2p" event={"ID":"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d","Type":"ContainerStarted","Data":"81c2fc126faa38cd627c0814225a882d113c0439776ae9c08105c81325a2f0d5"} Mar 13 10:42:39 crc kubenswrapper[4632]: I0313 10:42:39.375910 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gnw2p" podStartSLOduration=2.8662370839999998 podStartE2EDuration="5.375892s" podCreationTimestamp="2026-03-13 10:42:34 +0000 UTC" firstStartedPulling="2026-03-13 10:42:36.311667874 +0000 UTC m=+2330.334198007" lastFinishedPulling="2026-03-13 10:42:38.82132279 +0000 UTC m=+2332.843852923" observedRunningTime="2026-03-13 10:42:39.365892244 +0000 UTC m=+2333.388422387" watchObservedRunningTime="2026-03-13 10:42:39.375892 +0000 UTC m=+2333.398422133" Mar 13 10:42:40 crc kubenswrapper[4632]: I0313 10:42:40.461431 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:42:40 crc kubenswrapper[4632]: I0313 10:42:40.461757 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:42:40 crc kubenswrapper[4632]: I0313 10:42:40.461818 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:42:40 crc kubenswrapper[4632]: I0313 10:42:40.462926 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:42:40 crc kubenswrapper[4632]: I0313 10:42:40.463023 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" gracePeriod=600 Mar 13 10:42:40 crc kubenswrapper[4632]: E0313 10:42:40.593310 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:42:41 crc kubenswrapper[4632]: I0313 10:42:41.360184 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" exitCode=0 Mar 13 10:42:41 crc kubenswrapper[4632]: I0313 10:42:41.360710 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20"} Mar 13 10:42:41 crc kubenswrapper[4632]: I0313 10:42:41.360760 4632 scope.go:117] "RemoveContainer" containerID="2bb4e222f4f89a1d4e4bebc809fc60cc762d7ea9b6811f4bcc9cb78c179cd0bd" Mar 13 10:42:41 crc kubenswrapper[4632]: I0313 10:42:41.361441 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:42:41 crc kubenswrapper[4632]: E0313 10:42:41.361744 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:42:44 crc kubenswrapper[4632]: I0313 10:42:44.578052 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:44 crc kubenswrapper[4632]: I0313 10:42:44.578393 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:44 crc kubenswrapper[4632]: I0313 10:42:44.630381 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:45 crc kubenswrapper[4632]: I0313 10:42:45.463552 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:45 crc kubenswrapper[4632]: I0313 10:42:45.515359 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gnw2p"] Mar 13 10:42:47 crc kubenswrapper[4632]: I0313 10:42:47.432027 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gnw2p" podUID="cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" containerName="registry-server" containerID="cri-o://81c2fc126faa38cd627c0814225a882d113c0439776ae9c08105c81325a2f0d5" gracePeriod=2 Mar 13 10:42:47 crc kubenswrapper[4632]: I0313 10:42:47.896907 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.059728 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlqlv\" (UniqueName: \"kubernetes.io/projected/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-kube-api-access-vlqlv\") pod \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\" (UID: \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\") " Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.059833 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-catalog-content\") pod \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\" (UID: \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\") " Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.059924 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-utilities\") pod \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\" (UID: \"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d\") " Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.061606 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-utilities" (OuterVolumeSpecName: "utilities") pod "cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" (UID: "cf7c7e9e-1b37-4be4-916f-a5a90a6db26d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.084732 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-kube-api-access-vlqlv" (OuterVolumeSpecName: "kube-api-access-vlqlv") pod "cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" (UID: "cf7c7e9e-1b37-4be4-916f-a5a90a6db26d"). InnerVolumeSpecName "kube-api-access-vlqlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.133396 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" (UID: "cf7c7e9e-1b37-4be4-916f-a5a90a6db26d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.162163 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlqlv\" (UniqueName: \"kubernetes.io/projected/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-kube-api-access-vlqlv\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.162471 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.162483 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.447289 4632 generic.go:334] "Generic (PLEG): container finished" podID="cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" containerID="81c2fc126faa38cd627c0814225a882d113c0439776ae9c08105c81325a2f0d5" exitCode=0 Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.447345 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnw2p" event={"ID":"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d","Type":"ContainerDied","Data":"81c2fc126faa38cd627c0814225a882d113c0439776ae9c08105c81325a2f0d5"} Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.447381 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gnw2p" event={"ID":"cf7c7e9e-1b37-4be4-916f-a5a90a6db26d","Type":"ContainerDied","Data":"5d8fa1de309b7087e5b0807d70e759ae4655eeeff831ca9798400cbf8bccd783"} Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.447404 4632 scope.go:117] "RemoveContainer" containerID="81c2fc126faa38cd627c0814225a882d113c0439776ae9c08105c81325a2f0d5" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.447431 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gnw2p" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.482061 4632 scope.go:117] "RemoveContainer" containerID="8702aa0a52c8778ee09c744c907f2b88b34171b80b2c62359a090363a7ecc8ac" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.490749 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gnw2p"] Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.499134 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gnw2p"] Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.519105 4632 scope.go:117] "RemoveContainer" containerID="3dd84647257d4d428f925670105498da793e6d9758ec9cb89c02c2e2a54e558f" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.562214 4632 scope.go:117] "RemoveContainer" containerID="81c2fc126faa38cd627c0814225a882d113c0439776ae9c08105c81325a2f0d5" Mar 13 10:42:48 crc kubenswrapper[4632]: E0313 10:42:48.562642 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81c2fc126faa38cd627c0814225a882d113c0439776ae9c08105c81325a2f0d5\": container with ID starting with 81c2fc126faa38cd627c0814225a882d113c0439776ae9c08105c81325a2f0d5 not found: ID does not exist" containerID="81c2fc126faa38cd627c0814225a882d113c0439776ae9c08105c81325a2f0d5" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.562674 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81c2fc126faa38cd627c0814225a882d113c0439776ae9c08105c81325a2f0d5"} err="failed to get container status \"81c2fc126faa38cd627c0814225a882d113c0439776ae9c08105c81325a2f0d5\": rpc error: code = NotFound desc = could not find container \"81c2fc126faa38cd627c0814225a882d113c0439776ae9c08105c81325a2f0d5\": container with ID starting with 81c2fc126faa38cd627c0814225a882d113c0439776ae9c08105c81325a2f0d5 not found: ID does not exist" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.562696 4632 scope.go:117] "RemoveContainer" containerID="8702aa0a52c8778ee09c744c907f2b88b34171b80b2c62359a090363a7ecc8ac" Mar 13 10:42:48 crc kubenswrapper[4632]: E0313 10:42:48.563105 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8702aa0a52c8778ee09c744c907f2b88b34171b80b2c62359a090363a7ecc8ac\": container with ID starting with 8702aa0a52c8778ee09c744c907f2b88b34171b80b2c62359a090363a7ecc8ac not found: ID does not exist" containerID="8702aa0a52c8778ee09c744c907f2b88b34171b80b2c62359a090363a7ecc8ac" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.563131 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8702aa0a52c8778ee09c744c907f2b88b34171b80b2c62359a090363a7ecc8ac"} err="failed to get container status \"8702aa0a52c8778ee09c744c907f2b88b34171b80b2c62359a090363a7ecc8ac\": rpc error: code = NotFound desc = could not find container \"8702aa0a52c8778ee09c744c907f2b88b34171b80b2c62359a090363a7ecc8ac\": container with ID starting with 8702aa0a52c8778ee09c744c907f2b88b34171b80b2c62359a090363a7ecc8ac not found: ID does not exist" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.563145 4632 scope.go:117] "RemoveContainer" containerID="3dd84647257d4d428f925670105498da793e6d9758ec9cb89c02c2e2a54e558f" Mar 13 10:42:48 crc kubenswrapper[4632]: E0313 10:42:48.563417 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dd84647257d4d428f925670105498da793e6d9758ec9cb89c02c2e2a54e558f\": container with ID starting with 3dd84647257d4d428f925670105498da793e6d9758ec9cb89c02c2e2a54e558f not found: ID does not exist" containerID="3dd84647257d4d428f925670105498da793e6d9758ec9cb89c02c2e2a54e558f" Mar 13 10:42:48 crc kubenswrapper[4632]: I0313 10:42:48.563439 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dd84647257d4d428f925670105498da793e6d9758ec9cb89c02c2e2a54e558f"} err="failed to get container status \"3dd84647257d4d428f925670105498da793e6d9758ec9cb89c02c2e2a54e558f\": rpc error: code = NotFound desc = could not find container \"3dd84647257d4d428f925670105498da793e6d9758ec9cb89c02c2e2a54e558f\": container with ID starting with 3dd84647257d4d428f925670105498da793e6d9758ec9cb89c02c2e2a54e558f not found: ID does not exist" Mar 13 10:42:50 crc kubenswrapper[4632]: I0313 10:42:50.054655 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" path="/var/lib/kubelet/pods/cf7c7e9e-1b37-4be4-916f-a5a90a6db26d/volumes" Mar 13 10:42:54 crc kubenswrapper[4632]: I0313 10:42:54.043897 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:42:54 crc kubenswrapper[4632]: E0313 10:42:54.044773 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:43:08 crc kubenswrapper[4632]: I0313 10:43:08.059518 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:43:08 crc kubenswrapper[4632]: E0313 10:43:08.060405 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:43:19 crc kubenswrapper[4632]: I0313 10:43:19.043754 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:43:19 crc kubenswrapper[4632]: E0313 10:43:19.044642 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:43:19 crc kubenswrapper[4632]: I0313 10:43:19.720817 4632 generic.go:334] "Generic (PLEG): container finished" podID="96ca1247-6625-4b08-b155-34c56f02ec04" containerID="b9c8440031ae4a9073037d4c844efc1d48630f961db4b97500123a9880df8925" exitCode=0 Mar 13 10:43:19 crc kubenswrapper[4632]: I0313 10:43:19.720866 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" event={"ID":"96ca1247-6625-4b08-b155-34c56f02ec04","Type":"ContainerDied","Data":"b9c8440031ae4a9073037d4c844efc1d48630f961db4b97500123a9880df8925"} Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.293900 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.401577 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdp7x\" (UniqueName: \"kubernetes.io/projected/96ca1247-6625-4b08-b155-34c56f02ec04-kube-api-access-wdp7x\") pod \"96ca1247-6625-4b08-b155-34c56f02ec04\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.402186 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/96ca1247-6625-4b08-b155-34c56f02ec04-ovncontroller-config-0\") pod \"96ca1247-6625-4b08-b155-34c56f02ec04\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.402222 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-ovn-combined-ca-bundle\") pod \"96ca1247-6625-4b08-b155-34c56f02ec04\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.402476 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-inventory\") pod \"96ca1247-6625-4b08-b155-34c56f02ec04\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.402512 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-ssh-key-openstack-edpm-ipam\") pod \"96ca1247-6625-4b08-b155-34c56f02ec04\" (UID: \"96ca1247-6625-4b08-b155-34c56f02ec04\") " Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.410272 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "96ca1247-6625-4b08-b155-34c56f02ec04" (UID: "96ca1247-6625-4b08-b155-34c56f02ec04"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.410369 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96ca1247-6625-4b08-b155-34c56f02ec04-kube-api-access-wdp7x" (OuterVolumeSpecName: "kube-api-access-wdp7x") pod "96ca1247-6625-4b08-b155-34c56f02ec04" (UID: "96ca1247-6625-4b08-b155-34c56f02ec04"). InnerVolumeSpecName "kube-api-access-wdp7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.434223 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-inventory" (OuterVolumeSpecName: "inventory") pod "96ca1247-6625-4b08-b155-34c56f02ec04" (UID: "96ca1247-6625-4b08-b155-34c56f02ec04"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.436503 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "96ca1247-6625-4b08-b155-34c56f02ec04" (UID: "96ca1247-6625-4b08-b155-34c56f02ec04"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.441232 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/96ca1247-6625-4b08-b155-34c56f02ec04-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "96ca1247-6625-4b08-b155-34c56f02ec04" (UID: "96ca1247-6625-4b08-b155-34c56f02ec04"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.505026 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.505060 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.505074 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdp7x\" (UniqueName: \"kubernetes.io/projected/96ca1247-6625-4b08-b155-34c56f02ec04-kube-api-access-wdp7x\") on node \"crc\" DevicePath \"\"" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.505083 4632 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/96ca1247-6625-4b08-b155-34c56f02ec04-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.505093 4632 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96ca1247-6625-4b08-b155-34c56f02ec04-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.744093 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" event={"ID":"96ca1247-6625-4b08-b155-34c56f02ec04","Type":"ContainerDied","Data":"7cff807ce30996fc61ff5a53f3c16703d17afe3357255159f2daf0f8d5320802"} Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.744368 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cff807ce30996fc61ff5a53f3c16703d17afe3357255159f2daf0f8d5320802" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.744238 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-6t9b6" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.831700 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq"] Mar 13 10:43:21 crc kubenswrapper[4632]: E0313 10:43:21.832121 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" containerName="extract-utilities" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.832134 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" containerName="extract-utilities" Mar 13 10:43:21 crc kubenswrapper[4632]: E0313 10:43:21.832147 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" containerName="extract-content" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.832155 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" containerName="extract-content" Mar 13 10:43:21 crc kubenswrapper[4632]: E0313 10:43:21.832163 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96ca1247-6625-4b08-b155-34c56f02ec04" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.832170 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="96ca1247-6625-4b08-b155-34c56f02ec04" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 13 10:43:21 crc kubenswrapper[4632]: E0313 10:43:21.832192 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" containerName="registry-server" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.832198 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" containerName="registry-server" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.832385 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf7c7e9e-1b37-4be4-916f-a5a90a6db26d" containerName="registry-server" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.832400 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="96ca1247-6625-4b08-b155-34c56f02ec04" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.833001 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.838522 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.838603 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.838659 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.838712 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.839428 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.839518 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.856060 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq"] Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.912550 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktxnr\" (UniqueName: \"kubernetes.io/projected/96e4ce1c-8f09-4563-864f-da1f95bdd500-kube-api-access-ktxnr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.912608 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.912725 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.912815 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.912860 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:21 crc kubenswrapper[4632]: I0313 10:43:21.912962 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.015237 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktxnr\" (UniqueName: \"kubernetes.io/projected/96e4ce1c-8f09-4563-864f-da1f95bdd500-kube-api-access-ktxnr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.015298 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.015362 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.015419 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.015458 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.015526 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.020696 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.020916 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.021403 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.028868 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.035074 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.039516 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktxnr\" (UniqueName: \"kubernetes.io/projected/96e4ce1c-8f09-4563-864f-da1f95bdd500-kube-api-access-ktxnr\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.153671 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.716104 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq"] Mar 13 10:43:22 crc kubenswrapper[4632]: I0313 10:43:22.753841 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" event={"ID":"96e4ce1c-8f09-4563-864f-da1f95bdd500","Type":"ContainerStarted","Data":"63699ebe9acf1d6d33dc9f40302fd0b79959f65b39fbb9fa8c035bf1300ba29c"} Mar 13 10:43:23 crc kubenswrapper[4632]: I0313 10:43:23.772225 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" event={"ID":"96e4ce1c-8f09-4563-864f-da1f95bdd500","Type":"ContainerStarted","Data":"85251a26d676b79e5e052771251527a54fb8c42ed8c25a05daa652dca5b3df9e"} Mar 13 10:43:23 crc kubenswrapper[4632]: I0313 10:43:23.797549 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" podStartSLOduration=2.360757036 podStartE2EDuration="2.797525478s" podCreationTimestamp="2026-03-13 10:43:21 +0000 UTC" firstStartedPulling="2026-03-13 10:43:22.728535836 +0000 UTC m=+2376.751065969" lastFinishedPulling="2026-03-13 10:43:23.165304278 +0000 UTC m=+2377.187834411" observedRunningTime="2026-03-13 10:43:23.786457066 +0000 UTC m=+2377.808987229" watchObservedRunningTime="2026-03-13 10:43:23.797525478 +0000 UTC m=+2377.820055631" Mar 13 10:43:34 crc kubenswrapper[4632]: I0313 10:43:34.044498 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:43:34 crc kubenswrapper[4632]: E0313 10:43:34.045308 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:43:45 crc kubenswrapper[4632]: I0313 10:43:45.044466 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:43:45 crc kubenswrapper[4632]: E0313 10:43:45.045355 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:44:00 crc kubenswrapper[4632]: I0313 10:44:00.044433 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:44:00 crc kubenswrapper[4632]: E0313 10:44:00.045187 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:44:00 crc kubenswrapper[4632]: I0313 10:44:00.147676 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556644-dq2jd"] Mar 13 10:44:00 crc kubenswrapper[4632]: I0313 10:44:00.149494 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556644-dq2jd" Mar 13 10:44:00 crc kubenswrapper[4632]: I0313 10:44:00.152850 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:44:00 crc kubenswrapper[4632]: I0313 10:44:00.153116 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:44:00 crc kubenswrapper[4632]: I0313 10:44:00.155572 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:44:00 crc kubenswrapper[4632]: I0313 10:44:00.163171 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556644-dq2jd"] Mar 13 10:44:00 crc kubenswrapper[4632]: I0313 10:44:00.263779 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvzv7\" (UniqueName: \"kubernetes.io/projected/2462765e-6333-4e22-b4d7-ee2b2c6aa538-kube-api-access-kvzv7\") pod \"auto-csr-approver-29556644-dq2jd\" (UID: \"2462765e-6333-4e22-b4d7-ee2b2c6aa538\") " pod="openshift-infra/auto-csr-approver-29556644-dq2jd" Mar 13 10:44:00 crc kubenswrapper[4632]: I0313 10:44:00.366369 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvzv7\" (UniqueName: \"kubernetes.io/projected/2462765e-6333-4e22-b4d7-ee2b2c6aa538-kube-api-access-kvzv7\") pod \"auto-csr-approver-29556644-dq2jd\" (UID: \"2462765e-6333-4e22-b4d7-ee2b2c6aa538\") " pod="openshift-infra/auto-csr-approver-29556644-dq2jd" Mar 13 10:44:00 crc kubenswrapper[4632]: I0313 10:44:00.389404 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvzv7\" (UniqueName: \"kubernetes.io/projected/2462765e-6333-4e22-b4d7-ee2b2c6aa538-kube-api-access-kvzv7\") pod \"auto-csr-approver-29556644-dq2jd\" (UID: \"2462765e-6333-4e22-b4d7-ee2b2c6aa538\") " pod="openshift-infra/auto-csr-approver-29556644-dq2jd" Mar 13 10:44:00 crc kubenswrapper[4632]: I0313 10:44:00.475304 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556644-dq2jd" Mar 13 10:44:01 crc kubenswrapper[4632]: I0313 10:44:01.063624 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556644-dq2jd"] Mar 13 10:44:01 crc kubenswrapper[4632]: I0313 10:44:01.106579 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556644-dq2jd" event={"ID":"2462765e-6333-4e22-b4d7-ee2b2c6aa538","Type":"ContainerStarted","Data":"f56f0f67e1577faee7fca58378e1afee27b69eb2b7b50ae2443bfca5b6158d20"} Mar 13 10:44:03 crc kubenswrapper[4632]: I0313 10:44:03.130459 4632 generic.go:334] "Generic (PLEG): container finished" podID="2462765e-6333-4e22-b4d7-ee2b2c6aa538" containerID="07b2fe4a97569c9089b7972685eb914fd04195d02c9e7b239121095e54e42352" exitCode=0 Mar 13 10:44:03 crc kubenswrapper[4632]: I0313 10:44:03.131866 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556644-dq2jd" event={"ID":"2462765e-6333-4e22-b4d7-ee2b2c6aa538","Type":"ContainerDied","Data":"07b2fe4a97569c9089b7972685eb914fd04195d02c9e7b239121095e54e42352"} Mar 13 10:44:04 crc kubenswrapper[4632]: I0313 10:44:04.500878 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556644-dq2jd" Mar 13 10:44:04 crc kubenswrapper[4632]: I0313 10:44:04.658796 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvzv7\" (UniqueName: \"kubernetes.io/projected/2462765e-6333-4e22-b4d7-ee2b2c6aa538-kube-api-access-kvzv7\") pod \"2462765e-6333-4e22-b4d7-ee2b2c6aa538\" (UID: \"2462765e-6333-4e22-b4d7-ee2b2c6aa538\") " Mar 13 10:44:04 crc kubenswrapper[4632]: I0313 10:44:04.664414 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2462765e-6333-4e22-b4d7-ee2b2c6aa538-kube-api-access-kvzv7" (OuterVolumeSpecName: "kube-api-access-kvzv7") pod "2462765e-6333-4e22-b4d7-ee2b2c6aa538" (UID: "2462765e-6333-4e22-b4d7-ee2b2c6aa538"). InnerVolumeSpecName "kube-api-access-kvzv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:44:04 crc kubenswrapper[4632]: I0313 10:44:04.760984 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvzv7\" (UniqueName: \"kubernetes.io/projected/2462765e-6333-4e22-b4d7-ee2b2c6aa538-kube-api-access-kvzv7\") on node \"crc\" DevicePath \"\"" Mar 13 10:44:05 crc kubenswrapper[4632]: I0313 10:44:05.151972 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556644-dq2jd" event={"ID":"2462765e-6333-4e22-b4d7-ee2b2c6aa538","Type":"ContainerDied","Data":"f56f0f67e1577faee7fca58378e1afee27b69eb2b7b50ae2443bfca5b6158d20"} Mar 13 10:44:05 crc kubenswrapper[4632]: I0313 10:44:05.153415 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f56f0f67e1577faee7fca58378e1afee27b69eb2b7b50ae2443bfca5b6158d20" Mar 13 10:44:05 crc kubenswrapper[4632]: I0313 10:44:05.152415 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556644-dq2jd" Mar 13 10:44:05 crc kubenswrapper[4632]: I0313 10:44:05.577628 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556638-p7mdh"] Mar 13 10:44:05 crc kubenswrapper[4632]: I0313 10:44:05.587726 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556638-p7mdh"] Mar 13 10:44:06 crc kubenswrapper[4632]: I0313 10:44:06.056536 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="346e767a-d9dd-40e1-9ab3-2e4ec9184667" path="/var/lib/kubelet/pods/346e767a-d9dd-40e1-9ab3-2e4ec9184667/volumes" Mar 13 10:44:09 crc kubenswrapper[4632]: I0313 10:44:09.504441 4632 scope.go:117] "RemoveContainer" containerID="c166f0a830c16b65f03aba2171bb98a995fe4121f1b92036d629fce2afd52c26" Mar 13 10:44:12 crc kubenswrapper[4632]: I0313 10:44:12.044627 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:44:12 crc kubenswrapper[4632]: E0313 10:44:12.046356 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:44:14 crc kubenswrapper[4632]: I0313 10:44:14.229233 4632 generic.go:334] "Generic (PLEG): container finished" podID="96e4ce1c-8f09-4563-864f-da1f95bdd500" containerID="85251a26d676b79e5e052771251527a54fb8c42ed8c25a05daa652dca5b3df9e" exitCode=0 Mar 13 10:44:14 crc kubenswrapper[4632]: I0313 10:44:14.229327 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" event={"ID":"96e4ce1c-8f09-4563-864f-da1f95bdd500","Type":"ContainerDied","Data":"85251a26d676b79e5e052771251527a54fb8c42ed8c25a05daa652dca5b3df9e"} Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.692152 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.788903 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktxnr\" (UniqueName: \"kubernetes.io/projected/96e4ce1c-8f09-4563-864f-da1f95bdd500-kube-api-access-ktxnr\") pod \"96e4ce1c-8f09-4563-864f-da1f95bdd500\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.790101 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-ssh-key-openstack-edpm-ipam\") pod \"96e4ce1c-8f09-4563-864f-da1f95bdd500\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.790166 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-inventory\") pod \"96e4ce1c-8f09-4563-864f-da1f95bdd500\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.790212 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-nova-metadata-neutron-config-0\") pod \"96e4ce1c-8f09-4563-864f-da1f95bdd500\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.790258 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-neutron-ovn-metadata-agent-neutron-config-0\") pod \"96e4ce1c-8f09-4563-864f-da1f95bdd500\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.790416 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-neutron-metadata-combined-ca-bundle\") pod \"96e4ce1c-8f09-4563-864f-da1f95bdd500\" (UID: \"96e4ce1c-8f09-4563-864f-da1f95bdd500\") " Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.796393 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96e4ce1c-8f09-4563-864f-da1f95bdd500-kube-api-access-ktxnr" (OuterVolumeSpecName: "kube-api-access-ktxnr") pod "96e4ce1c-8f09-4563-864f-da1f95bdd500" (UID: "96e4ce1c-8f09-4563-864f-da1f95bdd500"). InnerVolumeSpecName "kube-api-access-ktxnr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.798607 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "96e4ce1c-8f09-4563-864f-da1f95bdd500" (UID: "96e4ce1c-8f09-4563-864f-da1f95bdd500"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.818081 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "96e4ce1c-8f09-4563-864f-da1f95bdd500" (UID: "96e4ce1c-8f09-4563-864f-da1f95bdd500"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.822407 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "96e4ce1c-8f09-4563-864f-da1f95bdd500" (UID: "96e4ce1c-8f09-4563-864f-da1f95bdd500"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.825475 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-inventory" (OuterVolumeSpecName: "inventory") pod "96e4ce1c-8f09-4563-864f-da1f95bdd500" (UID: "96e4ce1c-8f09-4563-864f-da1f95bdd500"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.828899 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "96e4ce1c-8f09-4563-864f-da1f95bdd500" (UID: "96e4ce1c-8f09-4563-864f-da1f95bdd500"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.893666 4632 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.893715 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktxnr\" (UniqueName: \"kubernetes.io/projected/96e4ce1c-8f09-4563-864f-da1f95bdd500-kube-api-access-ktxnr\") on node \"crc\" DevicePath \"\"" Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.893727 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.893738 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.893749 4632 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:44:15 crc kubenswrapper[4632]: I0313 10:44:15.893768 4632 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/96e4ce1c-8f09-4563-864f-da1f95bdd500-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.247472 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" event={"ID":"96e4ce1c-8f09-4563-864f-da1f95bdd500","Type":"ContainerDied","Data":"63699ebe9acf1d6d33dc9f40302fd0b79959f65b39fbb9fa8c035bf1300ba29c"} Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.247713 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63699ebe9acf1d6d33dc9f40302fd0b79959f65b39fbb9fa8c035bf1300ba29c" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.247518 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.395031 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh"] Mar 13 10:44:16 crc kubenswrapper[4632]: E0313 10:44:16.395517 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96e4ce1c-8f09-4563-864f-da1f95bdd500" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.395543 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="96e4ce1c-8f09-4563-864f-da1f95bdd500" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 13 10:44:16 crc kubenswrapper[4632]: E0313 10:44:16.395589 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2462765e-6333-4e22-b4d7-ee2b2c6aa538" containerName="oc" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.395601 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2462765e-6333-4e22-b4d7-ee2b2c6aa538" containerName="oc" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.395829 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="2462765e-6333-4e22-b4d7-ee2b2c6aa538" containerName="oc" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.395867 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="96e4ce1c-8f09-4563-864f-da1f95bdd500" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.396689 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.399062 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.399828 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.399996 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.400215 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.401274 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.405586 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh"] Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.507414 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.507471 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh7rq\" (UniqueName: \"kubernetes.io/projected/ed1a2c50-a476-43ca-9764-e0ebffb14134-kube-api-access-xh7rq\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.507522 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.507585 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.507694 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.610025 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.610180 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.610324 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.610354 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh7rq\" (UniqueName: \"kubernetes.io/projected/ed1a2c50-a476-43ca-9764-e0ebffb14134-kube-api-access-xh7rq\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.610379 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.613811 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.614322 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.614780 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.624076 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.630708 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh7rq\" (UniqueName: \"kubernetes.io/projected/ed1a2c50-a476-43ca-9764-e0ebffb14134-kube-api-access-xh7rq\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-skjrh\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:16 crc kubenswrapper[4632]: I0313 10:44:16.712546 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:44:17 crc kubenswrapper[4632]: I0313 10:44:17.273522 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh"] Mar 13 10:44:18 crc kubenswrapper[4632]: I0313 10:44:18.276000 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" event={"ID":"ed1a2c50-a476-43ca-9764-e0ebffb14134","Type":"ContainerStarted","Data":"e94f04b40e30694707c6fd5089936e853269c95465d0464597019d512ad17ad4"} Mar 13 10:44:18 crc kubenswrapper[4632]: I0313 10:44:18.276314 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" event={"ID":"ed1a2c50-a476-43ca-9764-e0ebffb14134","Type":"ContainerStarted","Data":"251492aaae45a40b4ef377f82a35e7e430a26d0b96081ac467780ad353a89a5b"} Mar 13 10:44:18 crc kubenswrapper[4632]: I0313 10:44:18.312046 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" podStartSLOduration=1.68438618 podStartE2EDuration="2.312021216s" podCreationTimestamp="2026-03-13 10:44:16 +0000 UTC" firstStartedPulling="2026-03-13 10:44:17.27578921 +0000 UTC m=+2431.298319343" lastFinishedPulling="2026-03-13 10:44:17.903424256 +0000 UTC m=+2431.925954379" observedRunningTime="2026-03-13 10:44:18.303675811 +0000 UTC m=+2432.326205944" watchObservedRunningTime="2026-03-13 10:44:18.312021216 +0000 UTC m=+2432.334551349" Mar 13 10:44:27 crc kubenswrapper[4632]: I0313 10:44:27.044518 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:44:27 crc kubenswrapper[4632]: E0313 10:44:27.046143 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:44:41 crc kubenswrapper[4632]: I0313 10:44:41.046274 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:44:41 crc kubenswrapper[4632]: E0313 10:44:41.047114 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:44:56 crc kubenswrapper[4632]: I0313 10:44:56.045448 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:44:56 crc kubenswrapper[4632]: E0313 10:44:56.046205 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.156129 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8"] Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.158107 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.160509 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.162511 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.221610 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8"] Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.295920 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cf52265-21b3-40f0-a2f5-d379c03cc045-config-volume\") pod \"collect-profiles-29556645-4btb8\" (UID: \"8cf52265-21b3-40f0-a2f5-d379c03cc045\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.296088 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8cf52265-21b3-40f0-a2f5-d379c03cc045-secret-volume\") pod \"collect-profiles-29556645-4btb8\" (UID: \"8cf52265-21b3-40f0-a2f5-d379c03cc045\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.296259 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8dpp\" (UniqueName: \"kubernetes.io/projected/8cf52265-21b3-40f0-a2f5-d379c03cc045-kube-api-access-h8dpp\") pod \"collect-profiles-29556645-4btb8\" (UID: \"8cf52265-21b3-40f0-a2f5-d379c03cc045\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.398071 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8cf52265-21b3-40f0-a2f5-d379c03cc045-secret-volume\") pod \"collect-profiles-29556645-4btb8\" (UID: \"8cf52265-21b3-40f0-a2f5-d379c03cc045\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.398580 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8dpp\" (UniqueName: \"kubernetes.io/projected/8cf52265-21b3-40f0-a2f5-d379c03cc045-kube-api-access-h8dpp\") pod \"collect-profiles-29556645-4btb8\" (UID: \"8cf52265-21b3-40f0-a2f5-d379c03cc045\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.398648 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cf52265-21b3-40f0-a2f5-d379c03cc045-config-volume\") pod \"collect-profiles-29556645-4btb8\" (UID: \"8cf52265-21b3-40f0-a2f5-d379c03cc045\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.399555 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cf52265-21b3-40f0-a2f5-d379c03cc045-config-volume\") pod \"collect-profiles-29556645-4btb8\" (UID: \"8cf52265-21b3-40f0-a2f5-d379c03cc045\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.414785 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8cf52265-21b3-40f0-a2f5-d379c03cc045-secret-volume\") pod \"collect-profiles-29556645-4btb8\" (UID: \"8cf52265-21b3-40f0-a2f5-d379c03cc045\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.421251 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8dpp\" (UniqueName: \"kubernetes.io/projected/8cf52265-21b3-40f0-a2f5-d379c03cc045-kube-api-access-h8dpp\") pod \"collect-profiles-29556645-4btb8\" (UID: \"8cf52265-21b3-40f0-a2f5-d379c03cc045\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" Mar 13 10:45:00 crc kubenswrapper[4632]: I0313 10:45:00.512683 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" Mar 13 10:45:01 crc kubenswrapper[4632]: I0313 10:45:01.057691 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8"] Mar 13 10:45:01 crc kubenswrapper[4632]: I0313 10:45:01.702465 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" event={"ID":"8cf52265-21b3-40f0-a2f5-d379c03cc045","Type":"ContainerStarted","Data":"8e20db958c001216e89a657171c617c2e4d78b297bcd654a9af9c2d8d32242ac"} Mar 13 10:45:01 crc kubenswrapper[4632]: I0313 10:45:01.704111 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" event={"ID":"8cf52265-21b3-40f0-a2f5-d379c03cc045","Type":"ContainerStarted","Data":"8d57aa4aed604811ff55c6ede5103947b8bd4fba767956c296f00e30b4b5ac65"} Mar 13 10:45:01 crc kubenswrapper[4632]: I0313 10:45:01.728271 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" podStartSLOduration=1.728248449 podStartE2EDuration="1.728248449s" podCreationTimestamp="2026-03-13 10:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:45:01.723911482 +0000 UTC m=+2475.746441615" watchObservedRunningTime="2026-03-13 10:45:01.728248449 +0000 UTC m=+2475.750778582" Mar 13 10:45:02 crc kubenswrapper[4632]: I0313 10:45:02.712828 4632 generic.go:334] "Generic (PLEG): container finished" podID="8cf52265-21b3-40f0-a2f5-d379c03cc045" containerID="8e20db958c001216e89a657171c617c2e4d78b297bcd654a9af9c2d8d32242ac" exitCode=0 Mar 13 10:45:02 crc kubenswrapper[4632]: I0313 10:45:02.712884 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" event={"ID":"8cf52265-21b3-40f0-a2f5-d379c03cc045","Type":"ContainerDied","Data":"8e20db958c001216e89a657171c617c2e4d78b297bcd654a9af9c2d8d32242ac"} Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.072013 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.190856 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cf52265-21b3-40f0-a2f5-d379c03cc045-config-volume\") pod \"8cf52265-21b3-40f0-a2f5-d379c03cc045\" (UID: \"8cf52265-21b3-40f0-a2f5-d379c03cc045\") " Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.190920 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8cf52265-21b3-40f0-a2f5-d379c03cc045-secret-volume\") pod \"8cf52265-21b3-40f0-a2f5-d379c03cc045\" (UID: \"8cf52265-21b3-40f0-a2f5-d379c03cc045\") " Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.191107 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8dpp\" (UniqueName: \"kubernetes.io/projected/8cf52265-21b3-40f0-a2f5-d379c03cc045-kube-api-access-h8dpp\") pod \"8cf52265-21b3-40f0-a2f5-d379c03cc045\" (UID: \"8cf52265-21b3-40f0-a2f5-d379c03cc045\") " Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.191569 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cf52265-21b3-40f0-a2f5-d379c03cc045-config-volume" (OuterVolumeSpecName: "config-volume") pod "8cf52265-21b3-40f0-a2f5-d379c03cc045" (UID: "8cf52265-21b3-40f0-a2f5-d379c03cc045"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.199127 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cf52265-21b3-40f0-a2f5-d379c03cc045-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8cf52265-21b3-40f0-a2f5-d379c03cc045" (UID: "8cf52265-21b3-40f0-a2f5-d379c03cc045"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.199211 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cf52265-21b3-40f0-a2f5-d379c03cc045-kube-api-access-h8dpp" (OuterVolumeSpecName: "kube-api-access-h8dpp") pod "8cf52265-21b3-40f0-a2f5-d379c03cc045" (UID: "8cf52265-21b3-40f0-a2f5-d379c03cc045"). InnerVolumeSpecName "kube-api-access-h8dpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.293712 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8dpp\" (UniqueName: \"kubernetes.io/projected/8cf52265-21b3-40f0-a2f5-d379c03cc045-kube-api-access-h8dpp\") on node \"crc\" DevicePath \"\"" Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.293758 4632 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8cf52265-21b3-40f0-a2f5-d379c03cc045-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.293768 4632 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8cf52265-21b3-40f0-a2f5-d379c03cc045-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.731685 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" event={"ID":"8cf52265-21b3-40f0-a2f5-d379c03cc045","Type":"ContainerDied","Data":"8d57aa4aed604811ff55c6ede5103947b8bd4fba767956c296f00e30b4b5ac65"} Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.731964 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d57aa4aed604811ff55c6ede5103947b8bd4fba767956c296f00e30b4b5ac65" Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.731793 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8" Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.807620 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg"] Mar 13 10:45:04 crc kubenswrapper[4632]: I0313 10:45:04.817549 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556600-r9flg"] Mar 13 10:45:06 crc kubenswrapper[4632]: I0313 10:45:06.058720 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="528d3aa9-10bf-4029-a4d2-85768264fde8" path="/var/lib/kubelet/pods/528d3aa9-10bf-4029-a4d2-85768264fde8/volumes" Mar 13 10:45:09 crc kubenswrapper[4632]: I0313 10:45:09.588802 4632 scope.go:117] "RemoveContainer" containerID="da165dd4ae62fa2ea1c777c8125fcd4bfe4bd102f508da056f1a058689bba35e" Mar 13 10:45:11 crc kubenswrapper[4632]: I0313 10:45:11.044987 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:45:11 crc kubenswrapper[4632]: E0313 10:45:11.045449 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:45:26 crc kubenswrapper[4632]: I0313 10:45:26.044287 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:45:26 crc kubenswrapper[4632]: E0313 10:45:26.045223 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:45:38 crc kubenswrapper[4632]: I0313 10:45:38.053586 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:45:38 crc kubenswrapper[4632]: E0313 10:45:38.054565 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:45:51 crc kubenswrapper[4632]: I0313 10:45:51.044349 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:45:51 crc kubenswrapper[4632]: E0313 10:45:51.045689 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:46:00 crc kubenswrapper[4632]: I0313 10:46:00.143904 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556646-blkxp"] Mar 13 10:46:00 crc kubenswrapper[4632]: E0313 10:46:00.144922 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cf52265-21b3-40f0-a2f5-d379c03cc045" containerName="collect-profiles" Mar 13 10:46:00 crc kubenswrapper[4632]: I0313 10:46:00.144939 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cf52265-21b3-40f0-a2f5-d379c03cc045" containerName="collect-profiles" Mar 13 10:46:00 crc kubenswrapper[4632]: I0313 10:46:00.145170 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cf52265-21b3-40f0-a2f5-d379c03cc045" containerName="collect-profiles" Mar 13 10:46:00 crc kubenswrapper[4632]: I0313 10:46:00.145872 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556646-blkxp" Mar 13 10:46:00 crc kubenswrapper[4632]: I0313 10:46:00.148151 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:46:00 crc kubenswrapper[4632]: I0313 10:46:00.148265 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:46:00 crc kubenswrapper[4632]: I0313 10:46:00.148265 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:46:00 crc kubenswrapper[4632]: I0313 10:46:00.164437 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556646-blkxp"] Mar 13 10:46:00 crc kubenswrapper[4632]: I0313 10:46:00.273900 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnzgm\" (UniqueName: \"kubernetes.io/projected/93948d53-dbf3-47ce-8af0-bee10cc7e246-kube-api-access-fnzgm\") pod \"auto-csr-approver-29556646-blkxp\" (UID: \"93948d53-dbf3-47ce-8af0-bee10cc7e246\") " pod="openshift-infra/auto-csr-approver-29556646-blkxp" Mar 13 10:46:00 crc kubenswrapper[4632]: I0313 10:46:00.376224 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnzgm\" (UniqueName: \"kubernetes.io/projected/93948d53-dbf3-47ce-8af0-bee10cc7e246-kube-api-access-fnzgm\") pod \"auto-csr-approver-29556646-blkxp\" (UID: \"93948d53-dbf3-47ce-8af0-bee10cc7e246\") " pod="openshift-infra/auto-csr-approver-29556646-blkxp" Mar 13 10:46:00 crc kubenswrapper[4632]: I0313 10:46:00.400791 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnzgm\" (UniqueName: \"kubernetes.io/projected/93948d53-dbf3-47ce-8af0-bee10cc7e246-kube-api-access-fnzgm\") pod \"auto-csr-approver-29556646-blkxp\" (UID: \"93948d53-dbf3-47ce-8af0-bee10cc7e246\") " pod="openshift-infra/auto-csr-approver-29556646-blkxp" Mar 13 10:46:00 crc kubenswrapper[4632]: I0313 10:46:00.471216 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556646-blkxp" Mar 13 10:46:01 crc kubenswrapper[4632]: I0313 10:46:01.002410 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556646-blkxp"] Mar 13 10:46:01 crc kubenswrapper[4632]: I0313 10:46:01.020300 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:46:01 crc kubenswrapper[4632]: I0313 10:46:01.285109 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556646-blkxp" event={"ID":"93948d53-dbf3-47ce-8af0-bee10cc7e246","Type":"ContainerStarted","Data":"1e499c382014b0aca0d9f283dcf3e408db5620f77e2f5a1e95fbd235bc60e907"} Mar 13 10:46:02 crc kubenswrapper[4632]: I0313 10:46:02.045181 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:46:02 crc kubenswrapper[4632]: E0313 10:46:02.045679 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:46:03 crc kubenswrapper[4632]: I0313 10:46:03.316718 4632 generic.go:334] "Generic (PLEG): container finished" podID="93948d53-dbf3-47ce-8af0-bee10cc7e246" containerID="92f6939c452dda4592aa326adcecce982f4fafb95f93ce909a101db10372c2ab" exitCode=0 Mar 13 10:46:03 crc kubenswrapper[4632]: I0313 10:46:03.317052 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556646-blkxp" event={"ID":"93948d53-dbf3-47ce-8af0-bee10cc7e246","Type":"ContainerDied","Data":"92f6939c452dda4592aa326adcecce982f4fafb95f93ce909a101db10372c2ab"} Mar 13 10:46:04 crc kubenswrapper[4632]: I0313 10:46:04.685559 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556646-blkxp" Mar 13 10:46:04 crc kubenswrapper[4632]: I0313 10:46:04.785404 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnzgm\" (UniqueName: \"kubernetes.io/projected/93948d53-dbf3-47ce-8af0-bee10cc7e246-kube-api-access-fnzgm\") pod \"93948d53-dbf3-47ce-8af0-bee10cc7e246\" (UID: \"93948d53-dbf3-47ce-8af0-bee10cc7e246\") " Mar 13 10:46:04 crc kubenswrapper[4632]: I0313 10:46:04.793271 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93948d53-dbf3-47ce-8af0-bee10cc7e246-kube-api-access-fnzgm" (OuterVolumeSpecName: "kube-api-access-fnzgm") pod "93948d53-dbf3-47ce-8af0-bee10cc7e246" (UID: "93948d53-dbf3-47ce-8af0-bee10cc7e246"). InnerVolumeSpecName "kube-api-access-fnzgm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:46:04 crc kubenswrapper[4632]: I0313 10:46:04.887908 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnzgm\" (UniqueName: \"kubernetes.io/projected/93948d53-dbf3-47ce-8af0-bee10cc7e246-kube-api-access-fnzgm\") on node \"crc\" DevicePath \"\"" Mar 13 10:46:05 crc kubenswrapper[4632]: I0313 10:46:05.334078 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556646-blkxp" event={"ID":"93948d53-dbf3-47ce-8af0-bee10cc7e246","Type":"ContainerDied","Data":"1e499c382014b0aca0d9f283dcf3e408db5620f77e2f5a1e95fbd235bc60e907"} Mar 13 10:46:05 crc kubenswrapper[4632]: I0313 10:46:05.334119 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e499c382014b0aca0d9f283dcf3e408db5620f77e2f5a1e95fbd235bc60e907" Mar 13 10:46:05 crc kubenswrapper[4632]: I0313 10:46:05.334150 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556646-blkxp" Mar 13 10:46:05 crc kubenswrapper[4632]: I0313 10:46:05.758874 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556640-zlhsq"] Mar 13 10:46:05 crc kubenswrapper[4632]: I0313 10:46:05.768504 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556640-zlhsq"] Mar 13 10:46:06 crc kubenswrapper[4632]: I0313 10:46:06.056334 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bcf0de2-27ca-4278-80a3-080ce237e6df" path="/var/lib/kubelet/pods/4bcf0de2-27ca-4278-80a3-080ce237e6df/volumes" Mar 13 10:46:09 crc kubenswrapper[4632]: I0313 10:46:09.673595 4632 scope.go:117] "RemoveContainer" containerID="84030d1b6c9dd12b070ed748955e52fe36ed2cac9f9bdddb744ca14dc6fbfa0a" Mar 13 10:46:16 crc kubenswrapper[4632]: I0313 10:46:16.045700 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:46:16 crc kubenswrapper[4632]: E0313 10:46:16.046671 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:46:31 crc kubenswrapper[4632]: I0313 10:46:31.044824 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:46:31 crc kubenswrapper[4632]: E0313 10:46:31.046418 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:46:42 crc kubenswrapper[4632]: I0313 10:46:42.044337 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:46:42 crc kubenswrapper[4632]: E0313 10:46:42.045322 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.340498 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8rs94"] Mar 13 10:46:55 crc kubenswrapper[4632]: E0313 10:46:55.342846 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93948d53-dbf3-47ce-8af0-bee10cc7e246" containerName="oc" Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.342971 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="93948d53-dbf3-47ce-8af0-bee10cc7e246" containerName="oc" Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.343276 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="93948d53-dbf3-47ce-8af0-bee10cc7e246" containerName="oc" Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.344611 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.404451 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8rs94"] Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.430310 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-utilities\") pod \"certified-operators-8rs94\" (UID: \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\") " pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.430424 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhljf\" (UniqueName: \"kubernetes.io/projected/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-kube-api-access-vhljf\") pod \"certified-operators-8rs94\" (UID: \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\") " pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.430544 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-catalog-content\") pod \"certified-operators-8rs94\" (UID: \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\") " pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.532651 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-utilities\") pod \"certified-operators-8rs94\" (UID: \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\") " pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.532752 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhljf\" (UniqueName: \"kubernetes.io/projected/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-kube-api-access-vhljf\") pod \"certified-operators-8rs94\" (UID: \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\") " pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.532823 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-catalog-content\") pod \"certified-operators-8rs94\" (UID: \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\") " pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.533434 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-utilities\") pod \"certified-operators-8rs94\" (UID: \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\") " pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.533363 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-catalog-content\") pod \"certified-operators-8rs94\" (UID: \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\") " pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.561192 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhljf\" (UniqueName: \"kubernetes.io/projected/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-kube-api-access-vhljf\") pod \"certified-operators-8rs94\" (UID: \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\") " pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:46:55 crc kubenswrapper[4632]: I0313 10:46:55.705747 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:46:56 crc kubenswrapper[4632]: I0313 10:46:56.047138 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:46:56 crc kubenswrapper[4632]: E0313 10:46:56.047351 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:46:56 crc kubenswrapper[4632]: I0313 10:46:56.298909 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8rs94"] Mar 13 10:46:56 crc kubenswrapper[4632]: I0313 10:46:56.962975 4632 generic.go:334] "Generic (PLEG): container finished" podID="4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" containerID="9a68434b5791583fc729ae45ace124d9bffa3419f6e081364198559e5e722d2d" exitCode=0 Mar 13 10:46:56 crc kubenswrapper[4632]: I0313 10:46:56.963163 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rs94" event={"ID":"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58","Type":"ContainerDied","Data":"9a68434b5791583fc729ae45ace124d9bffa3419f6e081364198559e5e722d2d"} Mar 13 10:46:56 crc kubenswrapper[4632]: I0313 10:46:56.963336 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rs94" event={"ID":"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58","Type":"ContainerStarted","Data":"2fcd4f33b9170a9e1f17871a87902d5c3a4bca4f5bff191585eaa7b78efa9537"} Mar 13 10:46:57 crc kubenswrapper[4632]: I0313 10:46:57.974734 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rs94" event={"ID":"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58","Type":"ContainerStarted","Data":"073ef57962037ed0b42aebc1b9b9c42dc3c21fd678c44de3f2fc1650fff2b40f"} Mar 13 10:46:59 crc kubenswrapper[4632]: I0313 10:46:59.993401 4632 generic.go:334] "Generic (PLEG): container finished" podID="4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" containerID="073ef57962037ed0b42aebc1b9b9c42dc3c21fd678c44de3f2fc1650fff2b40f" exitCode=0 Mar 13 10:46:59 crc kubenswrapper[4632]: I0313 10:46:59.993458 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rs94" event={"ID":"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58","Type":"ContainerDied","Data":"073ef57962037ed0b42aebc1b9b9c42dc3c21fd678c44de3f2fc1650fff2b40f"} Mar 13 10:47:01 crc kubenswrapper[4632]: I0313 10:47:01.005311 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rs94" event={"ID":"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58","Type":"ContainerStarted","Data":"debb15938677e3a625a25bb92735b8bbd7318e3df789b75b917f0f39fb974e78"} Mar 13 10:47:01 crc kubenswrapper[4632]: I0313 10:47:01.033069 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8rs94" podStartSLOduration=2.5485645 podStartE2EDuration="6.033051869s" podCreationTimestamp="2026-03-13 10:46:55 +0000 UTC" firstStartedPulling="2026-03-13 10:46:56.965795831 +0000 UTC m=+2590.988325964" lastFinishedPulling="2026-03-13 10:47:00.45028318 +0000 UTC m=+2594.472813333" observedRunningTime="2026-03-13 10:47:01.031188604 +0000 UTC m=+2595.053718737" watchObservedRunningTime="2026-03-13 10:47:01.033051869 +0000 UTC m=+2595.055582002" Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.595852 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-r7qtl"] Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.600554 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.636484 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r7qtl"] Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.648673 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-utilities\") pod \"redhat-marketplace-r7qtl\" (UID: \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\") " pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.648821 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5z5hv\" (UniqueName: \"kubernetes.io/projected/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-kube-api-access-5z5hv\") pod \"redhat-marketplace-r7qtl\" (UID: \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\") " pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.648908 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-catalog-content\") pod \"redhat-marketplace-r7qtl\" (UID: \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\") " pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.706816 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.707149 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.751569 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5z5hv\" (UniqueName: \"kubernetes.io/projected/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-kube-api-access-5z5hv\") pod \"redhat-marketplace-r7qtl\" (UID: \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\") " pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.751895 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-catalog-content\") pod \"redhat-marketplace-r7qtl\" (UID: \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\") " pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.752157 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-utilities\") pod \"redhat-marketplace-r7qtl\" (UID: \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\") " pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.752510 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-catalog-content\") pod \"redhat-marketplace-r7qtl\" (UID: \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\") " pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.752855 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-utilities\") pod \"redhat-marketplace-r7qtl\" (UID: \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\") " pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.778574 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5z5hv\" (UniqueName: \"kubernetes.io/projected/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-kube-api-access-5z5hv\") pod \"redhat-marketplace-r7qtl\" (UID: \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\") " pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:05 crc kubenswrapper[4632]: I0313 10:47:05.931040 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:06 crc kubenswrapper[4632]: I0313 10:47:06.539859 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-r7qtl"] Mar 13 10:47:06 crc kubenswrapper[4632]: I0313 10:47:06.765152 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-8rs94" podUID="4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" containerName="registry-server" probeResult="failure" output=< Mar 13 10:47:06 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:47:06 crc kubenswrapper[4632]: > Mar 13 10:47:07 crc kubenswrapper[4632]: I0313 10:47:07.081450 4632 generic.go:334] "Generic (PLEG): container finished" podID="e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" containerID="b3d79b73fcd58ac20d741acf9d96929a0df0db65f84dbb23ba98c1eff210414e" exitCode=0 Mar 13 10:47:07 crc kubenswrapper[4632]: I0313 10:47:07.081503 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7qtl" event={"ID":"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511","Type":"ContainerDied","Data":"b3d79b73fcd58ac20d741acf9d96929a0df0db65f84dbb23ba98c1eff210414e"} Mar 13 10:47:07 crc kubenswrapper[4632]: I0313 10:47:07.081569 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7qtl" event={"ID":"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511","Type":"ContainerStarted","Data":"874f90ecd39b4972bd013b4da6787ff75bf2a57a461e69b7db2804dbda134bb6"} Mar 13 10:47:08 crc kubenswrapper[4632]: I0313 10:47:08.093292 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7qtl" event={"ID":"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511","Type":"ContainerStarted","Data":"6869bdf68c35077e3e515fc4d4fc9c80e2292fab5161433ed27b5622d65b9902"} Mar 13 10:47:10 crc kubenswrapper[4632]: I0313 10:47:10.044624 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:47:10 crc kubenswrapper[4632]: E0313 10:47:10.045200 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:47:10 crc kubenswrapper[4632]: I0313 10:47:10.112488 4632 generic.go:334] "Generic (PLEG): container finished" podID="e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" containerID="6869bdf68c35077e3e515fc4d4fc9c80e2292fab5161433ed27b5622d65b9902" exitCode=0 Mar 13 10:47:10 crc kubenswrapper[4632]: I0313 10:47:10.112531 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7qtl" event={"ID":"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511","Type":"ContainerDied","Data":"6869bdf68c35077e3e515fc4d4fc9c80e2292fab5161433ed27b5622d65b9902"} Mar 13 10:47:11 crc kubenswrapper[4632]: I0313 10:47:11.124040 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7qtl" event={"ID":"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511","Type":"ContainerStarted","Data":"1182e2999759e9be37ef50a31fa5017ebfa813a7c00f6df460a1c1c6e3b78a8f"} Mar 13 10:47:15 crc kubenswrapper[4632]: I0313 10:47:15.763244 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:47:15 crc kubenswrapper[4632]: I0313 10:47:15.797959 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-r7qtl" podStartSLOduration=7.310138202 podStartE2EDuration="10.797914363s" podCreationTimestamp="2026-03-13 10:47:05 +0000 UTC" firstStartedPulling="2026-03-13 10:47:07.084582902 +0000 UTC m=+2601.107113045" lastFinishedPulling="2026-03-13 10:47:10.572359073 +0000 UTC m=+2604.594889206" observedRunningTime="2026-03-13 10:47:11.150374795 +0000 UTC m=+2605.172904948" watchObservedRunningTime="2026-03-13 10:47:15.797914363 +0000 UTC m=+2609.820444506" Mar 13 10:47:15 crc kubenswrapper[4632]: I0313 10:47:15.815091 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:47:15 crc kubenswrapper[4632]: I0313 10:47:15.931178 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:15 crc kubenswrapper[4632]: I0313 10:47:15.932283 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:15 crc kubenswrapper[4632]: I0313 10:47:15.976628 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:16 crc kubenswrapper[4632]: I0313 10:47:16.014003 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8rs94"] Mar 13 10:47:16 crc kubenswrapper[4632]: I0313 10:47:16.213657 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:17 crc kubenswrapper[4632]: I0313 10:47:17.170724 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8rs94" podUID="4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" containerName="registry-server" containerID="cri-o://debb15938677e3a625a25bb92735b8bbd7318e3df789b75b917f0f39fb974e78" gracePeriod=2 Mar 13 10:47:17 crc kubenswrapper[4632]: I0313 10:47:17.635040 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:47:17 crc kubenswrapper[4632]: I0313 10:47:17.694052 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-utilities\") pod \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\" (UID: \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\") " Mar 13 10:47:17 crc kubenswrapper[4632]: I0313 10:47:17.694294 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhljf\" (UniqueName: \"kubernetes.io/projected/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-kube-api-access-vhljf\") pod \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\" (UID: \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\") " Mar 13 10:47:17 crc kubenswrapper[4632]: I0313 10:47:17.694426 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-catalog-content\") pod \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\" (UID: \"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58\") " Mar 13 10:47:17 crc kubenswrapper[4632]: I0313 10:47:17.694912 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-utilities" (OuterVolumeSpecName: "utilities") pod "4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" (UID: "4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:47:17 crc kubenswrapper[4632]: I0313 10:47:17.701238 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-kube-api-access-vhljf" (OuterVolumeSpecName: "kube-api-access-vhljf") pod "4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" (UID: "4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58"). InnerVolumeSpecName "kube-api-access-vhljf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:47:17 crc kubenswrapper[4632]: I0313 10:47:17.749739 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" (UID: "4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:47:17 crc kubenswrapper[4632]: I0313 10:47:17.796253 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:47:17 crc kubenswrapper[4632]: I0313 10:47:17.796292 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:47:17 crc kubenswrapper[4632]: I0313 10:47:17.796302 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhljf\" (UniqueName: \"kubernetes.io/projected/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58-kube-api-access-vhljf\") on node \"crc\" DevicePath \"\"" Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.195229 4632 generic.go:334] "Generic (PLEG): container finished" podID="4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" containerID="debb15938677e3a625a25bb92735b8bbd7318e3df789b75b917f0f39fb974e78" exitCode=0 Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.197111 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8rs94" Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.197163 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rs94" event={"ID":"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58","Type":"ContainerDied","Data":"debb15938677e3a625a25bb92735b8bbd7318e3df789b75b917f0f39fb974e78"} Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.197235 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8rs94" event={"ID":"4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58","Type":"ContainerDied","Data":"2fcd4f33b9170a9e1f17871a87902d5c3a4bca4f5bff191585eaa7b78efa9537"} Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.197262 4632 scope.go:117] "RemoveContainer" containerID="debb15938677e3a625a25bb92735b8bbd7318e3df789b75b917f0f39fb974e78" Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.242718 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r7qtl"] Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.250572 4632 scope.go:117] "RemoveContainer" containerID="073ef57962037ed0b42aebc1b9b9c42dc3c21fd678c44de3f2fc1650fff2b40f" Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.260085 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8rs94"] Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.268765 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8rs94"] Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.272071 4632 scope.go:117] "RemoveContainer" containerID="9a68434b5791583fc729ae45ace124d9bffa3419f6e081364198559e5e722d2d" Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.316905 4632 scope.go:117] "RemoveContainer" containerID="debb15938677e3a625a25bb92735b8bbd7318e3df789b75b917f0f39fb974e78" Mar 13 10:47:18 crc kubenswrapper[4632]: E0313 10:47:18.317834 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"debb15938677e3a625a25bb92735b8bbd7318e3df789b75b917f0f39fb974e78\": container with ID starting with debb15938677e3a625a25bb92735b8bbd7318e3df789b75b917f0f39fb974e78 not found: ID does not exist" containerID="debb15938677e3a625a25bb92735b8bbd7318e3df789b75b917f0f39fb974e78" Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.318042 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"debb15938677e3a625a25bb92735b8bbd7318e3df789b75b917f0f39fb974e78"} err="failed to get container status \"debb15938677e3a625a25bb92735b8bbd7318e3df789b75b917f0f39fb974e78\": rpc error: code = NotFound desc = could not find container \"debb15938677e3a625a25bb92735b8bbd7318e3df789b75b917f0f39fb974e78\": container with ID starting with debb15938677e3a625a25bb92735b8bbd7318e3df789b75b917f0f39fb974e78 not found: ID does not exist" Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.318156 4632 scope.go:117] "RemoveContainer" containerID="073ef57962037ed0b42aebc1b9b9c42dc3c21fd678c44de3f2fc1650fff2b40f" Mar 13 10:47:18 crc kubenswrapper[4632]: E0313 10:47:18.318639 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"073ef57962037ed0b42aebc1b9b9c42dc3c21fd678c44de3f2fc1650fff2b40f\": container with ID starting with 073ef57962037ed0b42aebc1b9b9c42dc3c21fd678c44de3f2fc1650fff2b40f not found: ID does not exist" containerID="073ef57962037ed0b42aebc1b9b9c42dc3c21fd678c44de3f2fc1650fff2b40f" Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.318759 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"073ef57962037ed0b42aebc1b9b9c42dc3c21fd678c44de3f2fc1650fff2b40f"} err="failed to get container status \"073ef57962037ed0b42aebc1b9b9c42dc3c21fd678c44de3f2fc1650fff2b40f\": rpc error: code = NotFound desc = could not find container \"073ef57962037ed0b42aebc1b9b9c42dc3c21fd678c44de3f2fc1650fff2b40f\": container with ID starting with 073ef57962037ed0b42aebc1b9b9c42dc3c21fd678c44de3f2fc1650fff2b40f not found: ID does not exist" Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.318912 4632 scope.go:117] "RemoveContainer" containerID="9a68434b5791583fc729ae45ace124d9bffa3419f6e081364198559e5e722d2d" Mar 13 10:47:18 crc kubenswrapper[4632]: E0313 10:47:18.319559 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a68434b5791583fc729ae45ace124d9bffa3419f6e081364198559e5e722d2d\": container with ID starting with 9a68434b5791583fc729ae45ace124d9bffa3419f6e081364198559e5e722d2d not found: ID does not exist" containerID="9a68434b5791583fc729ae45ace124d9bffa3419f6e081364198559e5e722d2d" Mar 13 10:47:18 crc kubenswrapper[4632]: I0313 10:47:18.319673 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a68434b5791583fc729ae45ace124d9bffa3419f6e081364198559e5e722d2d"} err="failed to get container status \"9a68434b5791583fc729ae45ace124d9bffa3419f6e081364198559e5e722d2d\": rpc error: code = NotFound desc = could not find container \"9a68434b5791583fc729ae45ace124d9bffa3419f6e081364198559e5e722d2d\": container with ID starting with 9a68434b5791583fc729ae45ace124d9bffa3419f6e081364198559e5e722d2d not found: ID does not exist" Mar 13 10:47:19 crc kubenswrapper[4632]: I0313 10:47:19.206224 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-r7qtl" podUID="e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" containerName="registry-server" containerID="cri-o://1182e2999759e9be37ef50a31fa5017ebfa813a7c00f6df460a1c1c6e3b78a8f" gracePeriod=2 Mar 13 10:47:19 crc kubenswrapper[4632]: I0313 10:47:19.635932 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:19 crc kubenswrapper[4632]: I0313 10:47:19.782914 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-catalog-content\") pod \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\" (UID: \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\") " Mar 13 10:47:19 crc kubenswrapper[4632]: I0313 10:47:19.782995 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-utilities\") pod \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\" (UID: \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\") " Mar 13 10:47:19 crc kubenswrapper[4632]: I0313 10:47:19.783027 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5z5hv\" (UniqueName: \"kubernetes.io/projected/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-kube-api-access-5z5hv\") pod \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\" (UID: \"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511\") " Mar 13 10:47:19 crc kubenswrapper[4632]: I0313 10:47:19.784234 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-utilities" (OuterVolumeSpecName: "utilities") pod "e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" (UID: "e81415f8-ada1-4f0a-b1fd-92e2e5f5f511"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:47:19 crc kubenswrapper[4632]: I0313 10:47:19.789922 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-kube-api-access-5z5hv" (OuterVolumeSpecName: "kube-api-access-5z5hv") pod "e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" (UID: "e81415f8-ada1-4f0a-b1fd-92e2e5f5f511"). InnerVolumeSpecName "kube-api-access-5z5hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:47:19 crc kubenswrapper[4632]: I0313 10:47:19.817342 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" (UID: "e81415f8-ada1-4f0a-b1fd-92e2e5f5f511"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:47:19 crc kubenswrapper[4632]: I0313 10:47:19.884653 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:47:19 crc kubenswrapper[4632]: I0313 10:47:19.884686 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:47:19 crc kubenswrapper[4632]: I0313 10:47:19.884698 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5z5hv\" (UniqueName: \"kubernetes.io/projected/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511-kube-api-access-5z5hv\") on node \"crc\" DevicePath \"\"" Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.054616 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" path="/var/lib/kubelet/pods/4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58/volumes" Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.218237 4632 generic.go:334] "Generic (PLEG): container finished" podID="e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" containerID="1182e2999759e9be37ef50a31fa5017ebfa813a7c00f6df460a1c1c6e3b78a8f" exitCode=0 Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.218295 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7qtl" event={"ID":"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511","Type":"ContainerDied","Data":"1182e2999759e9be37ef50a31fa5017ebfa813a7c00f6df460a1c1c6e3b78a8f"} Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.218327 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-r7qtl" event={"ID":"e81415f8-ada1-4f0a-b1fd-92e2e5f5f511","Type":"ContainerDied","Data":"874f90ecd39b4972bd013b4da6787ff75bf2a57a461e69b7db2804dbda134bb6"} Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.218347 4632 scope.go:117] "RemoveContainer" containerID="1182e2999759e9be37ef50a31fa5017ebfa813a7c00f6df460a1c1c6e3b78a8f" Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.218477 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-r7qtl" Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.246344 4632 scope.go:117] "RemoveContainer" containerID="6869bdf68c35077e3e515fc4d4fc9c80e2292fab5161433ed27b5622d65b9902" Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.250168 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-r7qtl"] Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.266891 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-r7qtl"] Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.267378 4632 scope.go:117] "RemoveContainer" containerID="b3d79b73fcd58ac20d741acf9d96929a0df0db65f84dbb23ba98c1eff210414e" Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.309028 4632 scope.go:117] "RemoveContainer" containerID="1182e2999759e9be37ef50a31fa5017ebfa813a7c00f6df460a1c1c6e3b78a8f" Mar 13 10:47:20 crc kubenswrapper[4632]: E0313 10:47:20.310021 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1182e2999759e9be37ef50a31fa5017ebfa813a7c00f6df460a1c1c6e3b78a8f\": container with ID starting with 1182e2999759e9be37ef50a31fa5017ebfa813a7c00f6df460a1c1c6e3b78a8f not found: ID does not exist" containerID="1182e2999759e9be37ef50a31fa5017ebfa813a7c00f6df460a1c1c6e3b78a8f" Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.310074 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1182e2999759e9be37ef50a31fa5017ebfa813a7c00f6df460a1c1c6e3b78a8f"} err="failed to get container status \"1182e2999759e9be37ef50a31fa5017ebfa813a7c00f6df460a1c1c6e3b78a8f\": rpc error: code = NotFound desc = could not find container \"1182e2999759e9be37ef50a31fa5017ebfa813a7c00f6df460a1c1c6e3b78a8f\": container with ID starting with 1182e2999759e9be37ef50a31fa5017ebfa813a7c00f6df460a1c1c6e3b78a8f not found: ID does not exist" Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.310115 4632 scope.go:117] "RemoveContainer" containerID="6869bdf68c35077e3e515fc4d4fc9c80e2292fab5161433ed27b5622d65b9902" Mar 13 10:47:20 crc kubenswrapper[4632]: E0313 10:47:20.310650 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6869bdf68c35077e3e515fc4d4fc9c80e2292fab5161433ed27b5622d65b9902\": container with ID starting with 6869bdf68c35077e3e515fc4d4fc9c80e2292fab5161433ed27b5622d65b9902 not found: ID does not exist" containerID="6869bdf68c35077e3e515fc4d4fc9c80e2292fab5161433ed27b5622d65b9902" Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.310750 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6869bdf68c35077e3e515fc4d4fc9c80e2292fab5161433ed27b5622d65b9902"} err="failed to get container status \"6869bdf68c35077e3e515fc4d4fc9c80e2292fab5161433ed27b5622d65b9902\": rpc error: code = NotFound desc = could not find container \"6869bdf68c35077e3e515fc4d4fc9c80e2292fab5161433ed27b5622d65b9902\": container with ID starting with 6869bdf68c35077e3e515fc4d4fc9c80e2292fab5161433ed27b5622d65b9902 not found: ID does not exist" Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.310780 4632 scope.go:117] "RemoveContainer" containerID="b3d79b73fcd58ac20d741acf9d96929a0df0db65f84dbb23ba98c1eff210414e" Mar 13 10:47:20 crc kubenswrapper[4632]: E0313 10:47:20.311142 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3d79b73fcd58ac20d741acf9d96929a0df0db65f84dbb23ba98c1eff210414e\": container with ID starting with b3d79b73fcd58ac20d741acf9d96929a0df0db65f84dbb23ba98c1eff210414e not found: ID does not exist" containerID="b3d79b73fcd58ac20d741acf9d96929a0df0db65f84dbb23ba98c1eff210414e" Mar 13 10:47:20 crc kubenswrapper[4632]: I0313 10:47:20.311401 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3d79b73fcd58ac20d741acf9d96929a0df0db65f84dbb23ba98c1eff210414e"} err="failed to get container status \"b3d79b73fcd58ac20d741acf9d96929a0df0db65f84dbb23ba98c1eff210414e\": rpc error: code = NotFound desc = could not find container \"b3d79b73fcd58ac20d741acf9d96929a0df0db65f84dbb23ba98c1eff210414e\": container with ID starting with b3d79b73fcd58ac20d741acf9d96929a0df0db65f84dbb23ba98c1eff210414e not found: ID does not exist" Mar 13 10:47:22 crc kubenswrapper[4632]: I0313 10:47:22.055978 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" path="/var/lib/kubelet/pods/e81415f8-ada1-4f0a-b1fd-92e2e5f5f511/volumes" Mar 13 10:47:23 crc kubenswrapper[4632]: I0313 10:47:23.045759 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:47:23 crc kubenswrapper[4632]: E0313 10:47:23.045991 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:47:37 crc kubenswrapper[4632]: I0313 10:47:37.044305 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:47:37 crc kubenswrapper[4632]: E0313 10:47:37.045158 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:47:50 crc kubenswrapper[4632]: I0313 10:47:50.044179 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:47:50 crc kubenswrapper[4632]: I0313 10:47:50.493751 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"a5ebd5748d892637db30e6f25b4cdb7397d5f5e2a1d221a622054fbf7f8b83f2"} Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.165578 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556648-hbbkx"] Mar 13 10:48:00 crc kubenswrapper[4632]: E0313 10:48:00.166787 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" containerName="extract-content" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.166801 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" containerName="extract-content" Mar 13 10:48:00 crc kubenswrapper[4632]: E0313 10:48:00.166818 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" containerName="extract-utilities" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.166825 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" containerName="extract-utilities" Mar 13 10:48:00 crc kubenswrapper[4632]: E0313 10:48:00.166842 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" containerName="extract-utilities" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.166851 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" containerName="extract-utilities" Mar 13 10:48:00 crc kubenswrapper[4632]: E0313 10:48:00.166872 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" containerName="registry-server" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.166880 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" containerName="registry-server" Mar 13 10:48:00 crc kubenswrapper[4632]: E0313 10:48:00.166889 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" containerName="registry-server" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.166896 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" containerName="registry-server" Mar 13 10:48:00 crc kubenswrapper[4632]: E0313 10:48:00.166921 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" containerName="extract-content" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.166928 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" containerName="extract-content" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.167126 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="e81415f8-ada1-4f0a-b1fd-92e2e5f5f511" containerName="registry-server" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.167142 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a4a7db9-2c2b-42bc-a6fb-c33922f3ff58" containerName="registry-server" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.167798 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556648-hbbkx" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.172520 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.172863 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.175126 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.192402 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556648-hbbkx"] Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.289493 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhpjp\" (UniqueName: \"kubernetes.io/projected/5ab47075-381b-45d4-b6c8-c64ae6433ef1-kube-api-access-xhpjp\") pod \"auto-csr-approver-29556648-hbbkx\" (UID: \"5ab47075-381b-45d4-b6c8-c64ae6433ef1\") " pod="openshift-infra/auto-csr-approver-29556648-hbbkx" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.392000 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhpjp\" (UniqueName: \"kubernetes.io/projected/5ab47075-381b-45d4-b6c8-c64ae6433ef1-kube-api-access-xhpjp\") pod \"auto-csr-approver-29556648-hbbkx\" (UID: \"5ab47075-381b-45d4-b6c8-c64ae6433ef1\") " pod="openshift-infra/auto-csr-approver-29556648-hbbkx" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.411588 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhpjp\" (UniqueName: \"kubernetes.io/projected/5ab47075-381b-45d4-b6c8-c64ae6433ef1-kube-api-access-xhpjp\") pod \"auto-csr-approver-29556648-hbbkx\" (UID: \"5ab47075-381b-45d4-b6c8-c64ae6433ef1\") " pod="openshift-infra/auto-csr-approver-29556648-hbbkx" Mar 13 10:48:00 crc kubenswrapper[4632]: I0313 10:48:00.507144 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556648-hbbkx" Mar 13 10:48:01 crc kubenswrapper[4632]: I0313 10:48:01.085820 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556648-hbbkx"] Mar 13 10:48:01 crc kubenswrapper[4632]: I0313 10:48:01.611072 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556648-hbbkx" event={"ID":"5ab47075-381b-45d4-b6c8-c64ae6433ef1","Type":"ContainerStarted","Data":"da07aff5e6aa9184a5f24000561885b285e59acbd073ce5b368af221cda24a12"} Mar 13 10:48:02 crc kubenswrapper[4632]: I0313 10:48:02.622871 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556648-hbbkx" event={"ID":"5ab47075-381b-45d4-b6c8-c64ae6433ef1","Type":"ContainerStarted","Data":"ac1c75bd040311821d7426607144ebc256c3f11219f7a26012d50c7ce3c315ba"} Mar 13 10:48:02 crc kubenswrapper[4632]: I0313 10:48:02.640236 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556648-hbbkx" podStartSLOduration=1.9554097129999999 podStartE2EDuration="2.640216866s" podCreationTimestamp="2026-03-13 10:48:00 +0000 UTC" firstStartedPulling="2026-03-13 10:48:01.08372868 +0000 UTC m=+2655.106258813" lastFinishedPulling="2026-03-13 10:48:01.768535833 +0000 UTC m=+2655.791065966" observedRunningTime="2026-03-13 10:48:02.63755206 +0000 UTC m=+2656.660082213" watchObservedRunningTime="2026-03-13 10:48:02.640216866 +0000 UTC m=+2656.662746999" Mar 13 10:48:03 crc kubenswrapper[4632]: I0313 10:48:03.652288 4632 generic.go:334] "Generic (PLEG): container finished" podID="5ab47075-381b-45d4-b6c8-c64ae6433ef1" containerID="ac1c75bd040311821d7426607144ebc256c3f11219f7a26012d50c7ce3c315ba" exitCode=0 Mar 13 10:48:03 crc kubenswrapper[4632]: I0313 10:48:03.652965 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556648-hbbkx" event={"ID":"5ab47075-381b-45d4-b6c8-c64ae6433ef1","Type":"ContainerDied","Data":"ac1c75bd040311821d7426607144ebc256c3f11219f7a26012d50c7ce3c315ba"} Mar 13 10:48:05 crc kubenswrapper[4632]: I0313 10:48:05.024447 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556648-hbbkx" Mar 13 10:48:05 crc kubenswrapper[4632]: I0313 10:48:05.097381 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhpjp\" (UniqueName: \"kubernetes.io/projected/5ab47075-381b-45d4-b6c8-c64ae6433ef1-kube-api-access-xhpjp\") pod \"5ab47075-381b-45d4-b6c8-c64ae6433ef1\" (UID: \"5ab47075-381b-45d4-b6c8-c64ae6433ef1\") " Mar 13 10:48:05 crc kubenswrapper[4632]: I0313 10:48:05.108324 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ab47075-381b-45d4-b6c8-c64ae6433ef1-kube-api-access-xhpjp" (OuterVolumeSpecName: "kube-api-access-xhpjp") pod "5ab47075-381b-45d4-b6c8-c64ae6433ef1" (UID: "5ab47075-381b-45d4-b6c8-c64ae6433ef1"). InnerVolumeSpecName "kube-api-access-xhpjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:48:05 crc kubenswrapper[4632]: I0313 10:48:05.200433 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhpjp\" (UniqueName: \"kubernetes.io/projected/5ab47075-381b-45d4-b6c8-c64ae6433ef1-kube-api-access-xhpjp\") on node \"crc\" DevicePath \"\"" Mar 13 10:48:05 crc kubenswrapper[4632]: I0313 10:48:05.685176 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556648-hbbkx" event={"ID":"5ab47075-381b-45d4-b6c8-c64ae6433ef1","Type":"ContainerDied","Data":"da07aff5e6aa9184a5f24000561885b285e59acbd073ce5b368af221cda24a12"} Mar 13 10:48:05 crc kubenswrapper[4632]: I0313 10:48:05.685225 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="da07aff5e6aa9184a5f24000561885b285e59acbd073ce5b368af221cda24a12" Mar 13 10:48:05 crc kubenswrapper[4632]: I0313 10:48:05.685522 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556648-hbbkx" Mar 13 10:48:05 crc kubenswrapper[4632]: I0313 10:48:05.737170 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556642-f6lzb"] Mar 13 10:48:05 crc kubenswrapper[4632]: I0313 10:48:05.744811 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556642-f6lzb"] Mar 13 10:48:06 crc kubenswrapper[4632]: I0313 10:48:06.058147 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6df4a28-3b7b-4904-aa41-62caa26889a8" path="/var/lib/kubelet/pods/a6df4a28-3b7b-4904-aa41-62caa26889a8/volumes" Mar 13 10:48:09 crc kubenswrapper[4632]: I0313 10:48:09.788865 4632 scope.go:117] "RemoveContainer" containerID="2360a4309504ac747bcde26fcceae28cb04d811f34c5f4e463b65c45b06c70f5" Mar 13 10:48:23 crc kubenswrapper[4632]: I0313 10:48:23.858077 4632 generic.go:334] "Generic (PLEG): container finished" podID="ed1a2c50-a476-43ca-9764-e0ebffb14134" containerID="e94f04b40e30694707c6fd5089936e853269c95465d0464597019d512ad17ad4" exitCode=0 Mar 13 10:48:23 crc kubenswrapper[4632]: I0313 10:48:23.858164 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" event={"ID":"ed1a2c50-a476-43ca-9764-e0ebffb14134","Type":"ContainerDied","Data":"e94f04b40e30694707c6fd5089936e853269c95465d0464597019d512ad17ad4"} Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.267246 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.339221 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xh7rq\" (UniqueName: \"kubernetes.io/projected/ed1a2c50-a476-43ca-9764-e0ebffb14134-kube-api-access-xh7rq\") pod \"ed1a2c50-a476-43ca-9764-e0ebffb14134\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.339303 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-ssh-key-openstack-edpm-ipam\") pod \"ed1a2c50-a476-43ca-9764-e0ebffb14134\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.339426 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-inventory\") pod \"ed1a2c50-a476-43ca-9764-e0ebffb14134\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.339474 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-libvirt-combined-ca-bundle\") pod \"ed1a2c50-a476-43ca-9764-e0ebffb14134\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.339503 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-libvirt-secret-0\") pod \"ed1a2c50-a476-43ca-9764-e0ebffb14134\" (UID: \"ed1a2c50-a476-43ca-9764-e0ebffb14134\") " Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.349061 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "ed1a2c50-a476-43ca-9764-e0ebffb14134" (UID: "ed1a2c50-a476-43ca-9764-e0ebffb14134"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.359745 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed1a2c50-a476-43ca-9764-e0ebffb14134-kube-api-access-xh7rq" (OuterVolumeSpecName: "kube-api-access-xh7rq") pod "ed1a2c50-a476-43ca-9764-e0ebffb14134" (UID: "ed1a2c50-a476-43ca-9764-e0ebffb14134"). InnerVolumeSpecName "kube-api-access-xh7rq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.376378 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-inventory" (OuterVolumeSpecName: "inventory") pod "ed1a2c50-a476-43ca-9764-e0ebffb14134" (UID: "ed1a2c50-a476-43ca-9764-e0ebffb14134"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.377134 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ed1a2c50-a476-43ca-9764-e0ebffb14134" (UID: "ed1a2c50-a476-43ca-9764-e0ebffb14134"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.381431 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "ed1a2c50-a476-43ca-9764-e0ebffb14134" (UID: "ed1a2c50-a476-43ca-9764-e0ebffb14134"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.441816 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xh7rq\" (UniqueName: \"kubernetes.io/projected/ed1a2c50-a476-43ca-9764-e0ebffb14134-kube-api-access-xh7rq\") on node \"crc\" DevicePath \"\"" Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.442214 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.442231 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.442244 4632 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.442257 4632 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/ed1a2c50-a476-43ca-9764-e0ebffb14134-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.877421 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" event={"ID":"ed1a2c50-a476-43ca-9764-e0ebffb14134","Type":"ContainerDied","Data":"251492aaae45a40b4ef377f82a35e7e430a26d0b96081ac467780ad353a89a5b"} Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.877484 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="251492aaae45a40b4ef377f82a35e7e430a26d0b96081ac467780ad353a89a5b" Mar 13 10:48:25 crc kubenswrapper[4632]: I0313 10:48:25.877530 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-skjrh" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.008324 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq"] Mar 13 10:48:26 crc kubenswrapper[4632]: E0313 10:48:26.008798 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ab47075-381b-45d4-b6c8-c64ae6433ef1" containerName="oc" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.008822 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ab47075-381b-45d4-b6c8-c64ae6433ef1" containerName="oc" Mar 13 10:48:26 crc kubenswrapper[4632]: E0313 10:48:26.008859 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed1a2c50-a476-43ca-9764-e0ebffb14134" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.008870 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed1a2c50-a476-43ca-9764-e0ebffb14134" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.009161 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed1a2c50-a476-43ca-9764-e0ebffb14134" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.009189 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ab47075-381b-45d4-b6c8-c64ae6433ef1" containerName="oc" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.009829 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.024981 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.025422 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.025471 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.025521 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.025592 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.025521 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.025666 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.041404 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq"] Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.173343 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.173681 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.173722 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwrhx\" (UniqueName: \"kubernetes.io/projected/c897af06-c467-4ec3-aa76-c29a3ea3a462-kube-api-access-wwrhx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.173792 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.173822 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.173862 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.173926 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.173987 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.174061 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.174109 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.174149 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.275481 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.275565 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.275615 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.275681 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.275708 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.275740 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwrhx\" (UniqueName: \"kubernetes.io/projected/c897af06-c467-4ec3-aa76-c29a3ea3a462-kube-api-access-wwrhx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.275794 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.275823 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.275849 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.275897 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.275932 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.281026 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.281358 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.281463 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.281679 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.282662 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.282839 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.283095 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.284773 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.285655 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.286037 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.297265 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwrhx\" (UniqueName: \"kubernetes.io/projected/c897af06-c467-4ec3-aa76-c29a3ea3a462-kube-api-access-wwrhx\") pod \"nova-edpm-deployment-openstack-edpm-ipam-dl4cq\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.334688 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.865837 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq"] Mar 13 10:48:26 crc kubenswrapper[4632]: I0313 10:48:26.893694 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" event={"ID":"c897af06-c467-4ec3-aa76-c29a3ea3a462","Type":"ContainerStarted","Data":"d22525a75cd59bffbef3e23ac6c6e8d40f86fed8103fc04e25e403aefa74021b"} Mar 13 10:48:27 crc kubenswrapper[4632]: I0313 10:48:27.915392 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" event={"ID":"c897af06-c467-4ec3-aa76-c29a3ea3a462","Type":"ContainerStarted","Data":"3b4b530ae859a620ce6a4cb1762eca660904ffcd11138b81e9af76e94ecf0906"} Mar 13 10:48:27 crc kubenswrapper[4632]: I0313 10:48:27.942414 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" podStartSLOduration=2.538490863 podStartE2EDuration="2.942397696s" podCreationTimestamp="2026-03-13 10:48:25 +0000 UTC" firstStartedPulling="2026-03-13 10:48:26.882107038 +0000 UTC m=+2680.904637171" lastFinishedPulling="2026-03-13 10:48:27.286013871 +0000 UTC m=+2681.308544004" observedRunningTime="2026-03-13 10:48:27.934751119 +0000 UTC m=+2681.957281252" watchObservedRunningTime="2026-03-13 10:48:27.942397696 +0000 UTC m=+2681.964927829" Mar 13 10:50:00 crc kubenswrapper[4632]: I0313 10:50:00.153394 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556650-rskjc"] Mar 13 10:50:00 crc kubenswrapper[4632]: I0313 10:50:00.158252 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556650-rskjc" Mar 13 10:50:00 crc kubenswrapper[4632]: I0313 10:50:00.168726 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txjrr\" (UniqueName: \"kubernetes.io/projected/cc8dd4ae-e21e-4155-b617-19c85512d4fe-kube-api-access-txjrr\") pod \"auto-csr-approver-29556650-rskjc\" (UID: \"cc8dd4ae-e21e-4155-b617-19c85512d4fe\") " pod="openshift-infra/auto-csr-approver-29556650-rskjc" Mar 13 10:50:00 crc kubenswrapper[4632]: I0313 10:50:00.169400 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556650-rskjc"] Mar 13 10:50:00 crc kubenswrapper[4632]: I0313 10:50:00.171058 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:50:00 crc kubenswrapper[4632]: I0313 10:50:00.172033 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:50:00 crc kubenswrapper[4632]: I0313 10:50:00.172366 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:50:00 crc kubenswrapper[4632]: I0313 10:50:00.271003 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txjrr\" (UniqueName: \"kubernetes.io/projected/cc8dd4ae-e21e-4155-b617-19c85512d4fe-kube-api-access-txjrr\") pod \"auto-csr-approver-29556650-rskjc\" (UID: \"cc8dd4ae-e21e-4155-b617-19c85512d4fe\") " pod="openshift-infra/auto-csr-approver-29556650-rskjc" Mar 13 10:50:00 crc kubenswrapper[4632]: I0313 10:50:00.289458 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txjrr\" (UniqueName: \"kubernetes.io/projected/cc8dd4ae-e21e-4155-b617-19c85512d4fe-kube-api-access-txjrr\") pod \"auto-csr-approver-29556650-rskjc\" (UID: \"cc8dd4ae-e21e-4155-b617-19c85512d4fe\") " pod="openshift-infra/auto-csr-approver-29556650-rskjc" Mar 13 10:50:00 crc kubenswrapper[4632]: I0313 10:50:00.478983 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556650-rskjc" Mar 13 10:50:00 crc kubenswrapper[4632]: I0313 10:50:00.990254 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556650-rskjc"] Mar 13 10:50:01 crc kubenswrapper[4632]: I0313 10:50:01.776785 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556650-rskjc" event={"ID":"cc8dd4ae-e21e-4155-b617-19c85512d4fe","Type":"ContainerStarted","Data":"efa9d0470c5c6d7371d7df6219693ae11fcc4a5a3a75ca627a33c2dbd05b2574"} Mar 13 10:50:02 crc kubenswrapper[4632]: I0313 10:50:02.784749 4632 generic.go:334] "Generic (PLEG): container finished" podID="cc8dd4ae-e21e-4155-b617-19c85512d4fe" containerID="10fbcfced1ba7ba66a4ba615aa3b2aab72091e177631977720b60aac13ae9d0f" exitCode=0 Mar 13 10:50:02 crc kubenswrapper[4632]: I0313 10:50:02.784889 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556650-rskjc" event={"ID":"cc8dd4ae-e21e-4155-b617-19c85512d4fe","Type":"ContainerDied","Data":"10fbcfced1ba7ba66a4ba615aa3b2aab72091e177631977720b60aac13ae9d0f"} Mar 13 10:50:04 crc kubenswrapper[4632]: I0313 10:50:04.325155 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556650-rskjc" Mar 13 10:50:04 crc kubenswrapper[4632]: I0313 10:50:04.478868 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txjrr\" (UniqueName: \"kubernetes.io/projected/cc8dd4ae-e21e-4155-b617-19c85512d4fe-kube-api-access-txjrr\") pod \"cc8dd4ae-e21e-4155-b617-19c85512d4fe\" (UID: \"cc8dd4ae-e21e-4155-b617-19c85512d4fe\") " Mar 13 10:50:04 crc kubenswrapper[4632]: I0313 10:50:04.485499 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc8dd4ae-e21e-4155-b617-19c85512d4fe-kube-api-access-txjrr" (OuterVolumeSpecName: "kube-api-access-txjrr") pod "cc8dd4ae-e21e-4155-b617-19c85512d4fe" (UID: "cc8dd4ae-e21e-4155-b617-19c85512d4fe"). InnerVolumeSpecName "kube-api-access-txjrr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:50:04 crc kubenswrapper[4632]: I0313 10:50:04.580814 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txjrr\" (UniqueName: \"kubernetes.io/projected/cc8dd4ae-e21e-4155-b617-19c85512d4fe-kube-api-access-txjrr\") on node \"crc\" DevicePath \"\"" Mar 13 10:50:04 crc kubenswrapper[4632]: I0313 10:50:04.807397 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556650-rskjc" event={"ID":"cc8dd4ae-e21e-4155-b617-19c85512d4fe","Type":"ContainerDied","Data":"efa9d0470c5c6d7371d7df6219693ae11fcc4a5a3a75ca627a33c2dbd05b2574"} Mar 13 10:50:04 crc kubenswrapper[4632]: I0313 10:50:04.807875 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="efa9d0470c5c6d7371d7df6219693ae11fcc4a5a3a75ca627a33c2dbd05b2574" Mar 13 10:50:04 crc kubenswrapper[4632]: I0313 10:50:04.807677 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556650-rskjc" Mar 13 10:50:05 crc kubenswrapper[4632]: I0313 10:50:05.416752 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556644-dq2jd"] Mar 13 10:50:05 crc kubenswrapper[4632]: I0313 10:50:05.427494 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556644-dq2jd"] Mar 13 10:50:06 crc kubenswrapper[4632]: I0313 10:50:06.080775 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2462765e-6333-4e22-b4d7-ee2b2c6aa538" path="/var/lib/kubelet/pods/2462765e-6333-4e22-b4d7-ee2b2c6aa538/volumes" Mar 13 10:50:09 crc kubenswrapper[4632]: I0313 10:50:09.918425 4632 scope.go:117] "RemoveContainer" containerID="07b2fe4a97569c9089b7972685eb914fd04195d02c9e7b239121095e54e42352" Mar 13 10:50:10 crc kubenswrapper[4632]: I0313 10:50:10.461449 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:50:10 crc kubenswrapper[4632]: I0313 10:50:10.461525 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:50:40 crc kubenswrapper[4632]: I0313 10:50:40.460545 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:50:40 crc kubenswrapper[4632]: I0313 10:50:40.460990 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.555715 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-92wff"] Mar 13 10:50:57 crc kubenswrapper[4632]: E0313 10:50:57.558200 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc8dd4ae-e21e-4155-b617-19c85512d4fe" containerName="oc" Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.558227 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc8dd4ae-e21e-4155-b617-19c85512d4fe" containerName="oc" Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.558551 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc8dd4ae-e21e-4155-b617-19c85512d4fe" containerName="oc" Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.560329 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.563430 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-92wff"] Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.702105 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4kgss\" (UniqueName: \"kubernetes.io/projected/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-kube-api-access-4kgss\") pod \"redhat-operators-92wff\" (UID: \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\") " pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.702436 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-catalog-content\") pod \"redhat-operators-92wff\" (UID: \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\") " pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.702579 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-utilities\") pod \"redhat-operators-92wff\" (UID: \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\") " pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.804602 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-catalog-content\") pod \"redhat-operators-92wff\" (UID: \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\") " pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.804653 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-utilities\") pod \"redhat-operators-92wff\" (UID: \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\") " pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.804810 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4kgss\" (UniqueName: \"kubernetes.io/projected/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-kube-api-access-4kgss\") pod \"redhat-operators-92wff\" (UID: \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\") " pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.805185 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-catalog-content\") pod \"redhat-operators-92wff\" (UID: \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\") " pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.805414 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-utilities\") pod \"redhat-operators-92wff\" (UID: \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\") " pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.829638 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4kgss\" (UniqueName: \"kubernetes.io/projected/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-kube-api-access-4kgss\") pod \"redhat-operators-92wff\" (UID: \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\") " pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:50:57 crc kubenswrapper[4632]: I0313 10:50:57.891952 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:50:58 crc kubenswrapper[4632]: I0313 10:50:58.478790 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-92wff"] Mar 13 10:50:59 crc kubenswrapper[4632]: I0313 10:50:59.412827 4632 generic.go:334] "Generic (PLEG): container finished" podID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerID="b4b5f53f947256c5e10bbf4379bf34669d1d1bf56e886053918a7d4efb4d213d" exitCode=0 Mar 13 10:50:59 crc kubenswrapper[4632]: I0313 10:50:59.412999 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92wff" event={"ID":"2c8fcea0-c62d-4557-87e8-e46dee66bc0f","Type":"ContainerDied","Data":"b4b5f53f947256c5e10bbf4379bf34669d1d1bf56e886053918a7d4efb4d213d"} Mar 13 10:50:59 crc kubenswrapper[4632]: I0313 10:50:59.413184 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92wff" event={"ID":"2c8fcea0-c62d-4557-87e8-e46dee66bc0f","Type":"ContainerStarted","Data":"90a61bc58d69a6d619abaef868b90ed19460d1a8b36e5fcac632e0a2882d9502"} Mar 13 10:51:00 crc kubenswrapper[4632]: I0313 10:51:00.422861 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92wff" event={"ID":"2c8fcea0-c62d-4557-87e8-e46dee66bc0f","Type":"ContainerStarted","Data":"bcb3e10a9cf45d3b9e94b97d132bc40b54e004f985fa2291ed2a2caea2737a0a"} Mar 13 10:51:04 crc kubenswrapper[4632]: I0313 10:51:04.469255 4632 generic.go:334] "Generic (PLEG): container finished" podID="c897af06-c467-4ec3-aa76-c29a3ea3a462" containerID="3b4b530ae859a620ce6a4cb1762eca660904ffcd11138b81e9af76e94ecf0906" exitCode=0 Mar 13 10:51:04 crc kubenswrapper[4632]: I0313 10:51:04.469336 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" event={"ID":"c897af06-c467-4ec3-aa76-c29a3ea3a462","Type":"ContainerDied","Data":"3b4b530ae859a620ce6a4cb1762eca660904ffcd11138b81e9af76e94ecf0906"} Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.097124 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.178480 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-3\") pod \"c897af06-c467-4ec3-aa76-c29a3ea3a462\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.178576 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-combined-ca-bundle\") pod \"c897af06-c467-4ec3-aa76-c29a3ea3a462\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.178640 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-2\") pod \"c897af06-c467-4ec3-aa76-c29a3ea3a462\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.178678 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-inventory\") pod \"c897af06-c467-4ec3-aa76-c29a3ea3a462\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.178759 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-1\") pod \"c897af06-c467-4ec3-aa76-c29a3ea3a462\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.178795 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-extra-config-0\") pod \"c897af06-c467-4ec3-aa76-c29a3ea3a462\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.178880 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-0\") pod \"c897af06-c467-4ec3-aa76-c29a3ea3a462\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.179019 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-ssh-key-openstack-edpm-ipam\") pod \"c897af06-c467-4ec3-aa76-c29a3ea3a462\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.179049 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwrhx\" (UniqueName: \"kubernetes.io/projected/c897af06-c467-4ec3-aa76-c29a3ea3a462-kube-api-access-wwrhx\") pod \"c897af06-c467-4ec3-aa76-c29a3ea3a462\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.179082 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-migration-ssh-key-1\") pod \"c897af06-c467-4ec3-aa76-c29a3ea3a462\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.179104 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-migration-ssh-key-0\") pod \"c897af06-c467-4ec3-aa76-c29a3ea3a462\" (UID: \"c897af06-c467-4ec3-aa76-c29a3ea3a462\") " Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.185665 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c897af06-c467-4ec3-aa76-c29a3ea3a462-kube-api-access-wwrhx" (OuterVolumeSpecName: "kube-api-access-wwrhx") pod "c897af06-c467-4ec3-aa76-c29a3ea3a462" (UID: "c897af06-c467-4ec3-aa76-c29a3ea3a462"). InnerVolumeSpecName "kube-api-access-wwrhx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.190422 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "c897af06-c467-4ec3-aa76-c29a3ea3a462" (UID: "c897af06-c467-4ec3-aa76-c29a3ea3a462"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.220073 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "c897af06-c467-4ec3-aa76-c29a3ea3a462" (UID: "c897af06-c467-4ec3-aa76-c29a3ea3a462"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.227592 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "c897af06-c467-4ec3-aa76-c29a3ea3a462" (UID: "c897af06-c467-4ec3-aa76-c29a3ea3a462"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.234166 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c897af06-c467-4ec3-aa76-c29a3ea3a462" (UID: "c897af06-c467-4ec3-aa76-c29a3ea3a462"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.242847 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "c897af06-c467-4ec3-aa76-c29a3ea3a462" (UID: "c897af06-c467-4ec3-aa76-c29a3ea3a462"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.243262 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "c897af06-c467-4ec3-aa76-c29a3ea3a462" (UID: "c897af06-c467-4ec3-aa76-c29a3ea3a462"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.259839 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-inventory" (OuterVolumeSpecName: "inventory") pod "c897af06-c467-4ec3-aa76-c29a3ea3a462" (UID: "c897af06-c467-4ec3-aa76-c29a3ea3a462"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.260156 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "c897af06-c467-4ec3-aa76-c29a3ea3a462" (UID: "c897af06-c467-4ec3-aa76-c29a3ea3a462"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.267844 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "c897af06-c467-4ec3-aa76-c29a3ea3a462" (UID: "c897af06-c467-4ec3-aa76-c29a3ea3a462"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.277989 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "c897af06-c467-4ec3-aa76-c29a3ea3a462" (UID: "c897af06-c467-4ec3-aa76-c29a3ea3a462"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.281192 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.281222 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwrhx\" (UniqueName: \"kubernetes.io/projected/c897af06-c467-4ec3-aa76-c29a3ea3a462-kube-api-access-wwrhx\") on node \"crc\" DevicePath \"\"" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.281232 4632 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.281241 4632 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.281249 4632 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.281257 4632 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.281265 4632 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.281277 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.281287 4632 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.281295 4632 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.281305 4632 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/c897af06-c467-4ec3-aa76-c29a3ea3a462-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.489365 4632 generic.go:334] "Generic (PLEG): container finished" podID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerID="bcb3e10a9cf45d3b9e94b97d132bc40b54e004f985fa2291ed2a2caea2737a0a" exitCode=0 Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.489433 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92wff" event={"ID":"2c8fcea0-c62d-4557-87e8-e46dee66bc0f","Type":"ContainerDied","Data":"bcb3e10a9cf45d3b9e94b97d132bc40b54e004f985fa2291ed2a2caea2737a0a"} Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.491919 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" event={"ID":"c897af06-c467-4ec3-aa76-c29a3ea3a462","Type":"ContainerDied","Data":"d22525a75cd59bffbef3e23ac6c6e8d40f86fed8103fc04e25e403aefa74021b"} Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.492085 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d22525a75cd59bffbef3e23ac6c6e8d40f86fed8103fc04e25e403aefa74021b" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.492160 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-dl4cq" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.492884 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.625129 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw"] Mar 13 10:51:06 crc kubenswrapper[4632]: E0313 10:51:06.625612 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c897af06-c467-4ec3-aa76-c29a3ea3a462" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.625638 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c897af06-c467-4ec3-aa76-c29a3ea3a462" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.625881 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c897af06-c467-4ec3-aa76-c29a3ea3a462" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.626643 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.631065 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.634343 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.634343 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.634584 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.634879 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-qrzsx" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.639618 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw"] Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.688653 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.689002 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.689183 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.689302 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.689574 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxhtb\" (UniqueName: \"kubernetes.io/projected/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-kube-api-access-sxhtb\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.689708 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.689834 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.791370 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.791446 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.791493 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxhtb\" (UniqueName: \"kubernetes.io/projected/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-kube-api-access-sxhtb\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.791536 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.791583 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.791612 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.791636 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.796999 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.797042 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.797253 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.797460 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.797733 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.798084 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.812348 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxhtb\" (UniqueName: \"kubernetes.io/projected/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-kube-api-access-sxhtb\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:06 crc kubenswrapper[4632]: I0313 10:51:06.950789 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:51:07 crc kubenswrapper[4632]: I0313 10:51:07.502713 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92wff" event={"ID":"2c8fcea0-c62d-4557-87e8-e46dee66bc0f","Type":"ContainerStarted","Data":"9a3764407a73c02bf1afcfa64ffb7546776165b8c9cd2c18948a6549ce025879"} Mar 13 10:51:07 crc kubenswrapper[4632]: I0313 10:51:07.522864 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-92wff" podStartSLOduration=3.015256656 podStartE2EDuration="10.522844832s" podCreationTimestamp="2026-03-13 10:50:57 +0000 UTC" firstStartedPulling="2026-03-13 10:50:59.414910845 +0000 UTC m=+2833.437440978" lastFinishedPulling="2026-03-13 10:51:06.922499021 +0000 UTC m=+2840.945029154" observedRunningTime="2026-03-13 10:51:07.521073639 +0000 UTC m=+2841.543603782" watchObservedRunningTime="2026-03-13 10:51:07.522844832 +0000 UTC m=+2841.545374965" Mar 13 10:51:07 crc kubenswrapper[4632]: I0313 10:51:07.836227 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw"] Mar 13 10:51:07 crc kubenswrapper[4632]: I0313 10:51:07.893088 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:51:07 crc kubenswrapper[4632]: I0313 10:51:07.893151 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:51:08 crc kubenswrapper[4632]: I0313 10:51:08.510798 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" event={"ID":"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef","Type":"ContainerStarted","Data":"6ee75f7e51730eeb5489b2c1dd9bb7192cbbc507759459f37404c40b386bbdfc"} Mar 13 10:51:08 crc kubenswrapper[4632]: I0313 10:51:08.994043 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-92wff" podUID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerName="registry-server" probeResult="failure" output=< Mar 13 10:51:08 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:51:08 crc kubenswrapper[4632]: > Mar 13 10:51:09 crc kubenswrapper[4632]: I0313 10:51:09.529245 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" event={"ID":"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef","Type":"ContainerStarted","Data":"4b7c60398c3f33a538800ee1d366657cf73b7c4a7aff19c20e81ba7f933c624f"} Mar 13 10:51:10 crc kubenswrapper[4632]: I0313 10:51:10.461512 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:51:10 crc kubenswrapper[4632]: I0313 10:51:10.461593 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:51:10 crc kubenswrapper[4632]: I0313 10:51:10.461647 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:51:10 crc kubenswrapper[4632]: I0313 10:51:10.462439 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a5ebd5748d892637db30e6f25b4cdb7397d5f5e2a1d221a622054fbf7f8b83f2"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:51:10 crc kubenswrapper[4632]: I0313 10:51:10.462500 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://a5ebd5748d892637db30e6f25b4cdb7397d5f5e2a1d221a622054fbf7f8b83f2" gracePeriod=600 Mar 13 10:51:11 crc kubenswrapper[4632]: I0313 10:51:11.552714 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="a5ebd5748d892637db30e6f25b4cdb7397d5f5e2a1d221a622054fbf7f8b83f2" exitCode=0 Mar 13 10:51:11 crc kubenswrapper[4632]: I0313 10:51:11.552776 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"a5ebd5748d892637db30e6f25b4cdb7397d5f5e2a1d221a622054fbf7f8b83f2"} Mar 13 10:51:11 crc kubenswrapper[4632]: I0313 10:51:11.553256 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d"} Mar 13 10:51:11 crc kubenswrapper[4632]: I0313 10:51:11.553286 4632 scope.go:117] "RemoveContainer" containerID="7e6ad458a7a5f032b976d0f3e06f3cdb95d1f8fc235ab6b7d8f577ae0282cd20" Mar 13 10:51:11 crc kubenswrapper[4632]: I0313 10:51:11.574906 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" podStartSLOduration=5.099910812 podStartE2EDuration="5.574882118s" podCreationTimestamp="2026-03-13 10:51:06 +0000 UTC" firstStartedPulling="2026-03-13 10:51:07.808091547 +0000 UTC m=+2841.830621680" lastFinishedPulling="2026-03-13 10:51:08.283062853 +0000 UTC m=+2842.305592986" observedRunningTime="2026-03-13 10:51:09.55985902 +0000 UTC m=+2843.582389163" watchObservedRunningTime="2026-03-13 10:51:11.574882118 +0000 UTC m=+2845.597412251" Mar 13 10:51:18 crc kubenswrapper[4632]: I0313 10:51:18.933605 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-92wff" podUID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerName="registry-server" probeResult="failure" output=< Mar 13 10:51:18 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:51:18 crc kubenswrapper[4632]: > Mar 13 10:51:28 crc kubenswrapper[4632]: I0313 10:51:28.945721 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-92wff" podUID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerName="registry-server" probeResult="failure" output=< Mar 13 10:51:28 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:51:28 crc kubenswrapper[4632]: > Mar 13 10:51:38 crc kubenswrapper[4632]: I0313 10:51:38.938177 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-92wff" podUID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerName="registry-server" probeResult="failure" output=< Mar 13 10:51:38 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:51:38 crc kubenswrapper[4632]: > Mar 13 10:51:47 crc kubenswrapper[4632]: I0313 10:51:47.977074 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:51:48 crc kubenswrapper[4632]: I0313 10:51:48.061829 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:51:48 crc kubenswrapper[4632]: I0313 10:51:48.968619 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-92wff"] Mar 13 10:51:49 crc kubenswrapper[4632]: I0313 10:51:49.898650 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-92wff" podUID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerName="registry-server" containerID="cri-o://9a3764407a73c02bf1afcfa64ffb7546776165b8c9cd2c18948a6549ce025879" gracePeriod=2 Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.417417 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.473644 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-catalog-content\") pod \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\" (UID: \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\") " Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.473711 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-utilities\") pod \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\" (UID: \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\") " Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.473794 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4kgss\" (UniqueName: \"kubernetes.io/projected/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-kube-api-access-4kgss\") pod \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\" (UID: \"2c8fcea0-c62d-4557-87e8-e46dee66bc0f\") " Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.475464 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-utilities" (OuterVolumeSpecName: "utilities") pod "2c8fcea0-c62d-4557-87e8-e46dee66bc0f" (UID: "2c8fcea0-c62d-4557-87e8-e46dee66bc0f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.482179 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-kube-api-access-4kgss" (OuterVolumeSpecName: "kube-api-access-4kgss") pod "2c8fcea0-c62d-4557-87e8-e46dee66bc0f" (UID: "2c8fcea0-c62d-4557-87e8-e46dee66bc0f"). InnerVolumeSpecName "kube-api-access-4kgss". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.576578 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.576615 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4kgss\" (UniqueName: \"kubernetes.io/projected/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-kube-api-access-4kgss\") on node \"crc\" DevicePath \"\"" Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.639683 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2c8fcea0-c62d-4557-87e8-e46dee66bc0f" (UID: "2c8fcea0-c62d-4557-87e8-e46dee66bc0f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.678155 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2c8fcea0-c62d-4557-87e8-e46dee66bc0f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.908722 4632 generic.go:334] "Generic (PLEG): container finished" podID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerID="9a3764407a73c02bf1afcfa64ffb7546776165b8c9cd2c18948a6549ce025879" exitCode=0 Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.908761 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92wff" event={"ID":"2c8fcea0-c62d-4557-87e8-e46dee66bc0f","Type":"ContainerDied","Data":"9a3764407a73c02bf1afcfa64ffb7546776165b8c9cd2c18948a6549ce025879"} Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.908779 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-92wff" Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.909835 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-92wff" event={"ID":"2c8fcea0-c62d-4557-87e8-e46dee66bc0f","Type":"ContainerDied","Data":"90a61bc58d69a6d619abaef868b90ed19460d1a8b36e5fcac632e0a2882d9502"} Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.909947 4632 scope.go:117] "RemoveContainer" containerID="9a3764407a73c02bf1afcfa64ffb7546776165b8c9cd2c18948a6549ce025879" Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.957651 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-92wff"] Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.958043 4632 scope.go:117] "RemoveContainer" containerID="bcb3e10a9cf45d3b9e94b97d132bc40b54e004f985fa2291ed2a2caea2737a0a" Mar 13 10:51:50 crc kubenswrapper[4632]: I0313 10:51:50.968615 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-92wff"] Mar 13 10:51:51 crc kubenswrapper[4632]: I0313 10:51:51.001567 4632 scope.go:117] "RemoveContainer" containerID="b4b5f53f947256c5e10bbf4379bf34669d1d1bf56e886053918a7d4efb4d213d" Mar 13 10:51:51 crc kubenswrapper[4632]: I0313 10:51:51.039499 4632 scope.go:117] "RemoveContainer" containerID="9a3764407a73c02bf1afcfa64ffb7546776165b8c9cd2c18948a6549ce025879" Mar 13 10:51:51 crc kubenswrapper[4632]: E0313 10:51:51.040073 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9a3764407a73c02bf1afcfa64ffb7546776165b8c9cd2c18948a6549ce025879\": container with ID starting with 9a3764407a73c02bf1afcfa64ffb7546776165b8c9cd2c18948a6549ce025879 not found: ID does not exist" containerID="9a3764407a73c02bf1afcfa64ffb7546776165b8c9cd2c18948a6549ce025879" Mar 13 10:51:51 crc kubenswrapper[4632]: I0313 10:51:51.040143 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9a3764407a73c02bf1afcfa64ffb7546776165b8c9cd2c18948a6549ce025879"} err="failed to get container status \"9a3764407a73c02bf1afcfa64ffb7546776165b8c9cd2c18948a6549ce025879\": rpc error: code = NotFound desc = could not find container \"9a3764407a73c02bf1afcfa64ffb7546776165b8c9cd2c18948a6549ce025879\": container with ID starting with 9a3764407a73c02bf1afcfa64ffb7546776165b8c9cd2c18948a6549ce025879 not found: ID does not exist" Mar 13 10:51:51 crc kubenswrapper[4632]: I0313 10:51:51.040203 4632 scope.go:117] "RemoveContainer" containerID="bcb3e10a9cf45d3b9e94b97d132bc40b54e004f985fa2291ed2a2caea2737a0a" Mar 13 10:51:51 crc kubenswrapper[4632]: E0313 10:51:51.040679 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcb3e10a9cf45d3b9e94b97d132bc40b54e004f985fa2291ed2a2caea2737a0a\": container with ID starting with bcb3e10a9cf45d3b9e94b97d132bc40b54e004f985fa2291ed2a2caea2737a0a not found: ID does not exist" containerID="bcb3e10a9cf45d3b9e94b97d132bc40b54e004f985fa2291ed2a2caea2737a0a" Mar 13 10:51:51 crc kubenswrapper[4632]: I0313 10:51:51.040715 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcb3e10a9cf45d3b9e94b97d132bc40b54e004f985fa2291ed2a2caea2737a0a"} err="failed to get container status \"bcb3e10a9cf45d3b9e94b97d132bc40b54e004f985fa2291ed2a2caea2737a0a\": rpc error: code = NotFound desc = could not find container \"bcb3e10a9cf45d3b9e94b97d132bc40b54e004f985fa2291ed2a2caea2737a0a\": container with ID starting with bcb3e10a9cf45d3b9e94b97d132bc40b54e004f985fa2291ed2a2caea2737a0a not found: ID does not exist" Mar 13 10:51:51 crc kubenswrapper[4632]: I0313 10:51:51.040735 4632 scope.go:117] "RemoveContainer" containerID="b4b5f53f947256c5e10bbf4379bf34669d1d1bf56e886053918a7d4efb4d213d" Mar 13 10:51:51 crc kubenswrapper[4632]: E0313 10:51:51.041345 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4b5f53f947256c5e10bbf4379bf34669d1d1bf56e886053918a7d4efb4d213d\": container with ID starting with b4b5f53f947256c5e10bbf4379bf34669d1d1bf56e886053918a7d4efb4d213d not found: ID does not exist" containerID="b4b5f53f947256c5e10bbf4379bf34669d1d1bf56e886053918a7d4efb4d213d" Mar 13 10:51:51 crc kubenswrapper[4632]: I0313 10:51:51.041396 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4b5f53f947256c5e10bbf4379bf34669d1d1bf56e886053918a7d4efb4d213d"} err="failed to get container status \"b4b5f53f947256c5e10bbf4379bf34669d1d1bf56e886053918a7d4efb4d213d\": rpc error: code = NotFound desc = could not find container \"b4b5f53f947256c5e10bbf4379bf34669d1d1bf56e886053918a7d4efb4d213d\": container with ID starting with b4b5f53f947256c5e10bbf4379bf34669d1d1bf56e886053918a7d4efb4d213d not found: ID does not exist" Mar 13 10:51:52 crc kubenswrapper[4632]: I0313 10:51:52.057636 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" path="/var/lib/kubelet/pods/2c8fcea0-c62d-4557-87e8-e46dee66bc0f/volumes" Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.157823 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556652-ghswk"] Mar 13 10:52:00 crc kubenswrapper[4632]: E0313 10:52:00.158757 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerName="extract-content" Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.158770 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerName="extract-content" Mar 13 10:52:00 crc kubenswrapper[4632]: E0313 10:52:00.158784 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerName="registry-server" Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.158789 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerName="registry-server" Mar 13 10:52:00 crc kubenswrapper[4632]: E0313 10:52:00.158812 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerName="extract-utilities" Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.158820 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerName="extract-utilities" Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.159024 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c8fcea0-c62d-4557-87e8-e46dee66bc0f" containerName="registry-server" Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.159653 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556652-ghswk" Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.162861 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.162958 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.165230 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.184132 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556652-ghswk"] Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.212152 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ttnfv\" (UniqueName: \"kubernetes.io/projected/27a43bdf-8be4-458b-99ff-4135a684962a-kube-api-access-ttnfv\") pod \"auto-csr-approver-29556652-ghswk\" (UID: \"27a43bdf-8be4-458b-99ff-4135a684962a\") " pod="openshift-infra/auto-csr-approver-29556652-ghswk" Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.313386 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ttnfv\" (UniqueName: \"kubernetes.io/projected/27a43bdf-8be4-458b-99ff-4135a684962a-kube-api-access-ttnfv\") pod \"auto-csr-approver-29556652-ghswk\" (UID: \"27a43bdf-8be4-458b-99ff-4135a684962a\") " pod="openshift-infra/auto-csr-approver-29556652-ghswk" Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.334553 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ttnfv\" (UniqueName: \"kubernetes.io/projected/27a43bdf-8be4-458b-99ff-4135a684962a-kube-api-access-ttnfv\") pod \"auto-csr-approver-29556652-ghswk\" (UID: \"27a43bdf-8be4-458b-99ff-4135a684962a\") " pod="openshift-infra/auto-csr-approver-29556652-ghswk" Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.482554 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556652-ghswk" Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.940956 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556652-ghswk"] Mar 13 10:52:00 crc kubenswrapper[4632]: I0313 10:52:00.993723 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556652-ghswk" event={"ID":"27a43bdf-8be4-458b-99ff-4135a684962a","Type":"ContainerStarted","Data":"6f4c18d16927eaf56fce1884f8280121cccf26e3a8d837a33fa467c468e2742c"} Mar 13 10:52:03 crc kubenswrapper[4632]: I0313 10:52:03.016102 4632 generic.go:334] "Generic (PLEG): container finished" podID="27a43bdf-8be4-458b-99ff-4135a684962a" containerID="5c30aacc6e4fc680fbb7e668912e806d2164d1cac4b7e0e4c5e8b4688d3e76cd" exitCode=0 Mar 13 10:52:03 crc kubenswrapper[4632]: I0313 10:52:03.016529 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556652-ghswk" event={"ID":"27a43bdf-8be4-458b-99ff-4135a684962a","Type":"ContainerDied","Data":"5c30aacc6e4fc680fbb7e668912e806d2164d1cac4b7e0e4c5e8b4688d3e76cd"} Mar 13 10:52:04 crc kubenswrapper[4632]: I0313 10:52:04.441623 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556652-ghswk" Mar 13 10:52:04 crc kubenswrapper[4632]: I0313 10:52:04.497733 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ttnfv\" (UniqueName: \"kubernetes.io/projected/27a43bdf-8be4-458b-99ff-4135a684962a-kube-api-access-ttnfv\") pod \"27a43bdf-8be4-458b-99ff-4135a684962a\" (UID: \"27a43bdf-8be4-458b-99ff-4135a684962a\") " Mar 13 10:52:04 crc kubenswrapper[4632]: I0313 10:52:04.517863 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27a43bdf-8be4-458b-99ff-4135a684962a-kube-api-access-ttnfv" (OuterVolumeSpecName: "kube-api-access-ttnfv") pod "27a43bdf-8be4-458b-99ff-4135a684962a" (UID: "27a43bdf-8be4-458b-99ff-4135a684962a"). InnerVolumeSpecName "kube-api-access-ttnfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:52:04 crc kubenswrapper[4632]: I0313 10:52:04.601368 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ttnfv\" (UniqueName: \"kubernetes.io/projected/27a43bdf-8be4-458b-99ff-4135a684962a-kube-api-access-ttnfv\") on node \"crc\" DevicePath \"\"" Mar 13 10:52:05 crc kubenswrapper[4632]: I0313 10:52:05.035544 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556652-ghswk" event={"ID":"27a43bdf-8be4-458b-99ff-4135a684962a","Type":"ContainerDied","Data":"6f4c18d16927eaf56fce1884f8280121cccf26e3a8d837a33fa467c468e2742c"} Mar 13 10:52:05 crc kubenswrapper[4632]: I0313 10:52:05.035594 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6f4c18d16927eaf56fce1884f8280121cccf26e3a8d837a33fa467c468e2742c" Mar 13 10:52:05 crc kubenswrapper[4632]: I0313 10:52:05.035607 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556652-ghswk" Mar 13 10:52:05 crc kubenswrapper[4632]: I0313 10:52:05.562026 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556646-blkxp"] Mar 13 10:52:05 crc kubenswrapper[4632]: I0313 10:52:05.572578 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556646-blkxp"] Mar 13 10:52:06 crc kubenswrapper[4632]: I0313 10:52:06.055496 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93948d53-dbf3-47ce-8af0-bee10cc7e246" path="/var/lib/kubelet/pods/93948d53-dbf3-47ce-8af0-bee10cc7e246/volumes" Mar 13 10:52:10 crc kubenswrapper[4632]: I0313 10:52:10.013834 4632 scope.go:117] "RemoveContainer" containerID="92f6939c452dda4592aa326adcecce982f4fafb95f93ce909a101db10372c2ab" Mar 13 10:53:10 crc kubenswrapper[4632]: I0313 10:53:10.461094 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:53:10 crc kubenswrapper[4632]: I0313 10:53:10.461488 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:53:27 crc kubenswrapper[4632]: I0313 10:53:27.868320 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-774lb"] Mar 13 10:53:27 crc kubenswrapper[4632]: E0313 10:53:27.869320 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27a43bdf-8be4-458b-99ff-4135a684962a" containerName="oc" Mar 13 10:53:27 crc kubenswrapper[4632]: I0313 10:53:27.871811 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="27a43bdf-8be4-458b-99ff-4135a684962a" containerName="oc" Mar 13 10:53:27 crc kubenswrapper[4632]: I0313 10:53:27.872117 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="27a43bdf-8be4-458b-99ff-4135a684962a" containerName="oc" Mar 13 10:53:27 crc kubenswrapper[4632]: I0313 10:53:27.873534 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:27 crc kubenswrapper[4632]: I0313 10:53:27.888625 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-774lb"] Mar 13 10:53:28 crc kubenswrapper[4632]: I0313 10:53:28.036288 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/560629a7-9dec-4eb7-8c73-a8f097293daa-utilities\") pod \"community-operators-774lb\" (UID: \"560629a7-9dec-4eb7-8c73-a8f097293daa\") " pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:28 crc kubenswrapper[4632]: I0313 10:53:28.036338 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rc4f5\" (UniqueName: \"kubernetes.io/projected/560629a7-9dec-4eb7-8c73-a8f097293daa-kube-api-access-rc4f5\") pod \"community-operators-774lb\" (UID: \"560629a7-9dec-4eb7-8c73-a8f097293daa\") " pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:28 crc kubenswrapper[4632]: I0313 10:53:28.036477 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/560629a7-9dec-4eb7-8c73-a8f097293daa-catalog-content\") pod \"community-operators-774lb\" (UID: \"560629a7-9dec-4eb7-8c73-a8f097293daa\") " pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:28 crc kubenswrapper[4632]: I0313 10:53:28.137968 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/560629a7-9dec-4eb7-8c73-a8f097293daa-catalog-content\") pod \"community-operators-774lb\" (UID: \"560629a7-9dec-4eb7-8c73-a8f097293daa\") " pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:28 crc kubenswrapper[4632]: I0313 10:53:28.138068 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/560629a7-9dec-4eb7-8c73-a8f097293daa-utilities\") pod \"community-operators-774lb\" (UID: \"560629a7-9dec-4eb7-8c73-a8f097293daa\") " pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:28 crc kubenswrapper[4632]: I0313 10:53:28.138109 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rc4f5\" (UniqueName: \"kubernetes.io/projected/560629a7-9dec-4eb7-8c73-a8f097293daa-kube-api-access-rc4f5\") pod \"community-operators-774lb\" (UID: \"560629a7-9dec-4eb7-8c73-a8f097293daa\") " pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:28 crc kubenswrapper[4632]: I0313 10:53:28.138510 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/560629a7-9dec-4eb7-8c73-a8f097293daa-catalog-content\") pod \"community-operators-774lb\" (UID: \"560629a7-9dec-4eb7-8c73-a8f097293daa\") " pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:28 crc kubenswrapper[4632]: I0313 10:53:28.138610 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/560629a7-9dec-4eb7-8c73-a8f097293daa-utilities\") pod \"community-operators-774lb\" (UID: \"560629a7-9dec-4eb7-8c73-a8f097293daa\") " pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:28 crc kubenswrapper[4632]: I0313 10:53:28.162278 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rc4f5\" (UniqueName: \"kubernetes.io/projected/560629a7-9dec-4eb7-8c73-a8f097293daa-kube-api-access-rc4f5\") pod \"community-operators-774lb\" (UID: \"560629a7-9dec-4eb7-8c73-a8f097293daa\") " pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:28 crc kubenswrapper[4632]: I0313 10:53:28.195470 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:28 crc kubenswrapper[4632]: I0313 10:53:28.594357 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-774lb"] Mar 13 10:53:29 crc kubenswrapper[4632]: I0313 10:53:29.126036 4632 generic.go:334] "Generic (PLEG): container finished" podID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerID="f9fbff406d14d8da11f86810a1b1b035215dd5c6179ac20e5ddd29194bd3f5d6" exitCode=0 Mar 13 10:53:29 crc kubenswrapper[4632]: I0313 10:53:29.126109 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-774lb" event={"ID":"560629a7-9dec-4eb7-8c73-a8f097293daa","Type":"ContainerDied","Data":"f9fbff406d14d8da11f86810a1b1b035215dd5c6179ac20e5ddd29194bd3f5d6"} Mar 13 10:53:29 crc kubenswrapper[4632]: I0313 10:53:29.126300 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-774lb" event={"ID":"560629a7-9dec-4eb7-8c73-a8f097293daa","Type":"ContainerStarted","Data":"4ae9494de264dfa5dcfb2c9e6166d64886aa8f640f54445b6eadb498ad356c8c"} Mar 13 10:53:34 crc kubenswrapper[4632]: I0313 10:53:34.178701 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-774lb" event={"ID":"560629a7-9dec-4eb7-8c73-a8f097293daa","Type":"ContainerStarted","Data":"d5e77ef64ff23f92ed48258b81d7d0310ada291a691626009608a75068a59888"} Mar 13 10:53:35 crc kubenswrapper[4632]: I0313 10:53:35.192074 4632 generic.go:334] "Generic (PLEG): container finished" podID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerID="d5e77ef64ff23f92ed48258b81d7d0310ada291a691626009608a75068a59888" exitCode=0 Mar 13 10:53:35 crc kubenswrapper[4632]: I0313 10:53:35.192239 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-774lb" event={"ID":"560629a7-9dec-4eb7-8c73-a8f097293daa","Type":"ContainerDied","Data":"d5e77ef64ff23f92ed48258b81d7d0310ada291a691626009608a75068a59888"} Mar 13 10:53:35 crc kubenswrapper[4632]: I0313 10:53:35.192704 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-774lb" event={"ID":"560629a7-9dec-4eb7-8c73-a8f097293daa","Type":"ContainerStarted","Data":"1b995d3ea46318dbc1da1ae83e15d1a1943f08993ba4772ae9cb4b946ae10e86"} Mar 13 10:53:35 crc kubenswrapper[4632]: I0313 10:53:35.228176 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-774lb" podStartSLOduration=2.410377626 podStartE2EDuration="8.228154509s" podCreationTimestamp="2026-03-13 10:53:27 +0000 UTC" firstStartedPulling="2026-03-13 10:53:29.128371059 +0000 UTC m=+2983.150901202" lastFinishedPulling="2026-03-13 10:53:34.946147942 +0000 UTC m=+2988.968678085" observedRunningTime="2026-03-13 10:53:35.221011094 +0000 UTC m=+2989.243541227" watchObservedRunningTime="2026-03-13 10:53:35.228154509 +0000 UTC m=+2989.250684652" Mar 13 10:53:38 crc kubenswrapper[4632]: I0313 10:53:38.196349 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:38 crc kubenswrapper[4632]: I0313 10:53:38.196613 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:38 crc kubenswrapper[4632]: I0313 10:53:38.245906 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:40 crc kubenswrapper[4632]: I0313 10:53:40.460721 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:53:40 crc kubenswrapper[4632]: I0313 10:53:40.461083 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:53:48 crc kubenswrapper[4632]: I0313 10:53:48.253230 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-774lb" Mar 13 10:53:48 crc kubenswrapper[4632]: I0313 10:53:48.356738 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-774lb"] Mar 13 10:53:48 crc kubenswrapper[4632]: I0313 10:53:48.418740 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fdtl7"] Mar 13 10:53:48 crc kubenswrapper[4632]: I0313 10:53:48.418997 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fdtl7" podUID="a01bcaf0-e2c1-495b-bc6d-a57978c7817b" containerName="registry-server" containerID="cri-o://e2fda5f5d13b5663527978c3fcfd3cccd016f3f222280313a4a0c5c88b5212d7" gracePeriod=2 Mar 13 10:53:48 crc kubenswrapper[4632]: I0313 10:53:48.894769 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.068510 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-utilities\") pod \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\" (UID: \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\") " Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.068880 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgb8q\" (UniqueName: \"kubernetes.io/projected/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-kube-api-access-sgb8q\") pod \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\" (UID: \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\") " Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.069217 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-catalog-content\") pod \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\" (UID: \"a01bcaf0-e2c1-495b-bc6d-a57978c7817b\") " Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.069646 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-utilities" (OuterVolumeSpecName: "utilities") pod "a01bcaf0-e2c1-495b-bc6d-a57978c7817b" (UID: "a01bcaf0-e2c1-495b-bc6d-a57978c7817b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.070443 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.074570 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-kube-api-access-sgb8q" (OuterVolumeSpecName: "kube-api-access-sgb8q") pod "a01bcaf0-e2c1-495b-bc6d-a57978c7817b" (UID: "a01bcaf0-e2c1-495b-bc6d-a57978c7817b"). InnerVolumeSpecName "kube-api-access-sgb8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.131511 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a01bcaf0-e2c1-495b-bc6d-a57978c7817b" (UID: "a01bcaf0-e2c1-495b-bc6d-a57978c7817b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.172544 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.172584 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgb8q\" (UniqueName: \"kubernetes.io/projected/a01bcaf0-e2c1-495b-bc6d-a57978c7817b-kube-api-access-sgb8q\") on node \"crc\" DevicePath \"\"" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.328198 4632 generic.go:334] "Generic (PLEG): container finished" podID="a01bcaf0-e2c1-495b-bc6d-a57978c7817b" containerID="e2fda5f5d13b5663527978c3fcfd3cccd016f3f222280313a4a0c5c88b5212d7" exitCode=0 Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.328240 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdtl7" event={"ID":"a01bcaf0-e2c1-495b-bc6d-a57978c7817b","Type":"ContainerDied","Data":"e2fda5f5d13b5663527978c3fcfd3cccd016f3f222280313a4a0c5c88b5212d7"} Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.328266 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fdtl7" event={"ID":"a01bcaf0-e2c1-495b-bc6d-a57978c7817b","Type":"ContainerDied","Data":"b009a6765de2ecd804f8d033b53e49abb32cd27ba05ed7eacca8b430a75a2575"} Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.328285 4632 scope.go:117] "RemoveContainer" containerID="e2fda5f5d13b5663527978c3fcfd3cccd016f3f222280313a4a0c5c88b5212d7" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.328417 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fdtl7" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.364106 4632 scope.go:117] "RemoveContainer" containerID="90d2b05d55e78b9f9829c2ee4bf7bbc01510b17dfbff9b47dda76cd10b610010" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.379471 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fdtl7"] Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.387813 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fdtl7"] Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.412908 4632 scope.go:117] "RemoveContainer" containerID="e0b81f7099d2b6e974290907876a0f00faa065b0c428d84017f9ee0db229bc73" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.449767 4632 scope.go:117] "RemoveContainer" containerID="e2fda5f5d13b5663527978c3fcfd3cccd016f3f222280313a4a0c5c88b5212d7" Mar 13 10:53:49 crc kubenswrapper[4632]: E0313 10:53:49.450475 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e2fda5f5d13b5663527978c3fcfd3cccd016f3f222280313a4a0c5c88b5212d7\": container with ID starting with e2fda5f5d13b5663527978c3fcfd3cccd016f3f222280313a4a0c5c88b5212d7 not found: ID does not exist" containerID="e2fda5f5d13b5663527978c3fcfd3cccd016f3f222280313a4a0c5c88b5212d7" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.450527 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e2fda5f5d13b5663527978c3fcfd3cccd016f3f222280313a4a0c5c88b5212d7"} err="failed to get container status \"e2fda5f5d13b5663527978c3fcfd3cccd016f3f222280313a4a0c5c88b5212d7\": rpc error: code = NotFound desc = could not find container \"e2fda5f5d13b5663527978c3fcfd3cccd016f3f222280313a4a0c5c88b5212d7\": container with ID starting with e2fda5f5d13b5663527978c3fcfd3cccd016f3f222280313a4a0c5c88b5212d7 not found: ID does not exist" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.450560 4632 scope.go:117] "RemoveContainer" containerID="90d2b05d55e78b9f9829c2ee4bf7bbc01510b17dfbff9b47dda76cd10b610010" Mar 13 10:53:49 crc kubenswrapper[4632]: E0313 10:53:49.450970 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90d2b05d55e78b9f9829c2ee4bf7bbc01510b17dfbff9b47dda76cd10b610010\": container with ID starting with 90d2b05d55e78b9f9829c2ee4bf7bbc01510b17dfbff9b47dda76cd10b610010 not found: ID does not exist" containerID="90d2b05d55e78b9f9829c2ee4bf7bbc01510b17dfbff9b47dda76cd10b610010" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.451012 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90d2b05d55e78b9f9829c2ee4bf7bbc01510b17dfbff9b47dda76cd10b610010"} err="failed to get container status \"90d2b05d55e78b9f9829c2ee4bf7bbc01510b17dfbff9b47dda76cd10b610010\": rpc error: code = NotFound desc = could not find container \"90d2b05d55e78b9f9829c2ee4bf7bbc01510b17dfbff9b47dda76cd10b610010\": container with ID starting with 90d2b05d55e78b9f9829c2ee4bf7bbc01510b17dfbff9b47dda76cd10b610010 not found: ID does not exist" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.451036 4632 scope.go:117] "RemoveContainer" containerID="e0b81f7099d2b6e974290907876a0f00faa065b0c428d84017f9ee0db229bc73" Mar 13 10:53:49 crc kubenswrapper[4632]: E0313 10:53:49.451285 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0b81f7099d2b6e974290907876a0f00faa065b0c428d84017f9ee0db229bc73\": container with ID starting with e0b81f7099d2b6e974290907876a0f00faa065b0c428d84017f9ee0db229bc73 not found: ID does not exist" containerID="e0b81f7099d2b6e974290907876a0f00faa065b0c428d84017f9ee0db229bc73" Mar 13 10:53:49 crc kubenswrapper[4632]: I0313 10:53:49.451312 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0b81f7099d2b6e974290907876a0f00faa065b0c428d84017f9ee0db229bc73"} err="failed to get container status \"e0b81f7099d2b6e974290907876a0f00faa065b0c428d84017f9ee0db229bc73\": rpc error: code = NotFound desc = could not find container \"e0b81f7099d2b6e974290907876a0f00faa065b0c428d84017f9ee0db229bc73\": container with ID starting with e0b81f7099d2b6e974290907876a0f00faa065b0c428d84017f9ee0db229bc73 not found: ID does not exist" Mar 13 10:53:50 crc kubenswrapper[4632]: I0313 10:53:50.058502 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a01bcaf0-e2c1-495b-bc6d-a57978c7817b" path="/var/lib/kubelet/pods/a01bcaf0-e2c1-495b-bc6d-a57978c7817b/volumes" Mar 13 10:54:00 crc kubenswrapper[4632]: I0313 10:54:00.143986 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556654-htcrm"] Mar 13 10:54:00 crc kubenswrapper[4632]: E0313 10:54:00.144900 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01bcaf0-e2c1-495b-bc6d-a57978c7817b" containerName="extract-utilities" Mar 13 10:54:00 crc kubenswrapper[4632]: I0313 10:54:00.144914 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01bcaf0-e2c1-495b-bc6d-a57978c7817b" containerName="extract-utilities" Mar 13 10:54:00 crc kubenswrapper[4632]: E0313 10:54:00.144956 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01bcaf0-e2c1-495b-bc6d-a57978c7817b" containerName="registry-server" Mar 13 10:54:00 crc kubenswrapper[4632]: I0313 10:54:00.144962 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01bcaf0-e2c1-495b-bc6d-a57978c7817b" containerName="registry-server" Mar 13 10:54:00 crc kubenswrapper[4632]: E0313 10:54:00.144976 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a01bcaf0-e2c1-495b-bc6d-a57978c7817b" containerName="extract-content" Mar 13 10:54:00 crc kubenswrapper[4632]: I0313 10:54:00.144983 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a01bcaf0-e2c1-495b-bc6d-a57978c7817b" containerName="extract-content" Mar 13 10:54:00 crc kubenswrapper[4632]: I0313 10:54:00.145161 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="a01bcaf0-e2c1-495b-bc6d-a57978c7817b" containerName="registry-server" Mar 13 10:54:00 crc kubenswrapper[4632]: I0313 10:54:00.145729 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556654-htcrm" Mar 13 10:54:00 crc kubenswrapper[4632]: I0313 10:54:00.155805 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:54:00 crc kubenswrapper[4632]: I0313 10:54:00.156009 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:54:00 crc kubenswrapper[4632]: I0313 10:54:00.156193 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:54:00 crc kubenswrapper[4632]: I0313 10:54:00.160781 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556654-htcrm"] Mar 13 10:54:00 crc kubenswrapper[4632]: I0313 10:54:00.190033 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8t4l\" (UniqueName: \"kubernetes.io/projected/ba5d59c2-ece8-4b66-9a10-c3ef740d7e45-kube-api-access-x8t4l\") pod \"auto-csr-approver-29556654-htcrm\" (UID: \"ba5d59c2-ece8-4b66-9a10-c3ef740d7e45\") " pod="openshift-infra/auto-csr-approver-29556654-htcrm" Mar 13 10:54:00 crc kubenswrapper[4632]: I0313 10:54:00.292154 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8t4l\" (UniqueName: \"kubernetes.io/projected/ba5d59c2-ece8-4b66-9a10-c3ef740d7e45-kube-api-access-x8t4l\") pod \"auto-csr-approver-29556654-htcrm\" (UID: \"ba5d59c2-ece8-4b66-9a10-c3ef740d7e45\") " pod="openshift-infra/auto-csr-approver-29556654-htcrm" Mar 13 10:54:00 crc kubenswrapper[4632]: I0313 10:54:00.312174 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8t4l\" (UniqueName: \"kubernetes.io/projected/ba5d59c2-ece8-4b66-9a10-c3ef740d7e45-kube-api-access-x8t4l\") pod \"auto-csr-approver-29556654-htcrm\" (UID: \"ba5d59c2-ece8-4b66-9a10-c3ef740d7e45\") " pod="openshift-infra/auto-csr-approver-29556654-htcrm" Mar 13 10:54:00 crc kubenswrapper[4632]: I0313 10:54:00.472044 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556654-htcrm" Mar 13 10:54:01 crc kubenswrapper[4632]: I0313 10:54:01.060807 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556654-htcrm"] Mar 13 10:54:01 crc kubenswrapper[4632]: I0313 10:54:01.446286 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556654-htcrm" event={"ID":"ba5d59c2-ece8-4b66-9a10-c3ef740d7e45","Type":"ContainerStarted","Data":"56848811c020d11ba44de736b6154580ea88d07320adf5703633c722f7971769"} Mar 13 10:54:03 crc kubenswrapper[4632]: I0313 10:54:03.464900 4632 generic.go:334] "Generic (PLEG): container finished" podID="ba5d59c2-ece8-4b66-9a10-c3ef740d7e45" containerID="0151dac58382ec9dba1fe485dee8519ba248333bc8e6aeae5349b66a4c5fa931" exitCode=0 Mar 13 10:54:03 crc kubenswrapper[4632]: I0313 10:54:03.464991 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556654-htcrm" event={"ID":"ba5d59c2-ece8-4b66-9a10-c3ef740d7e45","Type":"ContainerDied","Data":"0151dac58382ec9dba1fe485dee8519ba248333bc8e6aeae5349b66a4c5fa931"} Mar 13 10:54:04 crc kubenswrapper[4632]: I0313 10:54:04.870354 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556654-htcrm" Mar 13 10:54:04 crc kubenswrapper[4632]: I0313 10:54:04.982005 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8t4l\" (UniqueName: \"kubernetes.io/projected/ba5d59c2-ece8-4b66-9a10-c3ef740d7e45-kube-api-access-x8t4l\") pod \"ba5d59c2-ece8-4b66-9a10-c3ef740d7e45\" (UID: \"ba5d59c2-ece8-4b66-9a10-c3ef740d7e45\") " Mar 13 10:54:04 crc kubenswrapper[4632]: I0313 10:54:04.990699 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba5d59c2-ece8-4b66-9a10-c3ef740d7e45-kube-api-access-x8t4l" (OuterVolumeSpecName: "kube-api-access-x8t4l") pod "ba5d59c2-ece8-4b66-9a10-c3ef740d7e45" (UID: "ba5d59c2-ece8-4b66-9a10-c3ef740d7e45"). InnerVolumeSpecName "kube-api-access-x8t4l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:54:05 crc kubenswrapper[4632]: I0313 10:54:05.084919 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8t4l\" (UniqueName: \"kubernetes.io/projected/ba5d59c2-ece8-4b66-9a10-c3ef740d7e45-kube-api-access-x8t4l\") on node \"crc\" DevicePath \"\"" Mar 13 10:54:05 crc kubenswrapper[4632]: I0313 10:54:05.485044 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556654-htcrm" event={"ID":"ba5d59c2-ece8-4b66-9a10-c3ef740d7e45","Type":"ContainerDied","Data":"56848811c020d11ba44de736b6154580ea88d07320adf5703633c722f7971769"} Mar 13 10:54:05 crc kubenswrapper[4632]: I0313 10:54:05.485376 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56848811c020d11ba44de736b6154580ea88d07320adf5703633c722f7971769" Mar 13 10:54:05 crc kubenswrapper[4632]: I0313 10:54:05.485087 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556654-htcrm" Mar 13 10:54:05 crc kubenswrapper[4632]: I0313 10:54:05.966493 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556648-hbbkx"] Mar 13 10:54:05 crc kubenswrapper[4632]: I0313 10:54:05.973502 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556648-hbbkx"] Mar 13 10:54:06 crc kubenswrapper[4632]: I0313 10:54:06.057036 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ab47075-381b-45d4-b6c8-c64ae6433ef1" path="/var/lib/kubelet/pods/5ab47075-381b-45d4-b6c8-c64ae6433ef1/volumes" Mar 13 10:54:08 crc kubenswrapper[4632]: I0313 10:54:08.518196 4632 generic.go:334] "Generic (PLEG): container finished" podID="4656b24f-4b10-481a-ba5b-1c17e5f2f7ef" containerID="4b7c60398c3f33a538800ee1d366657cf73b7c4a7aff19c20e81ba7f933c624f" exitCode=0 Mar 13 10:54:08 crc kubenswrapper[4632]: I0313 10:54:08.518463 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" event={"ID":"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef","Type":"ContainerDied","Data":"4b7c60398c3f33a538800ee1d366657cf73b7c4a7aff19c20e81ba7f933c624f"} Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.155854 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.238758 4632 scope.go:117] "RemoveContainer" containerID="ac1c75bd040311821d7426607144ebc256c3f11219f7a26012d50c7ce3c315ba" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.295791 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ssh-key-openstack-edpm-ipam\") pod \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.295907 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-1\") pod \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.295973 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-telemetry-combined-ca-bundle\") pod \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.295999 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-2\") pod \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.296049 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sxhtb\" (UniqueName: \"kubernetes.io/projected/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-kube-api-access-sxhtb\") pod \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.296151 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-inventory\") pod \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.296206 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-0\") pod \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\" (UID: \"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef\") " Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.302662 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-kube-api-access-sxhtb" (OuterVolumeSpecName: "kube-api-access-sxhtb") pod "4656b24f-4b10-481a-ba5b-1c17e5f2f7ef" (UID: "4656b24f-4b10-481a-ba5b-1c17e5f2f7ef"). InnerVolumeSpecName "kube-api-access-sxhtb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.303551 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "4656b24f-4b10-481a-ba5b-1c17e5f2f7ef" (UID: "4656b24f-4b10-481a-ba5b-1c17e5f2f7ef"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.330077 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "4656b24f-4b10-481a-ba5b-1c17e5f2f7ef" (UID: "4656b24f-4b10-481a-ba5b-1c17e5f2f7ef"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.330346 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "4656b24f-4b10-481a-ba5b-1c17e5f2f7ef" (UID: "4656b24f-4b10-481a-ba5b-1c17e5f2f7ef"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.333272 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "4656b24f-4b10-481a-ba5b-1c17e5f2f7ef" (UID: "4656b24f-4b10-481a-ba5b-1c17e5f2f7ef"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.340309 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-inventory" (OuterVolumeSpecName: "inventory") pod "4656b24f-4b10-481a-ba5b-1c17e5f2f7ef" (UID: "4656b24f-4b10-481a-ba5b-1c17e5f2f7ef"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.348049 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "4656b24f-4b10-481a-ba5b-1c17e5f2f7ef" (UID: "4656b24f-4b10-481a-ba5b-1c17e5f2f7ef"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.398797 4632 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.398840 4632 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.398859 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.398877 4632 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.398891 4632 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.398903 4632 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.398918 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sxhtb\" (UniqueName: \"kubernetes.io/projected/4656b24f-4b10-481a-ba5b-1c17e5f2f7ef-kube-api-access-sxhtb\") on node \"crc\" DevicePath \"\"" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.461120 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.461196 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.461259 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.462421 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.462528 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" gracePeriod=600 Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.542713 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" event={"ID":"4656b24f-4b10-481a-ba5b-1c17e5f2f7ef","Type":"ContainerDied","Data":"6ee75f7e51730eeb5489b2c1dd9bb7192cbbc507759459f37404c40b386bbdfc"} Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.543022 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ee75f7e51730eeb5489b2c1dd9bb7192cbbc507759459f37404c40b386bbdfc" Mar 13 10:54:10 crc kubenswrapper[4632]: I0313 10:54:10.542770 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw" Mar 13 10:54:10 crc kubenswrapper[4632]: E0313 10:54:10.605504 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:54:11 crc kubenswrapper[4632]: I0313 10:54:11.559032 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" exitCode=0 Mar 13 10:54:11 crc kubenswrapper[4632]: I0313 10:54:11.559088 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d"} Mar 13 10:54:11 crc kubenswrapper[4632]: I0313 10:54:11.559133 4632 scope.go:117] "RemoveContainer" containerID="a5ebd5748d892637db30e6f25b4cdb7397d5f5e2a1d221a622054fbf7f8b83f2" Mar 13 10:54:11 crc kubenswrapper[4632]: I0313 10:54:11.559804 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:54:11 crc kubenswrapper[4632]: E0313 10:54:11.560175 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:54:25 crc kubenswrapper[4632]: I0313 10:54:25.044701 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:54:25 crc kubenswrapper[4632]: E0313 10:54:25.045661 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:54:37 crc kubenswrapper[4632]: I0313 10:54:37.044204 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:54:37 crc kubenswrapper[4632]: E0313 10:54:37.045543 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:54:47 crc kubenswrapper[4632]: E0313 10:54:47.523413 4632 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.182:57266->38.102.83.182:37465: write tcp 38.102.83.182:57266->38.102.83.182:37465: write: broken pipe Mar 13 10:54:51 crc kubenswrapper[4632]: I0313 10:54:51.044715 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:54:51 crc kubenswrapper[4632]: E0313 10:54:51.046973 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:55:04 crc kubenswrapper[4632]: I0313 10:55:04.044706 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:55:04 crc kubenswrapper[4632]: E0313 10:55:04.045598 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:55:08 crc kubenswrapper[4632]: I0313 10:55:08.981923 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s00-multi-thread-testing"] Mar 13 10:55:08 crc kubenswrapper[4632]: E0313 10:55:08.982753 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5d59c2-ece8-4b66-9a10-c3ef740d7e45" containerName="oc" Mar 13 10:55:08 crc kubenswrapper[4632]: I0313 10:55:08.982769 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5d59c2-ece8-4b66-9a10-c3ef740d7e45" containerName="oc" Mar 13 10:55:08 crc kubenswrapper[4632]: E0313 10:55:08.982796 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4656b24f-4b10-481a-ba5b-1c17e5f2f7ef" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 13 10:55:08 crc kubenswrapper[4632]: I0313 10:55:08.982804 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="4656b24f-4b10-481a-ba5b-1c17e5f2f7ef" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 13 10:55:08 crc kubenswrapper[4632]: I0313 10:55:08.983031 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="4656b24f-4b10-481a-ba5b-1c17e5f2f7ef" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 13 10:55:08 crc kubenswrapper[4632]: I0313 10:55:08.983050 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba5d59c2-ece8-4b66-9a10-c3ef740d7e45" containerName="oc" Mar 13 10:55:08 crc kubenswrapper[4632]: I0313 10:55:08.983706 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:08 crc kubenswrapper[4632]: I0313 10:55:08.986672 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Mar 13 10:55:08 crc kubenswrapper[4632]: I0313 10:55:08.987220 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9w9qk" Mar 13 10:55:08 crc kubenswrapper[4632]: I0313 10:55:08.988126 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 13 10:55:08 crc kubenswrapper[4632]: I0313 10:55:08.988482 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.004326 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-multi-thread-testing"] Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.104519 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-openstack-config-secret\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.104587 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a62e0eae-95dd-40a3-a489-80646fde4301-openstack-config\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.104610 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a62e0eae-95dd-40a3-a489-80646fde4301-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.104635 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8w5vp\" (UniqueName: \"kubernetes.io/projected/a62e0eae-95dd-40a3-a489-80646fde4301-kube-api-access-8w5vp\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.105237 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-ssh-key\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.105300 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-ca-certs\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.105323 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.105351 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a62e0eae-95dd-40a3-a489-80646fde4301-config-data\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.105368 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a62e0eae-95dd-40a3-a489-80646fde4301-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.208257 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-ca-certs\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.208326 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.208387 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a62e0eae-95dd-40a3-a489-80646fde4301-config-data\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.208408 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a62e0eae-95dd-40a3-a489-80646fde4301-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.208629 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-openstack-config-secret\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.208659 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a62e0eae-95dd-40a3-a489-80646fde4301-openstack-config\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.208710 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a62e0eae-95dd-40a3-a489-80646fde4301-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.208735 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8w5vp\" (UniqueName: \"kubernetes.io/projected/a62e0eae-95dd-40a3-a489-80646fde4301-kube-api-access-8w5vp\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.208844 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-ssh-key\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.209127 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a62e0eae-95dd-40a3-a489-80646fde4301-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.209259 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a62e0eae-95dd-40a3-a489-80646fde4301-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.211315 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a62e0eae-95dd-40a3-a489-80646fde4301-openstack-config\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.211994 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a62e0eae-95dd-40a3-a489-80646fde4301-config-data\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.212031 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.214519 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-ssh-key\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.214643 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-openstack-config-secret\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.229451 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-ca-certs\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.231634 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8w5vp\" (UniqueName: \"kubernetes.io/projected/a62e0eae-95dd-40a3-a489-80646fde4301-kube-api-access-8w5vp\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.257082 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-multi-thread-testing\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.308473 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 10:55:09 crc kubenswrapper[4632]: I0313 10:55:09.899269 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-multi-thread-testing"] Mar 13 10:55:10 crc kubenswrapper[4632]: I0313 10:55:10.132120 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" event={"ID":"a62e0eae-95dd-40a3-a489-80646fde4301","Type":"ContainerStarted","Data":"959458908fe1f2c8aa4edafce9f9395e573f668491b9554e12daf71db7b5cc6a"} Mar 13 10:55:15 crc kubenswrapper[4632]: I0313 10:55:15.044452 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:55:15 crc kubenswrapper[4632]: E0313 10:55:15.045204 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:55:28 crc kubenswrapper[4632]: I0313 10:55:28.050499 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:55:28 crc kubenswrapper[4632]: E0313 10:55:28.051327 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:55:39 crc kubenswrapper[4632]: I0313 10:55:39.044832 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:55:39 crc kubenswrapper[4632]: E0313 10:55:39.045624 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:55:48 crc kubenswrapper[4632]: E0313 10:55:48.152463 4632 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.132:5001/podified-antelope-centos9/openstack-tempest-all:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:55:48 crc kubenswrapper[4632]: E0313 10:55:48.153453 4632 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.132:5001/podified-antelope-centos9/openstack-tempest-all:e43235cb19da04699a53f42b6a75afe9" Mar 13 10:55:48 crc kubenswrapper[4632]: E0313 10:55:48.156685 4632 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:38.102.83.132:5001/podified-antelope-centos9/openstack-tempest-all:e43235cb19da04699a53f42b6a75afe9,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8w5vp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest-s00-multi-thread-testing_openstack(a62e0eae-95dd-40a3-a489-80646fde4301): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:55:48 crc kubenswrapper[4632]: E0313 10:55:48.157908 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" podUID="a62e0eae-95dd-40a3-a489-80646fde4301" Mar 13 10:55:48 crc kubenswrapper[4632]: E0313 10:55:48.576216 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.132:5001/podified-antelope-centos9/openstack-tempest-all:e43235cb19da04699a53f42b6a75afe9\\\"\"" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" podUID="a62e0eae-95dd-40a3-a489-80646fde4301" Mar 13 10:55:52 crc kubenswrapper[4632]: I0313 10:55:52.044895 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:55:52 crc kubenswrapper[4632]: E0313 10:55:52.045811 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:56:00 crc kubenswrapper[4632]: I0313 10:56:00.165313 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556656-vzb8p"] Mar 13 10:56:00 crc kubenswrapper[4632]: I0313 10:56:00.169904 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556656-vzb8p" Mar 13 10:56:00 crc kubenswrapper[4632]: I0313 10:56:00.172601 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:56:00 crc kubenswrapper[4632]: I0313 10:56:00.174108 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:56:00 crc kubenswrapper[4632]: I0313 10:56:00.174187 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:56:00 crc kubenswrapper[4632]: I0313 10:56:00.186672 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556656-vzb8p"] Mar 13 10:56:00 crc kubenswrapper[4632]: I0313 10:56:00.334744 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26vmr\" (UniqueName: \"kubernetes.io/projected/f70e2037-a5d6-4479-af7f-18fe8ff9e952-kube-api-access-26vmr\") pod \"auto-csr-approver-29556656-vzb8p\" (UID: \"f70e2037-a5d6-4479-af7f-18fe8ff9e952\") " pod="openshift-infra/auto-csr-approver-29556656-vzb8p" Mar 13 10:56:00 crc kubenswrapper[4632]: I0313 10:56:00.436886 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26vmr\" (UniqueName: \"kubernetes.io/projected/f70e2037-a5d6-4479-af7f-18fe8ff9e952-kube-api-access-26vmr\") pod \"auto-csr-approver-29556656-vzb8p\" (UID: \"f70e2037-a5d6-4479-af7f-18fe8ff9e952\") " pod="openshift-infra/auto-csr-approver-29556656-vzb8p" Mar 13 10:56:00 crc kubenswrapper[4632]: I0313 10:56:00.464435 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26vmr\" (UniqueName: \"kubernetes.io/projected/f70e2037-a5d6-4479-af7f-18fe8ff9e952-kube-api-access-26vmr\") pod \"auto-csr-approver-29556656-vzb8p\" (UID: \"f70e2037-a5d6-4479-af7f-18fe8ff9e952\") " pod="openshift-infra/auto-csr-approver-29556656-vzb8p" Mar 13 10:56:00 crc kubenswrapper[4632]: I0313 10:56:00.499673 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556656-vzb8p" Mar 13 10:56:01 crc kubenswrapper[4632]: I0313 10:56:01.064067 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556656-vzb8p"] Mar 13 10:56:01 crc kubenswrapper[4632]: I0313 10:56:01.707224 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556656-vzb8p" event={"ID":"f70e2037-a5d6-4479-af7f-18fe8ff9e952","Type":"ContainerStarted","Data":"461a346934711ed3bdfd4aa55ae2ec81a65d1c4197e9cf719d8b6ec477df66ce"} Mar 13 10:56:02 crc kubenswrapper[4632]: I0313 10:56:02.716603 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556656-vzb8p" event={"ID":"f70e2037-a5d6-4479-af7f-18fe8ff9e952","Type":"ContainerStarted","Data":"6dd075c6962fa13da67ea22e1c7e0f24f4fdd06a675abd3b301b6ea671a2f51e"} Mar 13 10:56:02 crc kubenswrapper[4632]: I0313 10:56:02.737331 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556656-vzb8p" podStartSLOduration=1.809424227 podStartE2EDuration="2.737312503s" podCreationTimestamp="2026-03-13 10:56:00 +0000 UTC" firstStartedPulling="2026-03-13 10:56:01.07056543 +0000 UTC m=+3135.093095563" lastFinishedPulling="2026-03-13 10:56:01.998453716 +0000 UTC m=+3136.020983839" observedRunningTime="2026-03-13 10:56:02.730576237 +0000 UTC m=+3136.753106390" watchObservedRunningTime="2026-03-13 10:56:02.737312503 +0000 UTC m=+3136.759842626" Mar 13 10:56:03 crc kubenswrapper[4632]: I0313 10:56:03.729259 4632 generic.go:334] "Generic (PLEG): container finished" podID="f70e2037-a5d6-4479-af7f-18fe8ff9e952" containerID="6dd075c6962fa13da67ea22e1c7e0f24f4fdd06a675abd3b301b6ea671a2f51e" exitCode=0 Mar 13 10:56:03 crc kubenswrapper[4632]: I0313 10:56:03.729330 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556656-vzb8p" event={"ID":"f70e2037-a5d6-4479-af7f-18fe8ff9e952","Type":"ContainerDied","Data":"6dd075c6962fa13da67ea22e1c7e0f24f4fdd06a675abd3b301b6ea671a2f51e"} Mar 13 10:56:04 crc kubenswrapper[4632]: I0313 10:56:04.115600 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 13 10:56:05 crc kubenswrapper[4632]: I0313 10:56:05.124321 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556656-vzb8p" Mar 13 10:56:05 crc kubenswrapper[4632]: I0313 10:56:05.249121 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-26vmr\" (UniqueName: \"kubernetes.io/projected/f70e2037-a5d6-4479-af7f-18fe8ff9e952-kube-api-access-26vmr\") pod \"f70e2037-a5d6-4479-af7f-18fe8ff9e952\" (UID: \"f70e2037-a5d6-4479-af7f-18fe8ff9e952\") " Mar 13 10:56:05 crc kubenswrapper[4632]: I0313 10:56:05.256172 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f70e2037-a5d6-4479-af7f-18fe8ff9e952-kube-api-access-26vmr" (OuterVolumeSpecName: "kube-api-access-26vmr") pod "f70e2037-a5d6-4479-af7f-18fe8ff9e952" (UID: "f70e2037-a5d6-4479-af7f-18fe8ff9e952"). InnerVolumeSpecName "kube-api-access-26vmr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:56:05 crc kubenswrapper[4632]: I0313 10:56:05.351770 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-26vmr\" (UniqueName: \"kubernetes.io/projected/f70e2037-a5d6-4479-af7f-18fe8ff9e952-kube-api-access-26vmr\") on node \"crc\" DevicePath \"\"" Mar 13 10:56:05 crc kubenswrapper[4632]: I0313 10:56:05.749840 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556656-vzb8p" event={"ID":"f70e2037-a5d6-4479-af7f-18fe8ff9e952","Type":"ContainerDied","Data":"461a346934711ed3bdfd4aa55ae2ec81a65d1c4197e9cf719d8b6ec477df66ce"} Mar 13 10:56:05 crc kubenswrapper[4632]: I0313 10:56:05.750147 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="461a346934711ed3bdfd4aa55ae2ec81a65d1c4197e9cf719d8b6ec477df66ce" Mar 13 10:56:05 crc kubenswrapper[4632]: I0313 10:56:05.750073 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556656-vzb8p" Mar 13 10:56:05 crc kubenswrapper[4632]: I0313 10:56:05.752467 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" event={"ID":"a62e0eae-95dd-40a3-a489-80646fde4301","Type":"ContainerStarted","Data":"ec15c016ac8280363b8fb347025993466f5b7492f2d0ac470ef8fc423974c0e2"} Mar 13 10:56:05 crc kubenswrapper[4632]: I0313 10:56:05.788026 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" podStartSLOduration=4.587990695 podStartE2EDuration="58.787996943s" podCreationTimestamp="2026-03-13 10:55:07 +0000 UTC" firstStartedPulling="2026-03-13 10:55:09.913419149 +0000 UTC m=+3083.935949282" lastFinishedPulling="2026-03-13 10:56:04.113425397 +0000 UTC m=+3138.135955530" observedRunningTime="2026-03-13 10:56:05.775993357 +0000 UTC m=+3139.798523500" watchObservedRunningTime="2026-03-13 10:56:05.787996943 +0000 UTC m=+3139.810527096" Mar 13 10:56:05 crc kubenswrapper[4632]: I0313 10:56:05.816928 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556650-rskjc"] Mar 13 10:56:05 crc kubenswrapper[4632]: I0313 10:56:05.824956 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556650-rskjc"] Mar 13 10:56:06 crc kubenswrapper[4632]: I0313 10:56:06.044888 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:56:06 crc kubenswrapper[4632]: E0313 10:56:06.045191 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:56:06 crc kubenswrapper[4632]: I0313 10:56:06.056508 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc8dd4ae-e21e-4155-b617-19c85512d4fe" path="/var/lib/kubelet/pods/cc8dd4ae-e21e-4155-b617-19c85512d4fe/volumes" Mar 13 10:56:10 crc kubenswrapper[4632]: I0313 10:56:10.380595 4632 scope.go:117] "RemoveContainer" containerID="10fbcfced1ba7ba66a4ba615aa3b2aab72091e177631977720b60aac13ae9d0f" Mar 13 10:56:17 crc kubenswrapper[4632]: I0313 10:56:17.044200 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:56:17 crc kubenswrapper[4632]: E0313 10:56:17.045223 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:56:31 crc kubenswrapper[4632]: I0313 10:56:31.045629 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:56:31 crc kubenswrapper[4632]: E0313 10:56:31.046423 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:56:45 crc kubenswrapper[4632]: I0313 10:56:45.045908 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:56:45 crc kubenswrapper[4632]: E0313 10:56:45.047060 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:56:58 crc kubenswrapper[4632]: I0313 10:56:58.050354 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:56:58 crc kubenswrapper[4632]: E0313 10:56:58.051413 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:57:09 crc kubenswrapper[4632]: I0313 10:57:09.044137 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:57:09 crc kubenswrapper[4632]: E0313 10:57:09.045806 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:57:24 crc kubenswrapper[4632]: I0313 10:57:24.044226 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:57:24 crc kubenswrapper[4632]: E0313 10:57:24.044920 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:57:35 crc kubenswrapper[4632]: I0313 10:57:35.044492 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:57:35 crc kubenswrapper[4632]: E0313 10:57:35.045320 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:57:47 crc kubenswrapper[4632]: I0313 10:57:47.044518 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:57:47 crc kubenswrapper[4632]: E0313 10:57:47.045675 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.535540 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fmml7"] Mar 13 10:57:54 crc kubenswrapper[4632]: E0313 10:57:54.551066 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f70e2037-a5d6-4479-af7f-18fe8ff9e952" containerName="oc" Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.551895 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f70e2037-a5d6-4479-af7f-18fe8ff9e952" containerName="oc" Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.556673 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f70e2037-a5d6-4479-af7f-18fe8ff9e952" containerName="oc" Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.560429 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.664881 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c91e27-6596-4471-81e9-4a65e55379cc-utilities\") pod \"certified-operators-fmml7\" (UID: \"d7c91e27-6596-4471-81e9-4a65e55379cc\") " pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.665210 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c91e27-6596-4471-81e9-4a65e55379cc-catalog-content\") pod \"certified-operators-fmml7\" (UID: \"d7c91e27-6596-4471-81e9-4a65e55379cc\") " pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.665347 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzvbq\" (UniqueName: \"kubernetes.io/projected/d7c91e27-6596-4471-81e9-4a65e55379cc-kube-api-access-hzvbq\") pod \"certified-operators-fmml7\" (UID: \"d7c91e27-6596-4471-81e9-4a65e55379cc\") " pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.699096 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fmml7"] Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.767614 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c91e27-6596-4471-81e9-4a65e55379cc-utilities\") pod \"certified-operators-fmml7\" (UID: \"d7c91e27-6596-4471-81e9-4a65e55379cc\") " pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.767887 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c91e27-6596-4471-81e9-4a65e55379cc-catalog-content\") pod \"certified-operators-fmml7\" (UID: \"d7c91e27-6596-4471-81e9-4a65e55379cc\") " pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.767926 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzvbq\" (UniqueName: \"kubernetes.io/projected/d7c91e27-6596-4471-81e9-4a65e55379cc-kube-api-access-hzvbq\") pod \"certified-operators-fmml7\" (UID: \"d7c91e27-6596-4471-81e9-4a65e55379cc\") " pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.774245 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c91e27-6596-4471-81e9-4a65e55379cc-utilities\") pod \"certified-operators-fmml7\" (UID: \"d7c91e27-6596-4471-81e9-4a65e55379cc\") " pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.776025 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c91e27-6596-4471-81e9-4a65e55379cc-catalog-content\") pod \"certified-operators-fmml7\" (UID: \"d7c91e27-6596-4471-81e9-4a65e55379cc\") " pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.800043 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzvbq\" (UniqueName: \"kubernetes.io/projected/d7c91e27-6596-4471-81e9-4a65e55379cc-kube-api-access-hzvbq\") pod \"certified-operators-fmml7\" (UID: \"d7c91e27-6596-4471-81e9-4a65e55379cc\") " pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:57:54 crc kubenswrapper[4632]: I0313 10:57:54.890858 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:57:56 crc kubenswrapper[4632]: I0313 10:57:56.073402 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fmml7"] Mar 13 10:57:56 crc kubenswrapper[4632]: I0313 10:57:56.814466 4632 generic.go:334] "Generic (PLEG): container finished" podID="d7c91e27-6596-4471-81e9-4a65e55379cc" containerID="145a4083a5a5cb7cab6b80318c7e624a2791db3949fe0566ce5173ba6c6e5bc8" exitCode=0 Mar 13 10:57:56 crc kubenswrapper[4632]: I0313 10:57:56.814580 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fmml7" event={"ID":"d7c91e27-6596-4471-81e9-4a65e55379cc","Type":"ContainerDied","Data":"145a4083a5a5cb7cab6b80318c7e624a2791db3949fe0566ce5173ba6c6e5bc8"} Mar 13 10:57:56 crc kubenswrapper[4632]: I0313 10:57:56.814838 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fmml7" event={"ID":"d7c91e27-6596-4471-81e9-4a65e55379cc","Type":"ContainerStarted","Data":"4dd1def3762efd7a2c7321b10db0c95910e7988e1bb090633291fc5c968853b7"} Mar 13 10:57:56 crc kubenswrapper[4632]: I0313 10:57:56.822255 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:57:57 crc kubenswrapper[4632]: I0313 10:57:57.824576 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fmml7" event={"ID":"d7c91e27-6596-4471-81e9-4a65e55379cc","Type":"ContainerStarted","Data":"8d0330408b68a3dc8f003bcb5a971e2a87e41ec5aadc0fa59d29938015321612"} Mar 13 10:57:59 crc kubenswrapper[4632]: I0313 10:57:59.044568 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:57:59 crc kubenswrapper[4632]: E0313 10:57:59.045110 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:58:00 crc kubenswrapper[4632]: I0313 10:58:00.207091 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556658-mbpfd"] Mar 13 10:58:00 crc kubenswrapper[4632]: I0313 10:58:00.208478 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556658-mbpfd" Mar 13 10:58:00 crc kubenswrapper[4632]: I0313 10:58:00.213524 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:58:00 crc kubenswrapper[4632]: I0313 10:58:00.213907 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:58:00 crc kubenswrapper[4632]: I0313 10:58:00.214509 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 10:58:00 crc kubenswrapper[4632]: I0313 10:58:00.220060 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556658-mbpfd"] Mar 13 10:58:00 crc kubenswrapper[4632]: I0313 10:58:00.277271 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-984mt\" (UniqueName: \"kubernetes.io/projected/4db08ac3-f768-407d-a321-ed9032c5c015-kube-api-access-984mt\") pod \"auto-csr-approver-29556658-mbpfd\" (UID: \"4db08ac3-f768-407d-a321-ed9032c5c015\") " pod="openshift-infra/auto-csr-approver-29556658-mbpfd" Mar 13 10:58:00 crc kubenswrapper[4632]: I0313 10:58:00.379085 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-984mt\" (UniqueName: \"kubernetes.io/projected/4db08ac3-f768-407d-a321-ed9032c5c015-kube-api-access-984mt\") pod \"auto-csr-approver-29556658-mbpfd\" (UID: \"4db08ac3-f768-407d-a321-ed9032c5c015\") " pod="openshift-infra/auto-csr-approver-29556658-mbpfd" Mar 13 10:58:00 crc kubenswrapper[4632]: I0313 10:58:00.406813 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-984mt\" (UniqueName: \"kubernetes.io/projected/4db08ac3-f768-407d-a321-ed9032c5c015-kube-api-access-984mt\") pod \"auto-csr-approver-29556658-mbpfd\" (UID: \"4db08ac3-f768-407d-a321-ed9032c5c015\") " pod="openshift-infra/auto-csr-approver-29556658-mbpfd" Mar 13 10:58:00 crc kubenswrapper[4632]: I0313 10:58:00.564490 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556658-mbpfd" Mar 13 10:58:00 crc kubenswrapper[4632]: I0313 10:58:00.876894 4632 generic.go:334] "Generic (PLEG): container finished" podID="d7c91e27-6596-4471-81e9-4a65e55379cc" containerID="8d0330408b68a3dc8f003bcb5a971e2a87e41ec5aadc0fa59d29938015321612" exitCode=0 Mar 13 10:58:00 crc kubenswrapper[4632]: I0313 10:58:00.877048 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fmml7" event={"ID":"d7c91e27-6596-4471-81e9-4a65e55379cc","Type":"ContainerDied","Data":"8d0330408b68a3dc8f003bcb5a971e2a87e41ec5aadc0fa59d29938015321612"} Mar 13 10:58:01 crc kubenswrapper[4632]: I0313 10:58:01.152842 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556658-mbpfd"] Mar 13 10:58:01 crc kubenswrapper[4632]: I0313 10:58:01.900854 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fmml7" event={"ID":"d7c91e27-6596-4471-81e9-4a65e55379cc","Type":"ContainerStarted","Data":"d9aa3193c9ba513408053eae5959957c7f7db68e45389eba91fd0f384d6e744b"} Mar 13 10:58:01 crc kubenswrapper[4632]: I0313 10:58:01.902910 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556658-mbpfd" event={"ID":"4db08ac3-f768-407d-a321-ed9032c5c015","Type":"ContainerStarted","Data":"1ad1e2f33bb0810479cdd95e85cc645315ff0eeeea2f7b9d25a8d8a7b46bcceb"} Mar 13 10:58:01 crc kubenswrapper[4632]: I0313 10:58:01.969879 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fmml7" podStartSLOduration=3.414099043 podStartE2EDuration="7.969846258s" podCreationTimestamp="2026-03-13 10:57:54 +0000 UTC" firstStartedPulling="2026-03-13 10:57:56.816925341 +0000 UTC m=+3250.839455474" lastFinishedPulling="2026-03-13 10:58:01.372672556 +0000 UTC m=+3255.395202689" observedRunningTime="2026-03-13 10:58:01.930731526 +0000 UTC m=+3255.953261669" watchObservedRunningTime="2026-03-13 10:58:01.969846258 +0000 UTC m=+3255.992376391" Mar 13 10:58:03 crc kubenswrapper[4632]: I0313 10:58:03.926826 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556658-mbpfd" event={"ID":"4db08ac3-f768-407d-a321-ed9032c5c015","Type":"ContainerStarted","Data":"6a752c085ec4dd2121b36385f753ab45221d95dd428ca910155d9e3c585e4dbc"} Mar 13 10:58:04 crc kubenswrapper[4632]: I0313 10:58:04.892347 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:58:04 crc kubenswrapper[4632]: I0313 10:58:04.892844 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:58:04 crc kubenswrapper[4632]: I0313 10:58:04.939113 4632 generic.go:334] "Generic (PLEG): container finished" podID="4db08ac3-f768-407d-a321-ed9032c5c015" containerID="6a752c085ec4dd2121b36385f753ab45221d95dd428ca910155d9e3c585e4dbc" exitCode=0 Mar 13 10:58:04 crc kubenswrapper[4632]: I0313 10:58:04.941335 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556658-mbpfd" event={"ID":"4db08ac3-f768-407d-a321-ed9032c5c015","Type":"ContainerDied","Data":"6a752c085ec4dd2121b36385f753ab45221d95dd428ca910155d9e3c585e4dbc"} Mar 13 10:58:05 crc kubenswrapper[4632]: I0313 10:58:05.948375 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-fmml7" podUID="d7c91e27-6596-4471-81e9-4a65e55379cc" containerName="registry-server" probeResult="failure" output=< Mar 13 10:58:05 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:58:05 crc kubenswrapper[4632]: > Mar 13 10:58:06 crc kubenswrapper[4632]: I0313 10:58:06.566406 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556658-mbpfd" Mar 13 10:58:06 crc kubenswrapper[4632]: I0313 10:58:06.639361 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-984mt\" (UniqueName: \"kubernetes.io/projected/4db08ac3-f768-407d-a321-ed9032c5c015-kube-api-access-984mt\") pod \"4db08ac3-f768-407d-a321-ed9032c5c015\" (UID: \"4db08ac3-f768-407d-a321-ed9032c5c015\") " Mar 13 10:58:06 crc kubenswrapper[4632]: I0313 10:58:06.663166 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4db08ac3-f768-407d-a321-ed9032c5c015-kube-api-access-984mt" (OuterVolumeSpecName: "kube-api-access-984mt") pod "4db08ac3-f768-407d-a321-ed9032c5c015" (UID: "4db08ac3-f768-407d-a321-ed9032c5c015"). InnerVolumeSpecName "kube-api-access-984mt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:58:06 crc kubenswrapper[4632]: I0313 10:58:06.743923 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-984mt\" (UniqueName: \"kubernetes.io/projected/4db08ac3-f768-407d-a321-ed9032c5c015-kube-api-access-984mt\") on node \"crc\" DevicePath \"\"" Mar 13 10:58:06 crc kubenswrapper[4632]: I0313 10:58:06.959506 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556658-mbpfd" event={"ID":"4db08ac3-f768-407d-a321-ed9032c5c015","Type":"ContainerDied","Data":"1ad1e2f33bb0810479cdd95e85cc645315ff0eeeea2f7b9d25a8d8a7b46bcceb"} Mar 13 10:58:06 crc kubenswrapper[4632]: I0313 10:58:06.959548 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1ad1e2f33bb0810479cdd95e85cc645315ff0eeeea2f7b9d25a8d8a7b46bcceb" Mar 13 10:58:06 crc kubenswrapper[4632]: I0313 10:58:06.959596 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556658-mbpfd" Mar 13 10:58:07 crc kubenswrapper[4632]: I0313 10:58:07.095769 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556652-ghswk"] Mar 13 10:58:07 crc kubenswrapper[4632]: I0313 10:58:07.108682 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556652-ghswk"] Mar 13 10:58:08 crc kubenswrapper[4632]: I0313 10:58:08.063035 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27a43bdf-8be4-458b-99ff-4135a684962a" path="/var/lib/kubelet/pods/27a43bdf-8be4-458b-99ff-4135a684962a/volumes" Mar 13 10:58:10 crc kubenswrapper[4632]: I0313 10:58:10.045138 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:58:10 crc kubenswrapper[4632]: E0313 10:58:10.049139 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:58:10 crc kubenswrapper[4632]: I0313 10:58:10.501875 4632 scope.go:117] "RemoveContainer" containerID="5c30aacc6e4fc680fbb7e668912e806d2164d1cac4b7e0e4c5e8b4688d3e76cd" Mar 13 10:58:12 crc kubenswrapper[4632]: I0313 10:58:12.968807 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q9g65"] Mar 13 10:58:12 crc kubenswrapper[4632]: E0313 10:58:12.969496 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4db08ac3-f768-407d-a321-ed9032c5c015" containerName="oc" Mar 13 10:58:12 crc kubenswrapper[4632]: I0313 10:58:12.969510 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="4db08ac3-f768-407d-a321-ed9032c5c015" containerName="oc" Mar 13 10:58:12 crc kubenswrapper[4632]: I0313 10:58:12.969721 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="4db08ac3-f768-407d-a321-ed9032c5c015" containerName="oc" Mar 13 10:58:12 crc kubenswrapper[4632]: I0313 10:58:12.971102 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:13 crc kubenswrapper[4632]: I0313 10:58:13.011331 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9g65"] Mar 13 10:58:13 crc kubenswrapper[4632]: I0313 10:58:13.090258 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d9de0e-efea-443b-89a7-e02d3264020f-utilities\") pod \"redhat-marketplace-q9g65\" (UID: \"48d9de0e-efea-443b-89a7-e02d3264020f\") " pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:13 crc kubenswrapper[4632]: I0313 10:58:13.090351 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-459zd\" (UniqueName: \"kubernetes.io/projected/48d9de0e-efea-443b-89a7-e02d3264020f-kube-api-access-459zd\") pod \"redhat-marketplace-q9g65\" (UID: \"48d9de0e-efea-443b-89a7-e02d3264020f\") " pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:13 crc kubenswrapper[4632]: I0313 10:58:13.090391 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d9de0e-efea-443b-89a7-e02d3264020f-catalog-content\") pod \"redhat-marketplace-q9g65\" (UID: \"48d9de0e-efea-443b-89a7-e02d3264020f\") " pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:13 crc kubenswrapper[4632]: I0313 10:58:13.192843 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d9de0e-efea-443b-89a7-e02d3264020f-utilities\") pod \"redhat-marketplace-q9g65\" (UID: \"48d9de0e-efea-443b-89a7-e02d3264020f\") " pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:13 crc kubenswrapper[4632]: I0313 10:58:13.192995 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-459zd\" (UniqueName: \"kubernetes.io/projected/48d9de0e-efea-443b-89a7-e02d3264020f-kube-api-access-459zd\") pod \"redhat-marketplace-q9g65\" (UID: \"48d9de0e-efea-443b-89a7-e02d3264020f\") " pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:13 crc kubenswrapper[4632]: I0313 10:58:13.193069 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d9de0e-efea-443b-89a7-e02d3264020f-catalog-content\") pod \"redhat-marketplace-q9g65\" (UID: \"48d9de0e-efea-443b-89a7-e02d3264020f\") " pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:13 crc kubenswrapper[4632]: I0313 10:58:13.193616 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d9de0e-efea-443b-89a7-e02d3264020f-catalog-content\") pod \"redhat-marketplace-q9g65\" (UID: \"48d9de0e-efea-443b-89a7-e02d3264020f\") " pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:13 crc kubenswrapper[4632]: I0313 10:58:13.193633 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d9de0e-efea-443b-89a7-e02d3264020f-utilities\") pod \"redhat-marketplace-q9g65\" (UID: \"48d9de0e-efea-443b-89a7-e02d3264020f\") " pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:13 crc kubenswrapper[4632]: I0313 10:58:13.223304 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-459zd\" (UniqueName: \"kubernetes.io/projected/48d9de0e-efea-443b-89a7-e02d3264020f-kube-api-access-459zd\") pod \"redhat-marketplace-q9g65\" (UID: \"48d9de0e-efea-443b-89a7-e02d3264020f\") " pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:13 crc kubenswrapper[4632]: I0313 10:58:13.296634 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:14 crc kubenswrapper[4632]: I0313 10:58:14.070091 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9g65"] Mar 13 10:58:15 crc kubenswrapper[4632]: I0313 10:58:15.064678 4632 generic.go:334] "Generic (PLEG): container finished" podID="48d9de0e-efea-443b-89a7-e02d3264020f" containerID="89e85aed0ace971ea27aa6e4381a5b9c378475b156c657d5f7f64cf10fffb7e7" exitCode=0 Mar 13 10:58:15 crc kubenswrapper[4632]: I0313 10:58:15.064760 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9g65" event={"ID":"48d9de0e-efea-443b-89a7-e02d3264020f","Type":"ContainerDied","Data":"89e85aed0ace971ea27aa6e4381a5b9c378475b156c657d5f7f64cf10fffb7e7"} Mar 13 10:58:15 crc kubenswrapper[4632]: I0313 10:58:15.064961 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9g65" event={"ID":"48d9de0e-efea-443b-89a7-e02d3264020f","Type":"ContainerStarted","Data":"791d787bc94f7eeb2e183a040f421e97af028bfea047785789f3496908b775af"} Mar 13 10:58:15 crc kubenswrapper[4632]: I0313 10:58:15.952076 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-fmml7" podUID="d7c91e27-6596-4471-81e9-4a65e55379cc" containerName="registry-server" probeResult="failure" output=< Mar 13 10:58:15 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:58:15 crc kubenswrapper[4632]: > Mar 13 10:58:16 crc kubenswrapper[4632]: I0313 10:58:16.086044 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9g65" event={"ID":"48d9de0e-efea-443b-89a7-e02d3264020f","Type":"ContainerStarted","Data":"e3f2fccdad049285eb344aa03ec4c45ce2e653e042872421419872537e65b0ce"} Mar 13 10:58:18 crc kubenswrapper[4632]: I0313 10:58:18.104682 4632 generic.go:334] "Generic (PLEG): container finished" podID="48d9de0e-efea-443b-89a7-e02d3264020f" containerID="e3f2fccdad049285eb344aa03ec4c45ce2e653e042872421419872537e65b0ce" exitCode=0 Mar 13 10:58:18 crc kubenswrapper[4632]: I0313 10:58:18.104746 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9g65" event={"ID":"48d9de0e-efea-443b-89a7-e02d3264020f","Type":"ContainerDied","Data":"e3f2fccdad049285eb344aa03ec4c45ce2e653e042872421419872537e65b0ce"} Mar 13 10:58:20 crc kubenswrapper[4632]: I0313 10:58:20.126988 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9g65" event={"ID":"48d9de0e-efea-443b-89a7-e02d3264020f","Type":"ContainerStarted","Data":"a35447b594c379c0f9e9337a36f14fcb40355b64f78e7eff5cb0f3ce866933cc"} Mar 13 10:58:20 crc kubenswrapper[4632]: I0313 10:58:20.152932 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q9g65" podStartSLOduration=4.423860649 podStartE2EDuration="8.152907296s" podCreationTimestamp="2026-03-13 10:58:12 +0000 UTC" firstStartedPulling="2026-03-13 10:58:15.067169904 +0000 UTC m=+3269.089700037" lastFinishedPulling="2026-03-13 10:58:18.796216551 +0000 UTC m=+3272.818746684" observedRunningTime="2026-03-13 10:58:20.146087079 +0000 UTC m=+3274.168617212" watchObservedRunningTime="2026-03-13 10:58:20.152907296 +0000 UTC m=+3274.175437439" Mar 13 10:58:21 crc kubenswrapper[4632]: I0313 10:58:21.045206 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:58:21 crc kubenswrapper[4632]: E0313 10:58:21.045590 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:58:23 crc kubenswrapper[4632]: I0313 10:58:23.296723 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:23 crc kubenswrapper[4632]: I0313 10:58:23.298204 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:24 crc kubenswrapper[4632]: I0313 10:58:24.498462 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-q9g65" podUID="48d9de0e-efea-443b-89a7-e02d3264020f" containerName="registry-server" probeResult="failure" output=< Mar 13 10:58:24 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 10:58:24 crc kubenswrapper[4632]: > Mar 13 10:58:24 crc kubenswrapper[4632]: I0313 10:58:24.948306 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:58:25 crc kubenswrapper[4632]: I0313 10:58:25.004848 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:58:25 crc kubenswrapper[4632]: I0313 10:58:25.740099 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fmml7"] Mar 13 10:58:26 crc kubenswrapper[4632]: I0313 10:58:26.174185 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fmml7" podUID="d7c91e27-6596-4471-81e9-4a65e55379cc" containerName="registry-server" containerID="cri-o://d9aa3193c9ba513408053eae5959957c7f7db68e45389eba91fd0f384d6e744b" gracePeriod=2 Mar 13 10:58:27 crc kubenswrapper[4632]: I0313 10:58:27.195652 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fmml7" event={"ID":"d7c91e27-6596-4471-81e9-4a65e55379cc","Type":"ContainerDied","Data":"d9aa3193c9ba513408053eae5959957c7f7db68e45389eba91fd0f384d6e744b"} Mar 13 10:58:27 crc kubenswrapper[4632]: I0313 10:58:27.195492 4632 generic.go:334] "Generic (PLEG): container finished" podID="d7c91e27-6596-4471-81e9-4a65e55379cc" containerID="d9aa3193c9ba513408053eae5959957c7f7db68e45389eba91fd0f384d6e744b" exitCode=0 Mar 13 10:58:27 crc kubenswrapper[4632]: I0313 10:58:27.404186 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:58:27 crc kubenswrapper[4632]: I0313 10:58:27.605114 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c91e27-6596-4471-81e9-4a65e55379cc-utilities\") pod \"d7c91e27-6596-4471-81e9-4a65e55379cc\" (UID: \"d7c91e27-6596-4471-81e9-4a65e55379cc\") " Mar 13 10:58:27 crc kubenswrapper[4632]: I0313 10:58:27.605201 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c91e27-6596-4471-81e9-4a65e55379cc-catalog-content\") pod \"d7c91e27-6596-4471-81e9-4a65e55379cc\" (UID: \"d7c91e27-6596-4471-81e9-4a65e55379cc\") " Mar 13 10:58:27 crc kubenswrapper[4632]: I0313 10:58:27.605243 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzvbq\" (UniqueName: \"kubernetes.io/projected/d7c91e27-6596-4471-81e9-4a65e55379cc-kube-api-access-hzvbq\") pod \"d7c91e27-6596-4471-81e9-4a65e55379cc\" (UID: \"d7c91e27-6596-4471-81e9-4a65e55379cc\") " Mar 13 10:58:27 crc kubenswrapper[4632]: I0313 10:58:27.609572 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7c91e27-6596-4471-81e9-4a65e55379cc-utilities" (OuterVolumeSpecName: "utilities") pod "d7c91e27-6596-4471-81e9-4a65e55379cc" (UID: "d7c91e27-6596-4471-81e9-4a65e55379cc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:58:27 crc kubenswrapper[4632]: I0313 10:58:27.652409 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7c91e27-6596-4471-81e9-4a65e55379cc-kube-api-access-hzvbq" (OuterVolumeSpecName: "kube-api-access-hzvbq") pod "d7c91e27-6596-4471-81e9-4a65e55379cc" (UID: "d7c91e27-6596-4471-81e9-4a65e55379cc"). InnerVolumeSpecName "kube-api-access-hzvbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:58:27 crc kubenswrapper[4632]: I0313 10:58:27.710081 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d7c91e27-6596-4471-81e9-4a65e55379cc-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:58:27 crc kubenswrapper[4632]: I0313 10:58:27.710127 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzvbq\" (UniqueName: \"kubernetes.io/projected/d7c91e27-6596-4471-81e9-4a65e55379cc-kube-api-access-hzvbq\") on node \"crc\" DevicePath \"\"" Mar 13 10:58:27 crc kubenswrapper[4632]: I0313 10:58:27.876387 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d7c91e27-6596-4471-81e9-4a65e55379cc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d7c91e27-6596-4471-81e9-4a65e55379cc" (UID: "d7c91e27-6596-4471-81e9-4a65e55379cc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:58:27 crc kubenswrapper[4632]: I0313 10:58:27.921024 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d7c91e27-6596-4471-81e9-4a65e55379cc-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:58:28 crc kubenswrapper[4632]: I0313 10:58:28.212643 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fmml7" event={"ID":"d7c91e27-6596-4471-81e9-4a65e55379cc","Type":"ContainerDied","Data":"4dd1def3762efd7a2c7321b10db0c95910e7988e1bb090633291fc5c968853b7"} Mar 13 10:58:28 crc kubenswrapper[4632]: I0313 10:58:28.212708 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fmml7" Mar 13 10:58:28 crc kubenswrapper[4632]: I0313 10:58:28.212739 4632 scope.go:117] "RemoveContainer" containerID="d9aa3193c9ba513408053eae5959957c7f7db68e45389eba91fd0f384d6e744b" Mar 13 10:58:28 crc kubenswrapper[4632]: I0313 10:58:28.247987 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fmml7"] Mar 13 10:58:28 crc kubenswrapper[4632]: I0313 10:58:28.258080 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fmml7"] Mar 13 10:58:28 crc kubenswrapper[4632]: I0313 10:58:28.266592 4632 scope.go:117] "RemoveContainer" containerID="8d0330408b68a3dc8f003bcb5a971e2a87e41ec5aadc0fa59d29938015321612" Mar 13 10:58:28 crc kubenswrapper[4632]: I0313 10:58:28.304283 4632 scope.go:117] "RemoveContainer" containerID="145a4083a5a5cb7cab6b80318c7e624a2791db3949fe0566ce5173ba6c6e5bc8" Mar 13 10:58:30 crc kubenswrapper[4632]: I0313 10:58:30.054621 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7c91e27-6596-4471-81e9-4a65e55379cc" path="/var/lib/kubelet/pods/d7c91e27-6596-4471-81e9-4a65e55379cc/volumes" Mar 13 10:58:33 crc kubenswrapper[4632]: I0313 10:58:33.373883 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:33 crc kubenswrapper[4632]: I0313 10:58:33.430882 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:33 crc kubenswrapper[4632]: I0313 10:58:33.619414 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9g65"] Mar 13 10:58:35 crc kubenswrapper[4632]: I0313 10:58:35.288041 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q9g65" podUID="48d9de0e-efea-443b-89a7-e02d3264020f" containerName="registry-server" containerID="cri-o://a35447b594c379c0f9e9337a36f14fcb40355b64f78e7eff5cb0f3ce866933cc" gracePeriod=2 Mar 13 10:58:35 crc kubenswrapper[4632]: I0313 10:58:35.940288 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.045025 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:58:36 crc kubenswrapper[4632]: E0313 10:58:36.045380 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.069877 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-459zd\" (UniqueName: \"kubernetes.io/projected/48d9de0e-efea-443b-89a7-e02d3264020f-kube-api-access-459zd\") pod \"48d9de0e-efea-443b-89a7-e02d3264020f\" (UID: \"48d9de0e-efea-443b-89a7-e02d3264020f\") " Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.070141 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d9de0e-efea-443b-89a7-e02d3264020f-utilities\") pod \"48d9de0e-efea-443b-89a7-e02d3264020f\" (UID: \"48d9de0e-efea-443b-89a7-e02d3264020f\") " Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.070227 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d9de0e-efea-443b-89a7-e02d3264020f-catalog-content\") pod \"48d9de0e-efea-443b-89a7-e02d3264020f\" (UID: \"48d9de0e-efea-443b-89a7-e02d3264020f\") " Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.070740 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48d9de0e-efea-443b-89a7-e02d3264020f-utilities" (OuterVolumeSpecName: "utilities") pod "48d9de0e-efea-443b-89a7-e02d3264020f" (UID: "48d9de0e-efea-443b-89a7-e02d3264020f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.079497 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48d9de0e-efea-443b-89a7-e02d3264020f-kube-api-access-459zd" (OuterVolumeSpecName: "kube-api-access-459zd") pod "48d9de0e-efea-443b-89a7-e02d3264020f" (UID: "48d9de0e-efea-443b-89a7-e02d3264020f"). InnerVolumeSpecName "kube-api-access-459zd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.117683 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48d9de0e-efea-443b-89a7-e02d3264020f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48d9de0e-efea-443b-89a7-e02d3264020f" (UID: "48d9de0e-efea-443b-89a7-e02d3264020f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.173334 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-459zd\" (UniqueName: \"kubernetes.io/projected/48d9de0e-efea-443b-89a7-e02d3264020f-kube-api-access-459zd\") on node \"crc\" DevicePath \"\"" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.173368 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48d9de0e-efea-443b-89a7-e02d3264020f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.173379 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48d9de0e-efea-443b-89a7-e02d3264020f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.312348 4632 generic.go:334] "Generic (PLEG): container finished" podID="48d9de0e-efea-443b-89a7-e02d3264020f" containerID="a35447b594c379c0f9e9337a36f14fcb40355b64f78e7eff5cb0f3ce866933cc" exitCode=0 Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.312400 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9g65" event={"ID":"48d9de0e-efea-443b-89a7-e02d3264020f","Type":"ContainerDied","Data":"a35447b594c379c0f9e9337a36f14fcb40355b64f78e7eff5cb0f3ce866933cc"} Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.312429 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q9g65" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.312443 4632 scope.go:117] "RemoveContainer" containerID="a35447b594c379c0f9e9337a36f14fcb40355b64f78e7eff5cb0f3ce866933cc" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.312431 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q9g65" event={"ID":"48d9de0e-efea-443b-89a7-e02d3264020f","Type":"ContainerDied","Data":"791d787bc94f7eeb2e183a040f421e97af028bfea047785789f3496908b775af"} Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.357159 4632 scope.go:117] "RemoveContainer" containerID="e3f2fccdad049285eb344aa03ec4c45ce2e653e042872421419872537e65b0ce" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.362256 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9g65"] Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.375174 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q9g65"] Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.395386 4632 scope.go:117] "RemoveContainer" containerID="89e85aed0ace971ea27aa6e4381a5b9c378475b156c657d5f7f64cf10fffb7e7" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.430207 4632 scope.go:117] "RemoveContainer" containerID="a35447b594c379c0f9e9337a36f14fcb40355b64f78e7eff5cb0f3ce866933cc" Mar 13 10:58:36 crc kubenswrapper[4632]: E0313 10:58:36.438347 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a35447b594c379c0f9e9337a36f14fcb40355b64f78e7eff5cb0f3ce866933cc\": container with ID starting with a35447b594c379c0f9e9337a36f14fcb40355b64f78e7eff5cb0f3ce866933cc not found: ID does not exist" containerID="a35447b594c379c0f9e9337a36f14fcb40355b64f78e7eff5cb0f3ce866933cc" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.439203 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a35447b594c379c0f9e9337a36f14fcb40355b64f78e7eff5cb0f3ce866933cc"} err="failed to get container status \"a35447b594c379c0f9e9337a36f14fcb40355b64f78e7eff5cb0f3ce866933cc\": rpc error: code = NotFound desc = could not find container \"a35447b594c379c0f9e9337a36f14fcb40355b64f78e7eff5cb0f3ce866933cc\": container with ID starting with a35447b594c379c0f9e9337a36f14fcb40355b64f78e7eff5cb0f3ce866933cc not found: ID does not exist" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.439246 4632 scope.go:117] "RemoveContainer" containerID="e3f2fccdad049285eb344aa03ec4c45ce2e653e042872421419872537e65b0ce" Mar 13 10:58:36 crc kubenswrapper[4632]: E0313 10:58:36.439748 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3f2fccdad049285eb344aa03ec4c45ce2e653e042872421419872537e65b0ce\": container with ID starting with e3f2fccdad049285eb344aa03ec4c45ce2e653e042872421419872537e65b0ce not found: ID does not exist" containerID="e3f2fccdad049285eb344aa03ec4c45ce2e653e042872421419872537e65b0ce" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.439775 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3f2fccdad049285eb344aa03ec4c45ce2e653e042872421419872537e65b0ce"} err="failed to get container status \"e3f2fccdad049285eb344aa03ec4c45ce2e653e042872421419872537e65b0ce\": rpc error: code = NotFound desc = could not find container \"e3f2fccdad049285eb344aa03ec4c45ce2e653e042872421419872537e65b0ce\": container with ID starting with e3f2fccdad049285eb344aa03ec4c45ce2e653e042872421419872537e65b0ce not found: ID does not exist" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.439790 4632 scope.go:117] "RemoveContainer" containerID="89e85aed0ace971ea27aa6e4381a5b9c378475b156c657d5f7f64cf10fffb7e7" Mar 13 10:58:36 crc kubenswrapper[4632]: E0313 10:58:36.440032 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89e85aed0ace971ea27aa6e4381a5b9c378475b156c657d5f7f64cf10fffb7e7\": container with ID starting with 89e85aed0ace971ea27aa6e4381a5b9c378475b156c657d5f7f64cf10fffb7e7 not found: ID does not exist" containerID="89e85aed0ace971ea27aa6e4381a5b9c378475b156c657d5f7f64cf10fffb7e7" Mar 13 10:58:36 crc kubenswrapper[4632]: I0313 10:58:36.440051 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89e85aed0ace971ea27aa6e4381a5b9c378475b156c657d5f7f64cf10fffb7e7"} err="failed to get container status \"89e85aed0ace971ea27aa6e4381a5b9c378475b156c657d5f7f64cf10fffb7e7\": rpc error: code = NotFound desc = could not find container \"89e85aed0ace971ea27aa6e4381a5b9c378475b156c657d5f7f64cf10fffb7e7\": container with ID starting with 89e85aed0ace971ea27aa6e4381a5b9c378475b156c657d5f7f64cf10fffb7e7 not found: ID does not exist" Mar 13 10:58:38 crc kubenswrapper[4632]: I0313 10:58:38.062349 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48d9de0e-efea-443b-89a7-e02d3264020f" path="/var/lib/kubelet/pods/48d9de0e-efea-443b-89a7-e02d3264020f/volumes" Mar 13 10:58:49 crc kubenswrapper[4632]: I0313 10:58:49.044905 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:58:49 crc kubenswrapper[4632]: E0313 10:58:49.045937 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:59:00 crc kubenswrapper[4632]: I0313 10:59:00.044481 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:59:00 crc kubenswrapper[4632]: E0313 10:59:00.045661 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 10:59:12 crc kubenswrapper[4632]: I0313 10:59:12.045768 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 10:59:12 crc kubenswrapper[4632]: I0313 10:59:12.665883 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"06408f7526caaaeae759484ccf3ff85a146655a7d51ff7049c7be79b39fe96ba"} Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.674879 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556660-dxrjl"] Mar 13 11:00:00 crc kubenswrapper[4632]: E0313 11:00:00.683172 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c91e27-6596-4471-81e9-4a65e55379cc" containerName="registry-server" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.683213 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c91e27-6596-4471-81e9-4a65e55379cc" containerName="registry-server" Mar 13 11:00:00 crc kubenswrapper[4632]: E0313 11:00:00.683546 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c91e27-6596-4471-81e9-4a65e55379cc" containerName="extract-utilities" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.683561 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c91e27-6596-4471-81e9-4a65e55379cc" containerName="extract-utilities" Mar 13 11:00:00 crc kubenswrapper[4632]: E0313 11:00:00.683581 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48d9de0e-efea-443b-89a7-e02d3264020f" containerName="registry-server" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.683589 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="48d9de0e-efea-443b-89a7-e02d3264020f" containerName="registry-server" Mar 13 11:00:00 crc kubenswrapper[4632]: E0313 11:00:00.683609 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7c91e27-6596-4471-81e9-4a65e55379cc" containerName="extract-content" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.683616 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7c91e27-6596-4471-81e9-4a65e55379cc" containerName="extract-content" Mar 13 11:00:00 crc kubenswrapper[4632]: E0313 11:00:00.683627 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48d9de0e-efea-443b-89a7-e02d3264020f" containerName="extract-content" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.683633 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="48d9de0e-efea-443b-89a7-e02d3264020f" containerName="extract-content" Mar 13 11:00:00 crc kubenswrapper[4632]: E0313 11:00:00.683643 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48d9de0e-efea-443b-89a7-e02d3264020f" containerName="extract-utilities" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.683658 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="48d9de0e-efea-443b-89a7-e02d3264020f" containerName="extract-utilities" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.683906 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="48d9de0e-efea-443b-89a7-e02d3264020f" containerName="registry-server" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.683931 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7c91e27-6596-4471-81e9-4a65e55379cc" containerName="registry-server" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.691477 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8"] Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.696701 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.696691 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556660-dxrjl" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.723594 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.723604 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.723703 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.723733 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.726600 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.750923 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8"] Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.794075 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556660-dxrjl"] Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.849996 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ktg5\" (UniqueName: \"kubernetes.io/projected/f506e288-f3da-4d62-a6a2-bb598a62ed13-kube-api-access-2ktg5\") pod \"collect-profiles-29556660-7vph8\" (UID: \"f506e288-f3da-4d62-a6a2-bb598a62ed13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.850357 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f506e288-f3da-4d62-a6a2-bb598a62ed13-secret-volume\") pod \"collect-profiles-29556660-7vph8\" (UID: \"f506e288-f3da-4d62-a6a2-bb598a62ed13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.850738 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f506e288-f3da-4d62-a6a2-bb598a62ed13-config-volume\") pod \"collect-profiles-29556660-7vph8\" (UID: \"f506e288-f3da-4d62-a6a2-bb598a62ed13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.850819 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjskv\" (UniqueName: \"kubernetes.io/projected/ae414ebe-e9fa-4c30-965a-e368234bbb18-kube-api-access-cjskv\") pod \"auto-csr-approver-29556660-dxrjl\" (UID: \"ae414ebe-e9fa-4c30-965a-e368234bbb18\") " pod="openshift-infra/auto-csr-approver-29556660-dxrjl" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.952302 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f506e288-f3da-4d62-a6a2-bb598a62ed13-config-volume\") pod \"collect-profiles-29556660-7vph8\" (UID: \"f506e288-f3da-4d62-a6a2-bb598a62ed13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.952367 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjskv\" (UniqueName: \"kubernetes.io/projected/ae414ebe-e9fa-4c30-965a-e368234bbb18-kube-api-access-cjskv\") pod \"auto-csr-approver-29556660-dxrjl\" (UID: \"ae414ebe-e9fa-4c30-965a-e368234bbb18\") " pod="openshift-infra/auto-csr-approver-29556660-dxrjl" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.952492 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ktg5\" (UniqueName: \"kubernetes.io/projected/f506e288-f3da-4d62-a6a2-bb598a62ed13-kube-api-access-2ktg5\") pod \"collect-profiles-29556660-7vph8\" (UID: \"f506e288-f3da-4d62-a6a2-bb598a62ed13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.952522 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f506e288-f3da-4d62-a6a2-bb598a62ed13-secret-volume\") pod \"collect-profiles-29556660-7vph8\" (UID: \"f506e288-f3da-4d62-a6a2-bb598a62ed13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.962792 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f506e288-f3da-4d62-a6a2-bb598a62ed13-config-volume\") pod \"collect-profiles-29556660-7vph8\" (UID: \"f506e288-f3da-4d62-a6a2-bb598a62ed13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.977816 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f506e288-f3da-4d62-a6a2-bb598a62ed13-secret-volume\") pod \"collect-profiles-29556660-7vph8\" (UID: \"f506e288-f3da-4d62-a6a2-bb598a62ed13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.979704 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ktg5\" (UniqueName: \"kubernetes.io/projected/f506e288-f3da-4d62-a6a2-bb598a62ed13-kube-api-access-2ktg5\") pod \"collect-profiles-29556660-7vph8\" (UID: \"f506e288-f3da-4d62-a6a2-bb598a62ed13\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" Mar 13 11:00:00 crc kubenswrapper[4632]: I0313 11:00:00.983798 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjskv\" (UniqueName: \"kubernetes.io/projected/ae414ebe-e9fa-4c30-965a-e368234bbb18-kube-api-access-cjskv\") pod \"auto-csr-approver-29556660-dxrjl\" (UID: \"ae414ebe-e9fa-4c30-965a-e368234bbb18\") " pod="openshift-infra/auto-csr-approver-29556660-dxrjl" Mar 13 11:00:01 crc kubenswrapper[4632]: I0313 11:00:01.088618 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" Mar 13 11:00:01 crc kubenswrapper[4632]: I0313 11:00:01.126609 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556660-dxrjl" Mar 13 11:00:02 crc kubenswrapper[4632]: I0313 11:00:02.583698 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8"] Mar 13 11:00:02 crc kubenswrapper[4632]: I0313 11:00:02.598207 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556660-dxrjl"] Mar 13 11:00:03 crc kubenswrapper[4632]: I0313 11:00:03.245201 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" event={"ID":"f506e288-f3da-4d62-a6a2-bb598a62ed13","Type":"ContainerStarted","Data":"df99b126bcdc13810e89ae823dc76bf43eab9d932c52b6dd430fa449a698c642"} Mar 13 11:00:03 crc kubenswrapper[4632]: I0313 11:00:03.245525 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" event={"ID":"f506e288-f3da-4d62-a6a2-bb598a62ed13","Type":"ContainerStarted","Data":"34d565725937589b17417cd9a8d096a43573544687add3bac92bfe268458bb39"} Mar 13 11:00:03 crc kubenswrapper[4632]: I0313 11:00:03.247433 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556660-dxrjl" event={"ID":"ae414ebe-e9fa-4c30-965a-e368234bbb18","Type":"ContainerStarted","Data":"6ccc5a36abf2c1c45c76ea18c7f259f925ab314e33e7f60a14b3018f1f22a313"} Mar 13 11:00:03 crc kubenswrapper[4632]: I0313 11:00:03.273257 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" podStartSLOduration=3.271039179 podStartE2EDuration="3.271039179s" podCreationTimestamp="2026-03-13 11:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 11:00:03.262089487 +0000 UTC m=+3377.284619640" watchObservedRunningTime="2026-03-13 11:00:03.271039179 +0000 UTC m=+3377.293569322" Mar 13 11:00:04 crc kubenswrapper[4632]: I0313 11:00:04.259404 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" event={"ID":"f506e288-f3da-4d62-a6a2-bb598a62ed13","Type":"ContainerDied","Data":"df99b126bcdc13810e89ae823dc76bf43eab9d932c52b6dd430fa449a698c642"} Mar 13 11:00:04 crc kubenswrapper[4632]: I0313 11:00:04.259885 4632 generic.go:334] "Generic (PLEG): container finished" podID="f506e288-f3da-4d62-a6a2-bb598a62ed13" containerID="df99b126bcdc13810e89ae823dc76bf43eab9d932c52b6dd430fa449a698c642" exitCode=0 Mar 13 11:00:05 crc kubenswrapper[4632]: I0313 11:00:05.885673 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" Mar 13 11:00:05 crc kubenswrapper[4632]: I0313 11:00:05.955997 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f506e288-f3da-4d62-a6a2-bb598a62ed13-config-volume\") pod \"f506e288-f3da-4d62-a6a2-bb598a62ed13\" (UID: \"f506e288-f3da-4d62-a6a2-bb598a62ed13\") " Mar 13 11:00:05 crc kubenswrapper[4632]: I0313 11:00:05.956165 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f506e288-f3da-4d62-a6a2-bb598a62ed13-secret-volume\") pod \"f506e288-f3da-4d62-a6a2-bb598a62ed13\" (UID: \"f506e288-f3da-4d62-a6a2-bb598a62ed13\") " Mar 13 11:00:05 crc kubenswrapper[4632]: I0313 11:00:05.956280 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ktg5\" (UniqueName: \"kubernetes.io/projected/f506e288-f3da-4d62-a6a2-bb598a62ed13-kube-api-access-2ktg5\") pod \"f506e288-f3da-4d62-a6a2-bb598a62ed13\" (UID: \"f506e288-f3da-4d62-a6a2-bb598a62ed13\") " Mar 13 11:00:06 crc kubenswrapper[4632]: I0313 11:00:05.971904 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f506e288-f3da-4d62-a6a2-bb598a62ed13-config-volume" (OuterVolumeSpecName: "config-volume") pod "f506e288-f3da-4d62-a6a2-bb598a62ed13" (UID: "f506e288-f3da-4d62-a6a2-bb598a62ed13"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 11:00:06 crc kubenswrapper[4632]: I0313 11:00:06.062022 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f506e288-f3da-4d62-a6a2-bb598a62ed13-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f506e288-f3da-4d62-a6a2-bb598a62ed13" (UID: "f506e288-f3da-4d62-a6a2-bb598a62ed13"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:00:06 crc kubenswrapper[4632]: I0313 11:00:06.078878 4632 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f506e288-f3da-4d62-a6a2-bb598a62ed13-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 11:00:06 crc kubenswrapper[4632]: I0313 11:00:06.080221 4632 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f506e288-f3da-4d62-a6a2-bb598a62ed13-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 11:00:06 crc kubenswrapper[4632]: I0313 11:00:06.121199 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f506e288-f3da-4d62-a6a2-bb598a62ed13-kube-api-access-2ktg5" (OuterVolumeSpecName: "kube-api-access-2ktg5") pod "f506e288-f3da-4d62-a6a2-bb598a62ed13" (UID: "f506e288-f3da-4d62-a6a2-bb598a62ed13"). InnerVolumeSpecName "kube-api-access-2ktg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:00:06 crc kubenswrapper[4632]: I0313 11:00:06.182340 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ktg5\" (UniqueName: \"kubernetes.io/projected/f506e288-f3da-4d62-a6a2-bb598a62ed13-kube-api-access-2ktg5\") on node \"crc\" DevicePath \"\"" Mar 13 11:00:06 crc kubenswrapper[4632]: I0313 11:00:06.277411 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" event={"ID":"f506e288-f3da-4d62-a6a2-bb598a62ed13","Type":"ContainerDied","Data":"34d565725937589b17417cd9a8d096a43573544687add3bac92bfe268458bb39"} Mar 13 11:00:06 crc kubenswrapper[4632]: I0313 11:00:06.277469 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="34d565725937589b17417cd9a8d096a43573544687add3bac92bfe268458bb39" Mar 13 11:00:06 crc kubenswrapper[4632]: I0313 11:00:06.277552 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8" Mar 13 11:00:06 crc kubenswrapper[4632]: E0313 11:00:06.365949 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf506e288_f3da_4d62_a6a2_bb598a62ed13.slice/crio-34d565725937589b17417cd9a8d096a43573544687add3bac92bfe268458bb39\": RecentStats: unable to find data in memory cache]" Mar 13 11:00:07 crc kubenswrapper[4632]: I0313 11:00:07.023932 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm"] Mar 13 11:00:07 crc kubenswrapper[4632]: I0313 11:00:07.032314 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556615-zj2fm"] Mar 13 11:00:07 crc kubenswrapper[4632]: I0313 11:00:07.288073 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556660-dxrjl" event={"ID":"ae414ebe-e9fa-4c30-965a-e368234bbb18","Type":"ContainerStarted","Data":"b3d4b9e8bcea3a6dbdeee6316ce9071df3a8c8906a4c416a00caede29a1de5ca"} Mar 13 11:00:07 crc kubenswrapper[4632]: I0313 11:00:07.305740 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556660-dxrjl" podStartSLOduration=4.645795119 podStartE2EDuration="7.305713231s" podCreationTimestamp="2026-03-13 11:00:00 +0000 UTC" firstStartedPulling="2026-03-13 11:00:02.657927528 +0000 UTC m=+3376.680457661" lastFinishedPulling="2026-03-13 11:00:05.31784564 +0000 UTC m=+3379.340375773" observedRunningTime="2026-03-13 11:00:07.30042434 +0000 UTC m=+3381.322954473" watchObservedRunningTime="2026-03-13 11:00:07.305713231 +0000 UTC m=+3381.328243374" Mar 13 11:00:08 crc kubenswrapper[4632]: I0313 11:00:08.072524 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3c3c6392-454c-4131-90a0-6584565cef4c" path="/var/lib/kubelet/pods/3c3c6392-454c-4131-90a0-6584565cef4c/volumes" Mar 13 11:00:08 crc kubenswrapper[4632]: I0313 11:00:08.308821 4632 generic.go:334] "Generic (PLEG): container finished" podID="ae414ebe-e9fa-4c30-965a-e368234bbb18" containerID="b3d4b9e8bcea3a6dbdeee6316ce9071df3a8c8906a4c416a00caede29a1de5ca" exitCode=0 Mar 13 11:00:08 crc kubenswrapper[4632]: I0313 11:00:08.308876 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556660-dxrjl" event={"ID":"ae414ebe-e9fa-4c30-965a-e368234bbb18","Type":"ContainerDied","Data":"b3d4b9e8bcea3a6dbdeee6316ce9071df3a8c8906a4c416a00caede29a1de5ca"} Mar 13 11:00:09 crc kubenswrapper[4632]: I0313 11:00:09.897970 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556660-dxrjl" Mar 13 11:00:10 crc kubenswrapper[4632]: I0313 11:00:10.056228 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjskv\" (UniqueName: \"kubernetes.io/projected/ae414ebe-e9fa-4c30-965a-e368234bbb18-kube-api-access-cjskv\") pod \"ae414ebe-e9fa-4c30-965a-e368234bbb18\" (UID: \"ae414ebe-e9fa-4c30-965a-e368234bbb18\") " Mar 13 11:00:10 crc kubenswrapper[4632]: I0313 11:00:10.068403 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae414ebe-e9fa-4c30-965a-e368234bbb18-kube-api-access-cjskv" (OuterVolumeSpecName: "kube-api-access-cjskv") pod "ae414ebe-e9fa-4c30-965a-e368234bbb18" (UID: "ae414ebe-e9fa-4c30-965a-e368234bbb18"). InnerVolumeSpecName "kube-api-access-cjskv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:00:10 crc kubenswrapper[4632]: I0313 11:00:10.158983 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjskv\" (UniqueName: \"kubernetes.io/projected/ae414ebe-e9fa-4c30-965a-e368234bbb18-kube-api-access-cjskv\") on node \"crc\" DevicePath \"\"" Mar 13 11:00:10 crc kubenswrapper[4632]: I0313 11:00:10.325979 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556660-dxrjl" event={"ID":"ae414ebe-e9fa-4c30-965a-e368234bbb18","Type":"ContainerDied","Data":"6ccc5a36abf2c1c45c76ea18c7f259f925ab314e33e7f60a14b3018f1f22a313"} Mar 13 11:00:10 crc kubenswrapper[4632]: I0313 11:00:10.326027 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ccc5a36abf2c1c45c76ea18c7f259f925ab314e33e7f60a14b3018f1f22a313" Mar 13 11:00:10 crc kubenswrapper[4632]: I0313 11:00:10.326068 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556660-dxrjl" Mar 13 11:00:10 crc kubenswrapper[4632]: I0313 11:00:10.388699 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556654-htcrm"] Mar 13 11:00:10 crc kubenswrapper[4632]: I0313 11:00:10.420478 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556654-htcrm"] Mar 13 11:00:10 crc kubenswrapper[4632]: I0313 11:00:10.928040 4632 scope.go:117] "RemoveContainer" containerID="6acdfd407705651773e15ca9493f2efdac886dce6f04123c798b57f93aa775b6" Mar 13 11:00:12 crc kubenswrapper[4632]: I0313 11:00:12.060216 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba5d59c2-ece8-4b66-9a10-c3ef740d7e45" path="/var/lib/kubelet/pods/ba5d59c2-ece8-4b66-9a10-c3ef740d7e45/volumes" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.137196 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29556661-2p4wf"] Mar 13 11:01:01 crc kubenswrapper[4632]: E0313 11:01:01.142839 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f506e288-f3da-4d62-a6a2-bb598a62ed13" containerName="collect-profiles" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.142869 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f506e288-f3da-4d62-a6a2-bb598a62ed13" containerName="collect-profiles" Mar 13 11:01:01 crc kubenswrapper[4632]: E0313 11:01:01.143883 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae414ebe-e9fa-4c30-965a-e368234bbb18" containerName="oc" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.143913 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae414ebe-e9fa-4c30-965a-e368234bbb18" containerName="oc" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.145911 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f506e288-f3da-4d62-a6a2-bb598a62ed13" containerName="collect-profiles" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.145969 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae414ebe-e9fa-4c30-965a-e368234bbb18" containerName="oc" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.160794 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.262751 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwj9v\" (UniqueName: \"kubernetes.io/projected/6c20fa3e-2873-4076-b17a-3ee171199959-kube-api-access-hwj9v\") pod \"keystone-cron-29556661-2p4wf\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.263032 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-config-data\") pod \"keystone-cron-29556661-2p4wf\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.263138 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-fernet-keys\") pod \"keystone-cron-29556661-2p4wf\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.263172 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-combined-ca-bundle\") pod \"keystone-cron-29556661-2p4wf\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.363017 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29556661-2p4wf"] Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.364764 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-config-data\") pod \"keystone-cron-29556661-2p4wf\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.364864 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-fernet-keys\") pod \"keystone-cron-29556661-2p4wf\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.364900 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-combined-ca-bundle\") pod \"keystone-cron-29556661-2p4wf\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.365029 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwj9v\" (UniqueName: \"kubernetes.io/projected/6c20fa3e-2873-4076-b17a-3ee171199959-kube-api-access-hwj9v\") pod \"keystone-cron-29556661-2p4wf\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.411556 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwj9v\" (UniqueName: \"kubernetes.io/projected/6c20fa3e-2873-4076-b17a-3ee171199959-kube-api-access-hwj9v\") pod \"keystone-cron-29556661-2p4wf\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.414840 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-combined-ca-bundle\") pod \"keystone-cron-29556661-2p4wf\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.419393 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-config-data\") pod \"keystone-cron-29556661-2p4wf\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.423815 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-fernet-keys\") pod \"keystone-cron-29556661-2p4wf\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:01 crc kubenswrapper[4632]: I0313 11:01:01.528455 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:03 crc kubenswrapper[4632]: I0313 11:01:03.082761 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29556661-2p4wf"] Mar 13 11:01:03 crc kubenswrapper[4632]: I0313 11:01:03.806927 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556661-2p4wf" event={"ID":"6c20fa3e-2873-4076-b17a-3ee171199959","Type":"ContainerStarted","Data":"50957f6ac51a786ec085c100c45b3293f56673563593051170c3bcfe4bce73f5"} Mar 13 11:01:03 crc kubenswrapper[4632]: I0313 11:01:03.808427 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556661-2p4wf" event={"ID":"6c20fa3e-2873-4076-b17a-3ee171199959","Type":"ContainerStarted","Data":"c1622395247bf78009a91c0c227d8aea33bd7858a8489000b7092798dfc7bfe0"} Mar 13 11:01:03 crc kubenswrapper[4632]: I0313 11:01:03.864184 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29556661-2p4wf" podStartSLOduration=3.8623800299999997 podStartE2EDuration="3.86238003s" podCreationTimestamp="2026-03-13 11:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 11:01:03.856938545 +0000 UTC m=+3437.879468688" watchObservedRunningTime="2026-03-13 11:01:03.86238003 +0000 UTC m=+3437.884910163" Mar 13 11:01:09 crc kubenswrapper[4632]: I0313 11:01:09.888154 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556661-2p4wf" event={"ID":"6c20fa3e-2873-4076-b17a-3ee171199959","Type":"ContainerDied","Data":"50957f6ac51a786ec085c100c45b3293f56673563593051170c3bcfe4bce73f5"} Mar 13 11:01:09 crc kubenswrapper[4632]: I0313 11:01:09.888604 4632 generic.go:334] "Generic (PLEG): container finished" podID="6c20fa3e-2873-4076-b17a-3ee171199959" containerID="50957f6ac51a786ec085c100c45b3293f56673563593051170c3bcfe4bce73f5" exitCode=0 Mar 13 11:01:11 crc kubenswrapper[4632]: I0313 11:01:11.104744 4632 scope.go:117] "RemoveContainer" containerID="0151dac58382ec9dba1fe485dee8519ba248333bc8e6aeae5349b66a4c5fa931" Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.505436 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.642194 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-fernet-keys\") pod \"6c20fa3e-2873-4076-b17a-3ee171199959\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.642328 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwj9v\" (UniqueName: \"kubernetes.io/projected/6c20fa3e-2873-4076-b17a-3ee171199959-kube-api-access-hwj9v\") pod \"6c20fa3e-2873-4076-b17a-3ee171199959\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.642532 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-combined-ca-bundle\") pod \"6c20fa3e-2873-4076-b17a-3ee171199959\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.642588 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-config-data\") pod \"6c20fa3e-2873-4076-b17a-3ee171199959\" (UID: \"6c20fa3e-2873-4076-b17a-3ee171199959\") " Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.694791 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c20fa3e-2873-4076-b17a-3ee171199959-kube-api-access-hwj9v" (OuterVolumeSpecName: "kube-api-access-hwj9v") pod "6c20fa3e-2873-4076-b17a-3ee171199959" (UID: "6c20fa3e-2873-4076-b17a-3ee171199959"). InnerVolumeSpecName "kube-api-access-hwj9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.696660 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6c20fa3e-2873-4076-b17a-3ee171199959" (UID: "6c20fa3e-2873-4076-b17a-3ee171199959"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.724977 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6c20fa3e-2873-4076-b17a-3ee171199959" (UID: "6c20fa3e-2873-4076-b17a-3ee171199959"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.736869 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-config-data" (OuterVolumeSpecName: "config-data") pod "6c20fa3e-2873-4076-b17a-3ee171199959" (UID: "6c20fa3e-2873-4076-b17a-3ee171199959"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.745603 4632 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.745801 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwj9v\" (UniqueName: \"kubernetes.io/projected/6c20fa3e-2873-4076-b17a-3ee171199959-kube-api-access-hwj9v\") on node \"crc\" DevicePath \"\"" Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.745885 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.746170 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6c20fa3e-2873-4076-b17a-3ee171199959-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.921550 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556661-2p4wf" event={"ID":"6c20fa3e-2873-4076-b17a-3ee171199959","Type":"ContainerDied","Data":"c1622395247bf78009a91c0c227d8aea33bd7858a8489000b7092798dfc7bfe0"} Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.922125 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556661-2p4wf" Mar 13 11:01:12 crc kubenswrapper[4632]: I0313 11:01:12.922521 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1622395247bf78009a91c0c227d8aea33bd7858a8489000b7092798dfc7bfe0" Mar 13 11:01:40 crc kubenswrapper[4632]: I0313 11:01:40.470500 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:01:40 crc kubenswrapper[4632]: I0313 11:01:40.482260 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:02:00 crc kubenswrapper[4632]: I0313 11:02:00.817260 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556662-pw9tk"] Mar 13 11:02:00 crc kubenswrapper[4632]: E0313 11:02:00.821758 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c20fa3e-2873-4076-b17a-3ee171199959" containerName="keystone-cron" Mar 13 11:02:00 crc kubenswrapper[4632]: I0313 11:02:00.822245 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c20fa3e-2873-4076-b17a-3ee171199959" containerName="keystone-cron" Mar 13 11:02:00 crc kubenswrapper[4632]: I0313 11:02:00.824993 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c20fa3e-2873-4076-b17a-3ee171199959" containerName="keystone-cron" Mar 13 11:02:00 crc kubenswrapper[4632]: I0313 11:02:00.833305 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556662-pw9tk" Mar 13 11:02:00 crc kubenswrapper[4632]: I0313 11:02:00.851620 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:02:00 crc kubenswrapper[4632]: I0313 11:02:00.851636 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:02:00 crc kubenswrapper[4632]: I0313 11:02:00.851651 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:02:00 crc kubenswrapper[4632]: I0313 11:02:00.927962 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbhfw\" (UniqueName: \"kubernetes.io/projected/62b0f696-9e5c-4535-a181-fa2f4b645711-kube-api-access-hbhfw\") pod \"auto-csr-approver-29556662-pw9tk\" (UID: \"62b0f696-9e5c-4535-a181-fa2f4b645711\") " pod="openshift-infra/auto-csr-approver-29556662-pw9tk" Mar 13 11:02:00 crc kubenswrapper[4632]: I0313 11:02:00.998568 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556662-pw9tk"] Mar 13 11:02:01 crc kubenswrapper[4632]: I0313 11:02:01.030000 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbhfw\" (UniqueName: \"kubernetes.io/projected/62b0f696-9e5c-4535-a181-fa2f4b645711-kube-api-access-hbhfw\") pod \"auto-csr-approver-29556662-pw9tk\" (UID: \"62b0f696-9e5c-4535-a181-fa2f4b645711\") " pod="openshift-infra/auto-csr-approver-29556662-pw9tk" Mar 13 11:02:01 crc kubenswrapper[4632]: I0313 11:02:01.084852 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbhfw\" (UniqueName: \"kubernetes.io/projected/62b0f696-9e5c-4535-a181-fa2f4b645711-kube-api-access-hbhfw\") pod \"auto-csr-approver-29556662-pw9tk\" (UID: \"62b0f696-9e5c-4535-a181-fa2f4b645711\") " pod="openshift-infra/auto-csr-approver-29556662-pw9tk" Mar 13 11:02:01 crc kubenswrapper[4632]: I0313 11:02:01.169122 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556662-pw9tk" Mar 13 11:02:03 crc kubenswrapper[4632]: I0313 11:02:03.038114 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556662-pw9tk"] Mar 13 11:02:03 crc kubenswrapper[4632]: I0313 11:02:03.390324 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556662-pw9tk" event={"ID":"62b0f696-9e5c-4535-a181-fa2f4b645711","Type":"ContainerStarted","Data":"1af7d2f0857a2610ddaf36212c411393a094b387bb07f0eb12ad83a933bc22f0"} Mar 13 11:02:05 crc kubenswrapper[4632]: I0313 11:02:05.411467 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556662-pw9tk" event={"ID":"62b0f696-9e5c-4535-a181-fa2f4b645711","Type":"ContainerStarted","Data":"69d080c6683237a330690584133c6005521df29f2dcf4c21ed9a518e4de4e991"} Mar 13 11:02:05 crc kubenswrapper[4632]: I0313 11:02:05.435409 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556662-pw9tk" podStartSLOduration=4.413347009 podStartE2EDuration="5.434390052s" podCreationTimestamp="2026-03-13 11:02:00 +0000 UTC" firstStartedPulling="2026-03-13 11:02:03.090780692 +0000 UTC m=+3497.113310825" lastFinishedPulling="2026-03-13 11:02:04.111823735 +0000 UTC m=+3498.134353868" observedRunningTime="2026-03-13 11:02:05.430828474 +0000 UTC m=+3499.453358627" watchObservedRunningTime="2026-03-13 11:02:05.434390052 +0000 UTC m=+3499.456920185" Mar 13 11:02:07 crc kubenswrapper[4632]: I0313 11:02:07.454786 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556662-pw9tk" event={"ID":"62b0f696-9e5c-4535-a181-fa2f4b645711","Type":"ContainerDied","Data":"69d080c6683237a330690584133c6005521df29f2dcf4c21ed9a518e4de4e991"} Mar 13 11:02:07 crc kubenswrapper[4632]: I0313 11:02:07.454715 4632 generic.go:334] "Generic (PLEG): container finished" podID="62b0f696-9e5c-4535-a181-fa2f4b645711" containerID="69d080c6683237a330690584133c6005521df29f2dcf4c21ed9a518e4de4e991" exitCode=0 Mar 13 11:02:09 crc kubenswrapper[4632]: I0313 11:02:09.900893 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556662-pw9tk" Mar 13 11:02:10 crc kubenswrapper[4632]: I0313 11:02:09.999479 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hbhfw\" (UniqueName: \"kubernetes.io/projected/62b0f696-9e5c-4535-a181-fa2f4b645711-kube-api-access-hbhfw\") pod \"62b0f696-9e5c-4535-a181-fa2f4b645711\" (UID: \"62b0f696-9e5c-4535-a181-fa2f4b645711\") " Mar 13 11:02:10 crc kubenswrapper[4632]: I0313 11:02:10.044238 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62b0f696-9e5c-4535-a181-fa2f4b645711-kube-api-access-hbhfw" (OuterVolumeSpecName: "kube-api-access-hbhfw") pod "62b0f696-9e5c-4535-a181-fa2f4b645711" (UID: "62b0f696-9e5c-4535-a181-fa2f4b645711"). InnerVolumeSpecName "kube-api-access-hbhfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:02:10 crc kubenswrapper[4632]: I0313 11:02:10.101920 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hbhfw\" (UniqueName: \"kubernetes.io/projected/62b0f696-9e5c-4535-a181-fa2f4b645711-kube-api-access-hbhfw\") on node \"crc\" DevicePath \"\"" Mar 13 11:02:10 crc kubenswrapper[4632]: I0313 11:02:10.463498 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:02:10 crc kubenswrapper[4632]: I0313 11:02:10.465242 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:02:10 crc kubenswrapper[4632]: I0313 11:02:10.489492 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556662-pw9tk" event={"ID":"62b0f696-9e5c-4535-a181-fa2f4b645711","Type":"ContainerDied","Data":"1af7d2f0857a2610ddaf36212c411393a094b387bb07f0eb12ad83a933bc22f0"} Mar 13 11:02:10 crc kubenswrapper[4632]: I0313 11:02:10.489542 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1af7d2f0857a2610ddaf36212c411393a094b387bb07f0eb12ad83a933bc22f0" Mar 13 11:02:10 crc kubenswrapper[4632]: I0313 11:02:10.489553 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556662-pw9tk" Mar 13 11:02:11 crc kubenswrapper[4632]: I0313 11:02:11.049061 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556656-vzb8p"] Mar 13 11:02:11 crc kubenswrapper[4632]: I0313 11:02:11.068997 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556656-vzb8p"] Mar 13 11:02:12 crc kubenswrapper[4632]: I0313 11:02:12.057850 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f70e2037-a5d6-4479-af7f-18fe8ff9e952" path="/var/lib/kubelet/pods/f70e2037-a5d6-4479-af7f-18fe8ff9e952/volumes" Mar 13 11:02:40 crc kubenswrapper[4632]: I0313 11:02:40.465015 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:02:40 crc kubenswrapper[4632]: I0313 11:02:40.469605 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:02:40 crc kubenswrapper[4632]: I0313 11:02:40.472251 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 11:02:40 crc kubenswrapper[4632]: I0313 11:02:40.479160 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"06408f7526caaaeae759484ccf3ff85a146655a7d51ff7049c7be79b39fe96ba"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 11:02:40 crc kubenswrapper[4632]: I0313 11:02:40.479300 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://06408f7526caaaeae759484ccf3ff85a146655a7d51ff7049c7be79b39fe96ba" gracePeriod=600 Mar 13 11:02:40 crc kubenswrapper[4632]: I0313 11:02:40.806598 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="06408f7526caaaeae759484ccf3ff85a146655a7d51ff7049c7be79b39fe96ba" exitCode=0 Mar 13 11:02:40 crc kubenswrapper[4632]: I0313 11:02:40.806666 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"06408f7526caaaeae759484ccf3ff85a146655a7d51ff7049c7be79b39fe96ba"} Mar 13 11:02:40 crc kubenswrapper[4632]: I0313 11:02:40.809183 4632 scope.go:117] "RemoveContainer" containerID="d62fcc7d7dd37c1e59dee28bd69ab3bfac7e5412873fdfe93b8d8f0639424c9d" Mar 13 11:02:41 crc kubenswrapper[4632]: I0313 11:02:41.818695 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582"} Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.010437 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mm6fq"] Mar 13 11:02:54 crc kubenswrapper[4632]: E0313 11:02:54.012465 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62b0f696-9e5c-4535-a181-fa2f4b645711" containerName="oc" Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.012514 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="62b0f696-9e5c-4535-a181-fa2f4b645711" containerName="oc" Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.012965 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="62b0f696-9e5c-4535-a181-fa2f4b645711" containerName="oc" Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.022019 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.126361 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mm6fq"] Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.183092 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9tfk7\" (UniqueName: \"kubernetes.io/projected/269ac923-f4f9-43f2-934f-8b0f26f6c4af-kube-api-access-9tfk7\") pod \"redhat-operators-mm6fq\" (UID: \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\") " pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.183253 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/269ac923-f4f9-43f2-934f-8b0f26f6c4af-catalog-content\") pod \"redhat-operators-mm6fq\" (UID: \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\") " pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.183275 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/269ac923-f4f9-43f2-934f-8b0f26f6c4af-utilities\") pod \"redhat-operators-mm6fq\" (UID: \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\") " pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.285095 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/269ac923-f4f9-43f2-934f-8b0f26f6c4af-catalog-content\") pod \"redhat-operators-mm6fq\" (UID: \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\") " pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.285466 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/269ac923-f4f9-43f2-934f-8b0f26f6c4af-utilities\") pod \"redhat-operators-mm6fq\" (UID: \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\") " pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.285563 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9tfk7\" (UniqueName: \"kubernetes.io/projected/269ac923-f4f9-43f2-934f-8b0f26f6c4af-kube-api-access-9tfk7\") pod \"redhat-operators-mm6fq\" (UID: \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\") " pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.394140 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/269ac923-f4f9-43f2-934f-8b0f26f6c4af-utilities\") pod \"redhat-operators-mm6fq\" (UID: \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\") " pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.448901 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/269ac923-f4f9-43f2-934f-8b0f26f6c4af-catalog-content\") pod \"redhat-operators-mm6fq\" (UID: \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\") " pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.503779 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9tfk7\" (UniqueName: \"kubernetes.io/projected/269ac923-f4f9-43f2-934f-8b0f26f6c4af-kube-api-access-9tfk7\") pod \"redhat-operators-mm6fq\" (UID: \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\") " pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:02:54 crc kubenswrapper[4632]: I0313 11:02:54.670161 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:02:56 crc kubenswrapper[4632]: I0313 11:02:56.406099 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mm6fq"] Mar 13 11:02:56 crc kubenswrapper[4632]: W0313 11:02:56.447815 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod269ac923_f4f9_43f2_934f_8b0f26f6c4af.slice/crio-c53342fac7b10f1bef54be90bcf0e83cc2e423f561f4ea27cefd68e4947d5bb4 WatchSource:0}: Error finding container c53342fac7b10f1bef54be90bcf0e83cc2e423f561f4ea27cefd68e4947d5bb4: Status 404 returned error can't find the container with id c53342fac7b10f1bef54be90bcf0e83cc2e423f561f4ea27cefd68e4947d5bb4 Mar 13 11:02:56 crc kubenswrapper[4632]: I0313 11:02:56.981602 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6fq" event={"ID":"269ac923-f4f9-43f2-934f-8b0f26f6c4af","Type":"ContainerDied","Data":"93ddcb9911b3bbd33b20a1520c077d1ce20ed42dceb52f18631471d802d7e139"} Mar 13 11:02:56 crc kubenswrapper[4632]: I0313 11:02:56.982657 4632 generic.go:334] "Generic (PLEG): container finished" podID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerID="93ddcb9911b3bbd33b20a1520c077d1ce20ed42dceb52f18631471d802d7e139" exitCode=0 Mar 13 11:02:56 crc kubenswrapper[4632]: I0313 11:02:56.983723 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6fq" event={"ID":"269ac923-f4f9-43f2-934f-8b0f26f6c4af","Type":"ContainerStarted","Data":"c53342fac7b10f1bef54be90bcf0e83cc2e423f561f4ea27cefd68e4947d5bb4"} Mar 13 11:02:56 crc kubenswrapper[4632]: I0313 11:02:56.988395 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 11:03:00 crc kubenswrapper[4632]: I0313 11:03:00.028101 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6fq" event={"ID":"269ac923-f4f9-43f2-934f-8b0f26f6c4af","Type":"ContainerStarted","Data":"d03ceb30503b22a2ad94cc53347c3b0ae54c134bb2b9db1bd0c47dcfc27a8ece"} Mar 13 11:03:09 crc kubenswrapper[4632]: I0313 11:03:09.239060 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6fq" event={"ID":"269ac923-f4f9-43f2-934f-8b0f26f6c4af","Type":"ContainerDied","Data":"d03ceb30503b22a2ad94cc53347c3b0ae54c134bb2b9db1bd0c47dcfc27a8ece"} Mar 13 11:03:09 crc kubenswrapper[4632]: I0313 11:03:09.239561 4632 generic.go:334] "Generic (PLEG): container finished" podID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerID="d03ceb30503b22a2ad94cc53347c3b0ae54c134bb2b9db1bd0c47dcfc27a8ece" exitCode=0 Mar 13 11:03:11 crc kubenswrapper[4632]: I0313 11:03:11.262130 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6fq" event={"ID":"269ac923-f4f9-43f2-934f-8b0f26f6c4af","Type":"ContainerStarted","Data":"67b531d65834622b374c34e759c46150ba93cade0961705aa2b576c0c27e19d2"} Mar 13 11:03:11 crc kubenswrapper[4632]: I0313 11:03:11.306393 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mm6fq" podStartSLOduration=5.142205059 podStartE2EDuration="18.303378772s" podCreationTimestamp="2026-03-13 11:02:53 +0000 UTC" firstStartedPulling="2026-03-13 11:02:56.984219177 +0000 UTC m=+3551.006749310" lastFinishedPulling="2026-03-13 11:03:10.14539289 +0000 UTC m=+3564.167923023" observedRunningTime="2026-03-13 11:03:11.301077695 +0000 UTC m=+3565.323607828" watchObservedRunningTime="2026-03-13 11:03:11.303378772 +0000 UTC m=+3565.325908905" Mar 13 11:03:12 crc kubenswrapper[4632]: I0313 11:03:12.006167 4632 scope.go:117] "RemoveContainer" containerID="6dd075c6962fa13da67ea22e1c7e0f24f4fdd06a675abd3b301b6ea671a2f51e" Mar 13 11:03:14 crc kubenswrapper[4632]: I0313 11:03:14.671738 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:03:14 crc kubenswrapper[4632]: I0313 11:03:14.672371 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:03:15 crc kubenswrapper[4632]: I0313 11:03:15.734487 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:03:15 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:03:15 crc kubenswrapper[4632]: > Mar 13 11:03:25 crc kubenswrapper[4632]: I0313 11:03:25.753742 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:03:25 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:03:25 crc kubenswrapper[4632]: > Mar 13 11:03:35 crc kubenswrapper[4632]: I0313 11:03:35.725626 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:03:35 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:03:35 crc kubenswrapper[4632]: > Mar 13 11:03:45 crc kubenswrapper[4632]: I0313 11:03:45.768324 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:03:45 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:03:45 crc kubenswrapper[4632]: > Mar 13 11:03:55 crc kubenswrapper[4632]: I0313 11:03:55.736845 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:03:55 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:03:55 crc kubenswrapper[4632]: > Mar 13 11:04:01 crc kubenswrapper[4632]: I0313 11:04:01.342186 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556664-vgmtg"] Mar 13 11:04:01 crc kubenswrapper[4632]: I0313 11:04:01.375829 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556664-vgmtg" Mar 13 11:04:01 crc kubenswrapper[4632]: I0313 11:04:01.459644 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:04:01 crc kubenswrapper[4632]: I0313 11:04:01.459647 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:04:01 crc kubenswrapper[4632]: I0313 11:04:01.459653 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:04:01 crc kubenswrapper[4632]: I0313 11:04:01.558458 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqslx\" (UniqueName: \"kubernetes.io/projected/6ecca46c-1e06-43be-bacc-eae4a1a474b7-kube-api-access-xqslx\") pod \"auto-csr-approver-29556664-vgmtg\" (UID: \"6ecca46c-1e06-43be-bacc-eae4a1a474b7\") " pod="openshift-infra/auto-csr-approver-29556664-vgmtg" Mar 13 11:04:01 crc kubenswrapper[4632]: I0313 11:04:01.661263 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqslx\" (UniqueName: \"kubernetes.io/projected/6ecca46c-1e06-43be-bacc-eae4a1a474b7-kube-api-access-xqslx\") pod \"auto-csr-approver-29556664-vgmtg\" (UID: \"6ecca46c-1e06-43be-bacc-eae4a1a474b7\") " pod="openshift-infra/auto-csr-approver-29556664-vgmtg" Mar 13 11:04:01 crc kubenswrapper[4632]: I0313 11:04:01.914145 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqslx\" (UniqueName: \"kubernetes.io/projected/6ecca46c-1e06-43be-bacc-eae4a1a474b7-kube-api-access-xqslx\") pod \"auto-csr-approver-29556664-vgmtg\" (UID: \"6ecca46c-1e06-43be-bacc-eae4a1a474b7\") " pod="openshift-infra/auto-csr-approver-29556664-vgmtg" Mar 13 11:04:02 crc kubenswrapper[4632]: I0313 11:04:02.041550 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556664-vgmtg" Mar 13 11:04:02 crc kubenswrapper[4632]: I0313 11:04:02.504859 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556664-vgmtg"] Mar 13 11:04:04 crc kubenswrapper[4632]: I0313 11:04:04.031615 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556664-vgmtg"] Mar 13 11:04:04 crc kubenswrapper[4632]: I0313 11:04:04.748845 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556664-vgmtg" event={"ID":"6ecca46c-1e06-43be-bacc-eae4a1a474b7","Type":"ContainerStarted","Data":"280361caa31f92ed27e2f7c50d8a879ce6f0bb804fc4144396a778455ffd2cf2"} Mar 13 11:04:05 crc kubenswrapper[4632]: I0313 11:04:05.746012 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:04:05 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:04:05 crc kubenswrapper[4632]: > Mar 13 11:04:06 crc kubenswrapper[4632]: I0313 11:04:06.773565 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556664-vgmtg" event={"ID":"6ecca46c-1e06-43be-bacc-eae4a1a474b7","Type":"ContainerStarted","Data":"5e0a7ac81434eac7eff8520645fc1fc30caa50af82d06bce9d4415863d0b9aa2"} Mar 13 11:04:06 crc kubenswrapper[4632]: I0313 11:04:06.874377 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556664-vgmtg" podStartSLOduration=5.884281338 podStartE2EDuration="6.872581259s" podCreationTimestamp="2026-03-13 11:04:00 +0000 UTC" firstStartedPulling="2026-03-13 11:04:04.100679674 +0000 UTC m=+3618.123209807" lastFinishedPulling="2026-03-13 11:04:05.088979595 +0000 UTC m=+3619.111509728" observedRunningTime="2026-03-13 11:04:06.865171526 +0000 UTC m=+3620.887701699" watchObservedRunningTime="2026-03-13 11:04:06.872581259 +0000 UTC m=+3620.895111412" Mar 13 11:04:08 crc kubenswrapper[4632]: I0313 11:04:08.791636 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556664-vgmtg" event={"ID":"6ecca46c-1e06-43be-bacc-eae4a1a474b7","Type":"ContainerDied","Data":"5e0a7ac81434eac7eff8520645fc1fc30caa50af82d06bce9d4415863d0b9aa2"} Mar 13 11:04:08 crc kubenswrapper[4632]: I0313 11:04:08.793756 4632 generic.go:334] "Generic (PLEG): container finished" podID="6ecca46c-1e06-43be-bacc-eae4a1a474b7" containerID="5e0a7ac81434eac7eff8520645fc1fc30caa50af82d06bce9d4415863d0b9aa2" exitCode=0 Mar 13 11:04:10 crc kubenswrapper[4632]: I0313 11:04:10.954649 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556664-vgmtg" Mar 13 11:04:11 crc kubenswrapper[4632]: I0313 11:04:11.068830 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqslx\" (UniqueName: \"kubernetes.io/projected/6ecca46c-1e06-43be-bacc-eae4a1a474b7-kube-api-access-xqslx\") pod \"6ecca46c-1e06-43be-bacc-eae4a1a474b7\" (UID: \"6ecca46c-1e06-43be-bacc-eae4a1a474b7\") " Mar 13 11:04:11 crc kubenswrapper[4632]: I0313 11:04:11.127599 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ecca46c-1e06-43be-bacc-eae4a1a474b7-kube-api-access-xqslx" (OuterVolumeSpecName: "kube-api-access-xqslx") pod "6ecca46c-1e06-43be-bacc-eae4a1a474b7" (UID: "6ecca46c-1e06-43be-bacc-eae4a1a474b7"). InnerVolumeSpecName "kube-api-access-xqslx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:04:11 crc kubenswrapper[4632]: I0313 11:04:11.172114 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqslx\" (UniqueName: \"kubernetes.io/projected/6ecca46c-1e06-43be-bacc-eae4a1a474b7-kube-api-access-xqslx\") on node \"crc\" DevicePath \"\"" Mar 13 11:04:11 crc kubenswrapper[4632]: I0313 11:04:11.829357 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556664-vgmtg" event={"ID":"6ecca46c-1e06-43be-bacc-eae4a1a474b7","Type":"ContainerDied","Data":"280361caa31f92ed27e2f7c50d8a879ce6f0bb804fc4144396a778455ffd2cf2"} Mar 13 11:04:11 crc kubenswrapper[4632]: I0313 11:04:11.829589 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556664-vgmtg" Mar 13 11:04:11 crc kubenswrapper[4632]: I0313 11:04:11.830324 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="280361caa31f92ed27e2f7c50d8a879ce6f0bb804fc4144396a778455ffd2cf2" Mar 13 11:04:12 crc kubenswrapper[4632]: I0313 11:04:12.109528 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556658-mbpfd"] Mar 13 11:04:12 crc kubenswrapper[4632]: I0313 11:04:12.118748 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556658-mbpfd"] Mar 13 11:04:14 crc kubenswrapper[4632]: I0313 11:04:14.056894 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4db08ac3-f768-407d-a321-ed9032c5c015" path="/var/lib/kubelet/pods/4db08ac3-f768-407d-a321-ed9032c5c015/volumes" Mar 13 11:04:15 crc kubenswrapper[4632]: I0313 11:04:15.721319 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:04:15 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:04:15 crc kubenswrapper[4632]: > Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.042450 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6cbqr"] Mar 13 11:04:19 crc kubenswrapper[4632]: E0313 11:04:19.053066 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ecca46c-1e06-43be-bacc-eae4a1a474b7" containerName="oc" Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.053128 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ecca46c-1e06-43be-bacc-eae4a1a474b7" containerName="oc" Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.055485 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ecca46c-1e06-43be-bacc-eae4a1a474b7" containerName="oc" Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.063262 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.175647 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6cbqr"] Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.239321 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-utilities\") pod \"community-operators-6cbqr\" (UID: \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\") " pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.239641 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4gb7\" (UniqueName: \"kubernetes.io/projected/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-kube-api-access-d4gb7\") pod \"community-operators-6cbqr\" (UID: \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\") " pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.239717 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-catalog-content\") pod \"community-operators-6cbqr\" (UID: \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\") " pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.341910 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4gb7\" (UniqueName: \"kubernetes.io/projected/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-kube-api-access-d4gb7\") pod \"community-operators-6cbqr\" (UID: \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\") " pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.342033 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-catalog-content\") pod \"community-operators-6cbqr\" (UID: \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\") " pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.342109 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-utilities\") pod \"community-operators-6cbqr\" (UID: \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\") " pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.349894 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-catalog-content\") pod \"community-operators-6cbqr\" (UID: \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\") " pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.351014 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-utilities\") pod \"community-operators-6cbqr\" (UID: \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\") " pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.397641 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4gb7\" (UniqueName: \"kubernetes.io/projected/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-kube-api-access-d4gb7\") pod \"community-operators-6cbqr\" (UID: \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\") " pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:04:19 crc kubenswrapper[4632]: I0313 11:04:19.417848 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:04:21 crc kubenswrapper[4632]: I0313 11:04:21.082388 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6cbqr"] Mar 13 11:04:21 crc kubenswrapper[4632]: I0313 11:04:21.929011 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6cbqr" event={"ID":"1c07de1e-84e2-4dae-a3c3-ced19801c8c2","Type":"ContainerDied","Data":"c772f3e31ea57b90bb99ca7ef746eba3e104a41a227d8c675887f2261f06ab48"} Mar 13 11:04:21 crc kubenswrapper[4632]: I0313 11:04:21.931027 4632 generic.go:334] "Generic (PLEG): container finished" podID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerID="c772f3e31ea57b90bb99ca7ef746eba3e104a41a227d8c675887f2261f06ab48" exitCode=0 Mar 13 11:04:21 crc kubenswrapper[4632]: I0313 11:04:21.931140 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6cbqr" event={"ID":"1c07de1e-84e2-4dae-a3c3-ced19801c8c2","Type":"ContainerStarted","Data":"2dcd2355772a6d6f5fad347e41d6b1e9baff67bf2bdf1773631d76e760c8ca38"} Mar 13 11:04:22 crc kubenswrapper[4632]: I0313 11:04:22.951667 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6cbqr" event={"ID":"1c07de1e-84e2-4dae-a3c3-ced19801c8c2","Type":"ContainerStarted","Data":"c4e602b48052cce5414a1759fd3d99f56ebde469321edc3b351de56e308a589e"} Mar 13 11:04:25 crc kubenswrapper[4632]: I0313 11:04:25.751530 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:04:25 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:04:25 crc kubenswrapper[4632]: > Mar 13 11:04:25 crc kubenswrapper[4632]: I0313 11:04:25.982310 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6cbqr" event={"ID":"1c07de1e-84e2-4dae-a3c3-ced19801c8c2","Type":"ContainerDied","Data":"c4e602b48052cce5414a1759fd3d99f56ebde469321edc3b351de56e308a589e"} Mar 13 11:04:25 crc kubenswrapper[4632]: I0313 11:04:25.982169 4632 generic.go:334] "Generic (PLEG): container finished" podID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerID="c4e602b48052cce5414a1759fd3d99f56ebde469321edc3b351de56e308a589e" exitCode=0 Mar 13 11:04:26 crc kubenswrapper[4632]: I0313 11:04:26.995839 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6cbqr" event={"ID":"1c07de1e-84e2-4dae-a3c3-ced19801c8c2","Type":"ContainerStarted","Data":"e7435d3cb1416970cf6b4162802419aa1bcf01c76e855132a1393d7b353e8c78"} Mar 13 11:04:27 crc kubenswrapper[4632]: I0313 11:04:27.038204 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6cbqr" podStartSLOduration=4.357588645 podStartE2EDuration="9.036865848s" podCreationTimestamp="2026-03-13 11:04:18 +0000 UTC" firstStartedPulling="2026-03-13 11:04:21.933619966 +0000 UTC m=+3635.956150099" lastFinishedPulling="2026-03-13 11:04:26.612897169 +0000 UTC m=+3640.635427302" observedRunningTime="2026-03-13 11:04:27.035536915 +0000 UTC m=+3641.058067048" watchObservedRunningTime="2026-03-13 11:04:27.036865848 +0000 UTC m=+3641.059395981" Mar 13 11:04:29 crc kubenswrapper[4632]: I0313 11:04:29.418558 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:04:29 crc kubenswrapper[4632]: I0313 11:04:29.419068 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:04:30 crc kubenswrapper[4632]: I0313 11:04:30.481955 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-6cbqr" podUID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerName="registry-server" probeResult="failure" output=< Mar 13 11:04:30 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:04:30 crc kubenswrapper[4632]: > Mar 13 11:04:35 crc kubenswrapper[4632]: I0313 11:04:35.724988 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:04:35 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:04:35 crc kubenswrapper[4632]: > Mar 13 11:04:40 crc kubenswrapper[4632]: I0313 11:04:40.461614 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:04:40 crc kubenswrapper[4632]: I0313 11:04:40.464249 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:04:40 crc kubenswrapper[4632]: I0313 11:04:40.547538 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-6cbqr" podUID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerName="registry-server" probeResult="failure" output=< Mar 13 11:04:40 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:04:40 crc kubenswrapper[4632]: > Mar 13 11:04:45 crc kubenswrapper[4632]: I0313 11:04:45.726097 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:04:45 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:04:45 crc kubenswrapper[4632]: > Mar 13 11:04:45 crc kubenswrapper[4632]: I0313 11:04:45.731461 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:04:45 crc kubenswrapper[4632]: I0313 11:04:45.735651 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"67b531d65834622b374c34e759c46150ba93cade0961705aa2b576c0c27e19d2"} pod="openshift-marketplace/redhat-operators-mm6fq" containerMessage="Container registry-server failed startup probe, will be restarted" Mar 13 11:04:45 crc kubenswrapper[4632]: I0313 11:04:45.736727 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" containerID="cri-o://67b531d65834622b374c34e759c46150ba93cade0961705aa2b576c0c27e19d2" gracePeriod=30 Mar 13 11:04:50 crc kubenswrapper[4632]: I0313 11:04:50.251582 4632 generic.go:334] "Generic (PLEG): container finished" podID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerID="67b531d65834622b374c34e759c46150ba93cade0961705aa2b576c0c27e19d2" exitCode=0 Mar 13 11:04:50 crc kubenswrapper[4632]: I0313 11:04:50.252250 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6fq" event={"ID":"269ac923-f4f9-43f2-934f-8b0f26f6c4af","Type":"ContainerDied","Data":"67b531d65834622b374c34e759c46150ba93cade0961705aa2b576c0c27e19d2"} Mar 13 11:04:50 crc kubenswrapper[4632]: I0313 11:04:50.535247 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-6cbqr" podUID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerName="registry-server" probeResult="failure" output=< Mar 13 11:04:50 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:04:50 crc kubenswrapper[4632]: > Mar 13 11:04:51 crc kubenswrapper[4632]: I0313 11:04:51.263255 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6fq" event={"ID":"269ac923-f4f9-43f2-934f-8b0f26f6c4af","Type":"ContainerStarted","Data":"0bbbe65ea71f36a37f33d902708fe700b15c322c14c94c121a3ca523a54d026b"} Mar 13 11:04:54 crc kubenswrapper[4632]: I0313 11:04:54.692074 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:04:54 crc kubenswrapper[4632]: I0313 11:04:54.697461 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:04:55 crc kubenswrapper[4632]: I0313 11:04:55.737331 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:04:55 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:04:55 crc kubenswrapper[4632]: > Mar 13 11:04:59 crc kubenswrapper[4632]: I0313 11:04:59.636109 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:04:59 crc kubenswrapper[4632]: I0313 11:04:59.761512 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:05:00 crc kubenswrapper[4632]: I0313 11:05:00.766639 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6cbqr"] Mar 13 11:05:01 crc kubenswrapper[4632]: I0313 11:05:01.383657 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6cbqr" podUID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerName="registry-server" containerID="cri-o://e7435d3cb1416970cf6b4162802419aa1bcf01c76e855132a1393d7b353e8c78" gracePeriod=2 Mar 13 11:05:02 crc kubenswrapper[4632]: I0313 11:05:02.401814 4632 generic.go:334] "Generic (PLEG): container finished" podID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerID="e7435d3cb1416970cf6b4162802419aa1bcf01c76e855132a1393d7b353e8c78" exitCode=0 Mar 13 11:05:02 crc kubenswrapper[4632]: I0313 11:05:02.402472 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6cbqr" event={"ID":"1c07de1e-84e2-4dae-a3c3-ced19801c8c2","Type":"ContainerDied","Data":"e7435d3cb1416970cf6b4162802419aa1bcf01c76e855132a1393d7b353e8c78"} Mar 13 11:05:03 crc kubenswrapper[4632]: I0313 11:05:03.778604 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:03 crc kubenswrapper[4632]: I0313 11:05:03.778603 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:04 crc kubenswrapper[4632]: I0313 11:05:04.452316 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" podUID="712b2002-4fce-4983-926a-99a4b2dc7a8c" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.48:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:04 crc kubenswrapper[4632]: I0313 11:05:04.452322 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" podUID="712b2002-4fce-4983-926a-99a4b2dc7a8c" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.48:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:04 crc kubenswrapper[4632]: I0313 11:05:04.802953 4632 patch_prober.go:28] interesting pod/console-5678554f8b-n7dcv container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.45:8443/health\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:04 crc kubenswrapper[4632]: I0313 11:05:04.820893 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-5678554f8b-n7dcv" podUID="a59bb7d3-da4a-4275-9dcb-b851215a9cd0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.45:8443/health\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:05 crc kubenswrapper[4632]: I0313 11:05:05.221170 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:05 crc kubenswrapper[4632]: I0313 11:05:05.320109 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" podUID="2d221857-ee77-4165-a351-ecd5fc424970" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.84:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:05 crc kubenswrapper[4632]: I0313 11:05:05.489085 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:05 crc kubenswrapper[4632]: I0313 11:05:05.489442 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:05 crc kubenswrapper[4632]: I0313 11:05:05.489096 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:05 crc kubenswrapper[4632]: I0313 11:05:05.489516 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:05 crc kubenswrapper[4632]: I0313 11:05:05.895756 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:05 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:05 crc kubenswrapper[4632]: > Mar 13 11:05:06 crc kubenswrapper[4632]: I0313 11:05:06.203082 4632 patch_prober.go:28] interesting pod/controller-manager-7469657588-kpf64 container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:06 crc kubenswrapper[4632]: I0313 11:05:06.203140 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" podUID="a8ff14f9-e25c-4839-acab-a622f6f70f88" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:06 crc kubenswrapper[4632]: I0313 11:05:06.204332 4632 patch_prober.go:28] interesting pod/controller-manager-7469657588-kpf64 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:06 crc kubenswrapper[4632]: I0313 11:05:06.204371 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" podUID="a8ff14f9-e25c-4839-acab-a622f6f70f88" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:07 crc kubenswrapper[4632]: I0313 11:05:07.191046 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:07 crc kubenswrapper[4632]: I0313 11:05:07.197138 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:07 crc kubenswrapper[4632]: I0313 11:05:07.204101 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:07 crc kubenswrapper[4632]: I0313 11:05:07.204185 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:07 crc kubenswrapper[4632]: I0313 11:05:07.310158 4632 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-svhr5 container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:07 crc kubenswrapper[4632]: I0313 11:05:07.310248 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" podUID="353e9ca9-cb3b-4c6e-b1ca-446611a12dca" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:07 crc kubenswrapper[4632]: I0313 11:05:07.564200 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:07 crc kubenswrapper[4632]: I0313 11:05:07.564276 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:07 crc kubenswrapper[4632]: I0313 11:05:07.564220 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:07 crc kubenswrapper[4632]: I0313 11:05:07.564341 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:09 crc kubenswrapper[4632]: I0313 11:05:09.408182 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" podUID="9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.82:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:09 crc kubenswrapper[4632]: E0313 11:05:09.432307 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e7435d3cb1416970cf6b4162802419aa1bcf01c76e855132a1393d7b353e8c78 is running failed: container process not found" containerID="e7435d3cb1416970cf6b4162802419aa1bcf01c76e855132a1393d7b353e8c78" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 11:05:09 crc kubenswrapper[4632]: E0313 11:05:09.433498 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e7435d3cb1416970cf6b4162802419aa1bcf01c76e855132a1393d7b353e8c78 is running failed: container process not found" containerID="e7435d3cb1416970cf6b4162802419aa1bcf01c76e855132a1393d7b353e8c78" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 11:05:09 crc kubenswrapper[4632]: E0313 11:05:09.434212 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e7435d3cb1416970cf6b4162802419aa1bcf01c76e855132a1393d7b353e8c78 is running failed: container process not found" containerID="e7435d3cb1416970cf6b4162802419aa1bcf01c76e855132a1393d7b353e8c78" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 11:05:09 crc kubenswrapper[4632]: E0313 11:05:09.434264 4632 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e7435d3cb1416970cf6b4162802419aa1bcf01c76e855132a1393d7b353e8c78 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-6cbqr" podUID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerName="registry-server" Mar 13 11:05:10 crc kubenswrapper[4632]: I0313 11:05:10.476471 4632 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:10 crc kubenswrapper[4632]: I0313 11:05:10.487172 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:05:10 crc kubenswrapper[4632]: I0313 11:05:10.512984 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:05:10 crc kubenswrapper[4632]: I0313 11:05:10.513207 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:11 crc kubenswrapper[4632]: I0313 11:05:11.333366 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-2jqnk" podUID="7de02b7f-4e1c-4ba1-9659-c864e9080092" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:11 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:11 crc kubenswrapper[4632]: > Mar 13 11:05:11 crc kubenswrapper[4632]: I0313 11:05:11.333647 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-2jqnk" podUID="7de02b7f-4e1c-4ba1-9659-c864e9080092" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:11 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:11 crc kubenswrapper[4632]: > Mar 13 11:05:11 crc kubenswrapper[4632]: I0313 11:05:11.460117 4632 patch_prober.go:28] interesting pod/oauth-openshift-75bb75cfd7-8sh2x container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:11 crc kubenswrapper[4632]: I0313 11:05:11.460138 4632 patch_prober.go:28] interesting pod/oauth-openshift-75bb75cfd7-8sh2x container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:11 crc kubenswrapper[4632]: I0313 11:05:11.460182 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" podUID="48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:11 crc kubenswrapper[4632]: I0313 11:05:11.460197 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" podUID="48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:12 crc kubenswrapper[4632]: I0313 11:05:12.181078 4632 trace.go:236] Trace[978839734]: "Calculate volume metrics of registry-storage for pod openshift-image-registry/image-registry-66df7c8f76-sbzcm" (13-Mar-2026 11:05:10.654) (total time: 1498ms): Mar 13 11:05:12 crc kubenswrapper[4632]: Trace[978839734]: [1.498091103s] [1.498091103s] END Mar 13 11:05:12 crc kubenswrapper[4632]: I0313 11:05:12.409056 4632 scope.go:117] "RemoveContainer" containerID="6a752c085ec4dd2121b36385f753ab45221d95dd428ca910155d9e3c585e4dbc" Mar 13 11:05:13 crc kubenswrapper[4632]: I0313 11:05:13.878001 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:13 crc kubenswrapper[4632]: I0313 11:05:13.877998 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:14 crc kubenswrapper[4632]: I0313 11:05:14.471720 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" podUID="712b2002-4fce-4983-926a-99a4b2dc7a8c" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.48:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:14 crc kubenswrapper[4632]: I0313 11:05:14.471738 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" podUID="712b2002-4fce-4983-926a-99a4b2dc7a8c" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.48:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:14 crc kubenswrapper[4632]: I0313 11:05:14.693478 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" podUID="1542a9c8-92f6-4bc9-8231-829f649b0b8f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.67:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:14 crc kubenswrapper[4632]: I0313 11:05:14.693496 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" podUID="1542a9c8-92f6-4bc9-8231-829f649b0b8f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.67:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:15 crc kubenswrapper[4632]: I0313 11:05:15.073695 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6cbqr" event={"ID":"1c07de1e-84e2-4dae-a3c3-ced19801c8c2","Type":"ContainerDied","Data":"2dcd2355772a6d6f5fad347e41d6b1e9baff67bf2bdf1773631d76e760c8ca38"} Mar 13 11:05:15 crc kubenswrapper[4632]: I0313 11:05:15.074420 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dcd2355772a6d6f5fad347e41d6b1e9baff67bf2bdf1773631d76e760c8ca38" Mar 13 11:05:15 crc kubenswrapper[4632]: I0313 11:05:15.073672 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:05:15 crc kubenswrapper[4632]: I0313 11:05:15.114895 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-utilities\") pod \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\" (UID: \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\") " Mar 13 11:05:15 crc kubenswrapper[4632]: I0313 11:05:15.115153 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4gb7\" (UniqueName: \"kubernetes.io/projected/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-kube-api-access-d4gb7\") pod \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\" (UID: \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\") " Mar 13 11:05:15 crc kubenswrapper[4632]: I0313 11:05:15.115331 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-catalog-content\") pod \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\" (UID: \"1c07de1e-84e2-4dae-a3c3-ced19801c8c2\") " Mar 13 11:05:15 crc kubenswrapper[4632]: I0313 11:05:15.149735 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-utilities" (OuterVolumeSpecName: "utilities") pod "1c07de1e-84e2-4dae-a3c3-ced19801c8c2" (UID: "1c07de1e-84e2-4dae-a3c3-ced19801c8c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:05:15 crc kubenswrapper[4632]: I0313 11:05:15.161537 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-kube-api-access-d4gb7" (OuterVolumeSpecName: "kube-api-access-d4gb7") pod "1c07de1e-84e2-4dae-a3c3-ced19801c8c2" (UID: "1c07de1e-84e2-4dae-a3c3-ced19801c8c2"). InnerVolumeSpecName "kube-api-access-d4gb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:05:15 crc kubenswrapper[4632]: I0313 11:05:15.218160 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:05:15 crc kubenswrapper[4632]: I0313 11:05:15.218200 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4gb7\" (UniqueName: \"kubernetes.io/projected/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-kube-api-access-d4gb7\") on node \"crc\" DevicePath \"\"" Mar 13 11:05:15 crc kubenswrapper[4632]: I0313 11:05:15.246152 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:15 crc kubenswrapper[4632]: I0313 11:05:15.265511 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1c07de1e-84e2-4dae-a3c3-ced19801c8c2" (UID: "1c07de1e-84e2-4dae-a3c3-ced19801c8c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:05:15 crc kubenswrapper[4632]: I0313 11:05:15.335831 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1c07de1e-84e2-4dae-a3c3-ced19801c8c2-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:05:15 crc kubenswrapper[4632]: I0313 11:05:15.790379 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:15 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:15 crc kubenswrapper[4632]: > Mar 13 11:05:16 crc kubenswrapper[4632]: I0313 11:05:16.173664 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6cbqr" Mar 13 11:05:16 crc kubenswrapper[4632]: I0313 11:05:16.876999 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6cbqr"] Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.049338 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6cbqr"] Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.305209 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.305299 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.337158 4632 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-svhr5 container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.350807 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" podUID="353e9ca9-cb3b-4c6e-b1ca-446611a12dca" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.350720 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.350871 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.566297 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.566310 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.566721 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.566666 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.940257 4632 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-xfvsc container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.940603 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" podUID="4e100e6e-7259-4262-be47-9c2b5be7a53a" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.940261 4632 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-xfvsc container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:17 crc kubenswrapper[4632]: I0313 11:05:17.940710 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" podUID="4e100e6e-7259-4262-be47-9c2b5be7a53a" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:18 crc kubenswrapper[4632]: I0313 11:05:18.117450 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" path="/var/lib/kubelet/pods/1c07de1e-84e2-4dae-a3c3-ced19801c8c2/volumes" Mar 13 11:05:18 crc kubenswrapper[4632]: I0313 11:05:18.647257 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" podUID="68c5eb80-4214-42c5-a08d-de6012969621" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.58:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:18 crc kubenswrapper[4632]: I0313 11:05:18.689224 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" podUID="68c5eb80-4214-42c5-a08d-de6012969621" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.58:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:18 crc kubenswrapper[4632]: I0313 11:05:18.785154 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" podUID="ff6d4dcb-9eb8-44fc-951e-f2aecd77a639" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.62:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:18 crc kubenswrapper[4632]: I0313 11:05:18.785212 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" podUID="ff6d4dcb-9eb8-44fc-951e-f2aecd77a639" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.62:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:19 crc kubenswrapper[4632]: I0313 11:05:19.643063 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" podUID="f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.75:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:19 crc kubenswrapper[4632]: I0313 11:05:19.643075 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" podUID="f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.75:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:19 crc kubenswrapper[4632]: I0313 11:05:19.938197 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" podUID="ee081327-4c3f-4c0a-9085-71085c6487b5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.88:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:19 crc kubenswrapper[4632]: I0313 11:05:19.938234 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" podUID="ee081327-4c3f-4c0a-9085-71085c6487b5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.88:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:20 crc kubenswrapper[4632]: I0313 11:05:20.431360 4632 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:20 crc kubenswrapper[4632]: I0313 11:05:20.480901 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:21 crc kubenswrapper[4632]: I0313 11:05:21.382220 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-2jqnk" podUID="7de02b7f-4e1c-4ba1-9659-c864e9080092" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:21 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:21 crc kubenswrapper[4632]: > Mar 13 11:05:21 crc kubenswrapper[4632]: I0313 11:05:21.388906 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-2jqnk" podUID="7de02b7f-4e1c-4ba1-9659-c864e9080092" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:21 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:21 crc kubenswrapper[4632]: > Mar 13 11:05:21 crc kubenswrapper[4632]: I0313 11:05:21.451181 4632 patch_prober.go:28] interesting pod/oauth-openshift-75bb75cfd7-8sh2x container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:21 crc kubenswrapper[4632]: I0313 11:05:21.451158 4632 patch_prober.go:28] interesting pod/oauth-openshift-75bb75cfd7-8sh2x container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:21 crc kubenswrapper[4632]: I0313 11:05:21.451253 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" podUID="48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:21 crc kubenswrapper[4632]: I0313 11:05:21.451302 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" podUID="48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:25 crc kubenswrapper[4632]: I0313 11:05:25.248916 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:25 crc kubenswrapper[4632]: I0313 11:05:25.309163 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/frr-k8s-lvlxj" Mar 13 11:05:25 crc kubenswrapper[4632]: I0313 11:05:25.320349 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="frr" containerStatusID={"Type":"cri-o","ID":"59d96a138cf7adeb4d273db270ca9998a9b75447d7d6c92e875e751afba3f9b8"} pod="metallb-system/frr-k8s-lvlxj" containerMessage="Container frr failed liveness probe, will be restarted" Mar 13 11:05:25 crc kubenswrapper[4632]: I0313 11:05:25.321830 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="frr" containerID="cri-o://59d96a138cf7adeb4d273db270ca9998a9b75447d7d6c92e875e751afba3f9b8" gracePeriod=2 Mar 13 11:05:25 crc kubenswrapper[4632]: I0313 11:05:25.765604 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:25 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:25 crc kubenswrapper[4632]: > Mar 13 11:05:26 crc kubenswrapper[4632]: I0313 11:05:26.351969 4632 generic.go:334] "Generic (PLEG): container finished" podID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerID="59d96a138cf7adeb4d273db270ca9998a9b75447d7d6c92e875e751afba3f9b8" exitCode=143 Mar 13 11:05:26 crc kubenswrapper[4632]: I0313 11:05:26.351989 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lvlxj" event={"ID":"85b58bb0-63f5-4c85-8759-ce28d2c7db58","Type":"ContainerDied","Data":"59d96a138cf7adeb4d273db270ca9998a9b75447d7d6c92e875e751afba3f9b8"} Mar 13 11:05:27 crc kubenswrapper[4632]: I0313 11:05:27.497435 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lvlxj" event={"ID":"85b58bb0-63f5-4c85-8759-ce28d2c7db58","Type":"ContainerStarted","Data":"39cdf7120397ce6bc7ca222c9f810a74a54099eea3fa7559bd511f04a2d4ba8e"} Mar 13 11:05:29 crc kubenswrapper[4632]: I0313 11:05:29.212165 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-lvlxj" Mar 13 11:05:29 crc kubenswrapper[4632]: I0313 11:05:29.644812 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" podUID="e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.86:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:30 crc kubenswrapper[4632]: I0313 11:05:30.221185 4632 prober.go:107] "Probe failed" probeType="Startup" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:30 crc kubenswrapper[4632]: I0313 11:05:30.441852 4632 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:30 crc kubenswrapper[4632]: I0313 11:05:30.450132 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:30 crc kubenswrapper[4632]: I0313 11:05:30.450538 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 11:05:30 crc kubenswrapper[4632]: I0313 11:05:30.457212 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-scheduler" containerStatusID={"Type":"cri-o","ID":"ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1"} pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" containerMessage="Container kube-scheduler failed liveness probe, will be restarted" Mar 13 11:05:30 crc kubenswrapper[4632]: I0313 11:05:30.459351 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" containerID="cri-o://ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1" gracePeriod=30 Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.355228 4632 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": dial tcp 192.168.126.11:10259: connect: connection refused" start-of-body= Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.355295 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": dial tcp 192.168.126.11:10259: connect: connection refused" Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.367818 4632 patch_prober.go:28] interesting pod/oauth-openshift-75bb75cfd7-8sh2x container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.367832 4632 patch_prober.go:28] interesting pod/oauth-openshift-75bb75cfd7-8sh2x container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.367885 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" podUID="48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.367888 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" podUID="48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.375614 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.375654 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.378906 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="oauth-openshift" containerStatusID={"Type":"cri-o","ID":"224837f104bcdbc6545d62209161e349a9d07cdcaf5c66e47c1de75b3af4b369"} pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" containerMessage="Container oauth-openshift failed liveness probe, will be restarted" Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.425584 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-2jqnk" podUID="7de02b7f-4e1c-4ba1-9659-c864e9080092" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:31 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:31 crc kubenswrapper[4632]: > Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.425902 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.425855 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-2jqnk" podUID="7de02b7f-4e1c-4ba1-9659-c864e9080092" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:31 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:31 crc kubenswrapper[4632]: > Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.426056 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.429874 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"956a137089f814594898450d78be9fb64aa26d046a610f28da0c9756520f90c8"} pod="openstack-operators/openstack-operator-index-2jqnk" containerMessage="Container registry-server failed liveness probe, will be restarted" Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.429918 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-2jqnk" podUID="7de02b7f-4e1c-4ba1-9659-c864e9080092" containerName="registry-server" containerID="cri-o://956a137089f814594898450d78be9fb64aa26d046a610f28da0c9756520f90c8" gracePeriod=30 Mar 13 11:05:31 crc kubenswrapper[4632]: E0313 11:05:31.461135 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="956a137089f814594898450d78be9fb64aa26d046a610f28da0c9756520f90c8" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 11:05:31 crc kubenswrapper[4632]: E0313 11:05:31.470815 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="956a137089f814594898450d78be9fb64aa26d046a610f28da0c9756520f90c8" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 11:05:31 crc kubenswrapper[4632]: E0313 11:05:31.472764 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="956a137089f814594898450d78be9fb64aa26d046a610f28da0c9756520f90c8" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 11:05:31 crc kubenswrapper[4632]: E0313 11:05:31.472826 4632 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack-operators/openstack-operator-index-2jqnk" podUID="7de02b7f-4e1c-4ba1-9659-c864e9080092" containerName="registry-server" Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.967184 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" podUID="a0d52d98-fe87-4bc8-890e-5c5efb1f30d6" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.81:6080/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:31 crc kubenswrapper[4632]: I0313 11:05:31.967668 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" podUID="a0d52d98-fe87-4bc8-890e-5c5efb1f30d6" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.81:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:32 crc kubenswrapper[4632]: I0313 11:05:32.175228 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" podUID="3fdb377f-5a78-4687-82e1-50718514290d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.91:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:32 crc kubenswrapper[4632]: I0313 11:05:32.175244 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" podUID="3fdb377f-5a78-4687-82e1-50718514290d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.91:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:32 crc kubenswrapper[4632]: I0313 11:05:32.417223 4632 patch_prober.go:28] interesting pod/oauth-openshift-75bb75cfd7-8sh2x container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:32 crc kubenswrapper[4632]: I0313 11:05:32.417315 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" podUID="48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:32 crc kubenswrapper[4632]: I0313 11:05:32.697054 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1"} Mar 13 11:05:32 crc kubenswrapper[4632]: I0313 11:05:32.697102 4632 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="ba32b01674f111328387620c028407e11c9d44b3154c3dce8f415f79e1db54c1" exitCode=0 Mar 13 11:05:32 crc kubenswrapper[4632]: I0313 11:05:32.883959 4632 patch_prober.go:28] interesting pod/route-controller-manager-db6b8fbf8-pllt2 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:32 crc kubenswrapper[4632]: I0313 11:05:32.884033 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" podUID="2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:32 crc kubenswrapper[4632]: I0313 11:05:32.884390 4632 patch_prober.go:28] interesting pod/route-controller-manager-db6b8fbf8-pllt2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:32 crc kubenswrapper[4632]: I0313 11:05:32.884452 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" podUID="2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:33 crc kubenswrapper[4632]: I0313 11:05:33.720188 4632 generic.go:334] "Generic (PLEG): container finished" podID="7de02b7f-4e1c-4ba1-9659-c864e9080092" containerID="956a137089f814594898450d78be9fb64aa26d046a610f28da0c9756520f90c8" exitCode=0 Mar 13 11:05:33 crc kubenswrapper[4632]: I0313 11:05:33.720421 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2jqnk" event={"ID":"7de02b7f-4e1c-4ba1-9659-c864e9080092","Type":"ContainerDied","Data":"956a137089f814594898450d78be9fb64aa26d046a610f28da0c9756520f90c8"} Mar 13 11:05:33 crc kubenswrapper[4632]: I0313 11:05:33.756029 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:33 crc kubenswrapper[4632]: I0313 11:05:33.758629 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:34 crc kubenswrapper[4632]: I0313 11:05:34.738030 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a3aae7b1e1166a2d7565d7709da6786bb6123dbabf32f4ef389dd47642569340"} Mar 13 11:05:34 crc kubenswrapper[4632]: I0313 11:05:34.739444 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 11:05:34 crc kubenswrapper[4632]: I0313 11:05:34.756405 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="1761ca69-46fd-4375-af60-22b3e77c19a2" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:34 crc kubenswrapper[4632]: I0313 11:05:34.759799 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="1761ca69-46fd-4375-af60-22b3e77c19a2" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:34 crc kubenswrapper[4632]: I0313 11:05:34.812150 4632 patch_prober.go:28] interesting pod/console-5678554f8b-n7dcv container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.45:8443/health\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:34 crc kubenswrapper[4632]: I0313 11:05:34.844788 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-5678554f8b-n7dcv" podUID="a59bb7d3-da4a-4275-9dcb-b851215a9cd0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.45:8443/health\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:35 crc kubenswrapper[4632]: I0313 11:05:35.303123 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:35 crc kubenswrapper[4632]: I0313 11:05:35.303695 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:35 crc kubenswrapper[4632]: I0313 11:05:35.303741 4632 prober.go:107] "Probe failed" probeType="Startup" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:35 crc kubenswrapper[4632]: I0313 11:05:35.321146 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" podUID="2d221857-ee77-4165-a351-ecd5fc424970" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.84:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:35 crc kubenswrapper[4632]: I0313 11:05:35.754807 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:35 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:35 crc kubenswrapper[4632]: > Mar 13 11:05:35 crc kubenswrapper[4632]: I0313 11:05:35.777545 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-2jqnk" event={"ID":"7de02b7f-4e1c-4ba1-9659-c864e9080092","Type":"ContainerStarted","Data":"fa1f92cf71967f1ece5dd0d584d63e1d80dc100a2f3b28938056aa32105d1b6c"} Mar 13 11:05:35 crc kubenswrapper[4632]: I0313 11:05:35.875192 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" podUID="b33bccd8-6f28-4ffe-9500-069a52aab5df" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.50:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:35 crc kubenswrapper[4632]: I0313 11:05:35.875205 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" podUID="b33bccd8-6f28-4ffe-9500-069a52aab5df" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.50:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:36 crc kubenswrapper[4632]: I0313 11:05:36.202284 4632 patch_prober.go:28] interesting pod/controller-manager-7469657588-kpf64 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:36 crc kubenswrapper[4632]: I0313 11:05:36.202379 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" podUID="a8ff14f9-e25c-4839-acab-a622f6f70f88" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:36 crc kubenswrapper[4632]: I0313 11:05:36.212760 4632 patch_prober.go:28] interesting pod/controller-manager-7469657588-kpf64 container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:36 crc kubenswrapper[4632]: I0313 11:05:36.213099 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" podUID="a8ff14f9-e25c-4839-acab-a622f6f70f88" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:36 crc kubenswrapper[4632]: I0313 11:05:36.297511 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" podUID="09ddc697-7ac1-4896-b9e2-1ae6c59c6f47" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 11:05:36 crc kubenswrapper[4632]: I0313 11:05:36.811466 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-tztd9" podUID="8f51973a-596d-40dc-9b5b-b2c95a60ea0c" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:36 crc kubenswrapper[4632]: I0313 11:05:36.811491 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-tztd9" podUID="8f51973a-596d-40dc-9b5b-b2c95a60ea0c" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:36 crc kubenswrapper[4632]: I0313 11:05:36.928092 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" podUID="a0d52d98-fe87-4bc8-890e-5c5efb1f30d6" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.81:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.100962 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.101326 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.101080 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.101386 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.223918 4632 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-r5v5p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.224024 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" podUID="32f62e32-732b-4646-85f0-45b8ea6544a6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.265121 4632 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-r5v5p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.265187 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" podUID="32f62e32-732b-4646-85f0-45b8ea6544a6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.348107 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.348169 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.390176 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.390270 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.390178 4632 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-svhr5 container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.390384 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" podUID="353e9ca9-cb3b-4c6e-b1ca-446611a12dca" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.564204 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.564269 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.565414 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:37 crc kubenswrapper[4632]: I0313 11:05:37.565468 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:38 crc kubenswrapper[4632]: I0313 11:05:38.695804 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" podUID="68c5eb80-4214-42c5-a08d-de6012969621" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.58:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:38 crc kubenswrapper[4632]: I0313 11:05:38.695807 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" podUID="68c5eb80-4214-42c5-a08d-de6012969621" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.58:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:38 crc kubenswrapper[4632]: I0313 11:05:38.758759 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Mar 13 11:05:38 crc kubenswrapper[4632]: I0313 11:05:38.815336 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 11:05:38 crc kubenswrapper[4632]: I0313 11:05:38.815575 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 11:05:39 crc kubenswrapper[4632]: I0313 11:05:39.264009 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" podUID="f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.75:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:39 crc kubenswrapper[4632]: I0313 11:05:39.265540 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" podUID="f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.75:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:39 crc kubenswrapper[4632]: I0313 11:05:39.411347 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" podUID="d04e9aa6-f234-4ffa-81e2-1a2407addb77" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.83:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:39 crc kubenswrapper[4632]: I0313 11:05:39.411363 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" podUID="d04e9aa6-f234-4ffa-81e2-1a2407addb77" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.83:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:39 crc kubenswrapper[4632]: I0313 11:05:39.487776 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" podUID="9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.82:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:39 crc kubenswrapper[4632]: I0313 11:05:39.488191 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" podUID="9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.82:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:39 crc kubenswrapper[4632]: I0313 11:05:39.543277 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" podUID="0a9d48f4-d68b-4ef9-826e-ed619c761405" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.85:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:39 crc kubenswrapper[4632]: I0313 11:05:39.543336 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" podUID="0a9d48f4-d68b-4ef9-826e-ed619c761405" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.85:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:40 crc kubenswrapper[4632]: I0313 11:05:40.181066 4632 prober.go:107] "Probe failed" probeType="Startup" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:40 crc kubenswrapper[4632]: I0313 11:05:40.376031 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 11:05:40 crc kubenswrapper[4632]: I0313 11:05:40.460951 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:05:40 crc kubenswrapper[4632]: I0313 11:05:40.462528 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:05:40 crc kubenswrapper[4632]: I0313 11:05:40.462593 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 11:05:40 crc kubenswrapper[4632]: I0313 11:05:40.467180 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 11:05:40 crc kubenswrapper[4632]: I0313 11:05:40.468394 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" gracePeriod=600 Mar 13 11:05:41 crc kubenswrapper[4632]: E0313 11:05:41.342066 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:05:41 crc kubenswrapper[4632]: I0313 11:05:41.440661 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:41 crc kubenswrapper[4632]: I0313 11:05:41.441038 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:41 crc kubenswrapper[4632]: I0313 11:05:41.442460 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:41 crc kubenswrapper[4632]: I0313 11:05:41.442547 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:41 crc kubenswrapper[4632]: I0313 11:05:41.869793 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582"} Mar 13 11:05:41 crc kubenswrapper[4632]: I0313 11:05:41.870404 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" exitCode=0 Mar 13 11:05:41 crc kubenswrapper[4632]: I0313 11:05:41.874444 4632 scope.go:117] "RemoveContainer" containerID="06408f7526caaaeae759484ccf3ff85a146655a7d51ff7049c7be79b39fe96ba" Mar 13 11:05:41 crc kubenswrapper[4632]: I0313 11:05:41.874621 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:05:41 crc kubenswrapper[4632]: E0313 11:05:41.875040 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:05:41 crc kubenswrapper[4632]: I0313 11:05:41.891273 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-774lb" podUID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:41 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:41 crc kubenswrapper[4632]: > Mar 13 11:05:41 crc kubenswrapper[4632]: I0313 11:05:41.891273 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-774lb" podUID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:41 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:41 crc kubenswrapper[4632]: > Mar 13 11:05:42 crc kubenswrapper[4632]: I0313 11:05:42.290691 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" podUID="09ddc697-7ac1-4896-b9e2-1ae6c59c6f47" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 11:05:42 crc kubenswrapper[4632]: I0313 11:05:42.780625 4632 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-d9n25 container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.70:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:42 crc kubenswrapper[4632]: I0313 11:05:42.781763 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" podUID="023be687-a773-401c-981b-e3d7136f53b6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.70:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:42 crc kubenswrapper[4632]: I0313 11:05:42.780797 4632 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-d9n25 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.70:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:42 crc kubenswrapper[4632]: I0313 11:05:42.781866 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-d9n25" podUID="023be687-a773-401c-981b-e3d7136f53b6" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.70:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:42 crc kubenswrapper[4632]: I0313 11:05:42.890011 4632 patch_prober.go:28] interesting pod/route-controller-manager-db6b8fbf8-pllt2 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:42 crc kubenswrapper[4632]: I0313 11:05:42.890005 4632 patch_prober.go:28] interesting pod/route-controller-manager-db6b8fbf8-pllt2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:42 crc kubenswrapper[4632]: I0313 11:05:42.890078 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" podUID="2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:42 crc kubenswrapper[4632]: I0313 11:05:42.890107 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" podUID="2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:43 crc kubenswrapper[4632]: I0313 11:05:43.100533 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack-operators/openstack-operator-index-2jqnk" podUID="7de02b7f-4e1c-4ba1-9659-c864e9080092" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:43 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:43 crc kubenswrapper[4632]: > Mar 13 11:05:43 crc kubenswrapper[4632]: I0313 11:05:43.757981 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:43 crc kubenswrapper[4632]: I0313 11:05:43.757981 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:43 crc kubenswrapper[4632]: I0313 11:05:43.760587 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Mar 13 11:05:43 crc kubenswrapper[4632]: I0313 11:05:43.798352 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" podUID="09ddc697-7ac1-4896-b9e2-1ae6c59c6f47" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 11:05:44 crc kubenswrapper[4632]: I0313 11:05:44.475173 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" podUID="712b2002-4fce-4983-926a-99a4b2dc7a8c" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.48:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:44 crc kubenswrapper[4632]: I0313 11:05:44.475200 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" podUID="712b2002-4fce-4983-926a-99a4b2dc7a8c" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.48:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:44 crc kubenswrapper[4632]: I0313 11:05:44.558712 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:44 crc kubenswrapper[4632]: I0313 11:05:44.558782 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:44 crc kubenswrapper[4632]: I0313 11:05:44.558712 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:44 crc kubenswrapper[4632]: I0313 11:05:44.558828 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:44 crc kubenswrapper[4632]: I0313 11:05:44.713200 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" podUID="1542a9c8-92f6-4bc9-8231-829f649b0b8f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.67:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:44 crc kubenswrapper[4632]: I0313 11:05:44.714797 4632 patch_prober.go:28] interesting pod/nmstate-webhook-5f558f5558-gcngd container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.32:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:44 crc kubenswrapper[4632]: I0313 11:05:44.714876 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" podUID="9bf11778-d854-4c97-acd1-ed4822ee5f47" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.32:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:44 crc kubenswrapper[4632]: I0313 11:05:44.756389 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="1761ca69-46fd-4375-af60-22b3e77c19a2" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:44 crc kubenswrapper[4632]: I0313 11:05:44.756460 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="1761ca69-46fd-4375-af60-22b3e77c19a2" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:44 crc kubenswrapper[4632]: I0313 11:05:44.758480 4632 patch_prober.go:28] interesting pod/console-5678554f8b-n7dcv container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.45:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:44 crc kubenswrapper[4632]: I0313 11:05:44.758516 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-5678554f8b-n7dcv" podUID="a59bb7d3-da4a-4275-9dcb-b851215a9cd0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.45:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:45 crc kubenswrapper[4632]: I0313 11:05:45.303338 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:45 crc kubenswrapper[4632]: I0313 11:05:45.303858 4632 prober.go:107] "Probe failed" probeType="Startup" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:45 crc kubenswrapper[4632]: I0313 11:05:45.303969 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:45 crc kubenswrapper[4632]: I0313 11:05:45.345209 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" podUID="2d221857-ee77-4165-a351-ecd5fc424970" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.84:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:45 crc kubenswrapper[4632]: I0313 11:05:45.874235 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" podUID="b33bccd8-6f28-4ffe-9500-069a52aab5df" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.50:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:45 crc kubenswrapper[4632]: I0313 11:05:45.874660 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" podUID="b33bccd8-6f28-4ffe-9500-069a52aab5df" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.50:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:45 crc kubenswrapper[4632]: I0313 11:05:45.899812 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:45 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:45 crc kubenswrapper[4632]: > Mar 13 11:05:46 crc kubenswrapper[4632]: I0313 11:05:46.202447 4632 patch_prober.go:28] interesting pod/controller-manager-7469657588-kpf64 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:46 crc kubenswrapper[4632]: I0313 11:05:46.202529 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" podUID="a8ff14f9-e25c-4839-acab-a622f6f70f88" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:46 crc kubenswrapper[4632]: I0313 11:05:46.203643 4632 patch_prober.go:28] interesting pod/controller-manager-7469657588-kpf64 container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:46 crc kubenswrapper[4632]: I0313 11:05:46.203699 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" podUID="a8ff14f9-e25c-4839-acab-a622f6f70f88" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:46 crc kubenswrapper[4632]: I0313 11:05:46.861643 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-tztd9" podUID="8f51973a-596d-40dc-9b5b-b2c95a60ea0c" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:46 crc kubenswrapper[4632]: I0313 11:05:46.863412 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-tztd9" podUID="8f51973a-596d-40dc-9b5b-b2c95a60ea0c" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:46 crc kubenswrapper[4632]: I0313 11:05:46.926213 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" podUID="a0d52d98-fe87-4bc8-890e-5c5efb1f30d6" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.81:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.182189 4632 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-r5v5p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.182270 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" podUID="32f62e32-732b-4646-85f0-45b8ea6544a6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.182343 4632 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-r5v5p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.182396 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" podUID="32f62e32-732b-4646-85f0-45b8ea6544a6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.188556 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.188637 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.189271 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.189365 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.273573 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.273639 4632 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-svhr5 container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.273955 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" podUID="353e9ca9-cb3b-4c6e-b1ca-446611a12dca" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.273868 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.273731 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.274036 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.438201 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.438277 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.438154 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.438418 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.439616 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.439707 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.444157 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="openshift-config-operator" containerStatusID={"Type":"cri-o","ID":"e98f0e8253db82d7fc1c628a628a0d9ea91c85c3796f3abe0d968983b3e782e2"} pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" containerMessage="Container openshift-config-operator failed liveness probe, will be restarted" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.444605 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" containerID="cri-o://e98f0e8253db82d7fc1c628a628a0d9ea91c85c3796f3abe0d968983b3e782e2" gracePeriod=30 Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.564835 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.564890 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.564991 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.564908 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.945094 4632 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-tqbl9 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.945439 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" podUID="ebf1040d-57dd-47ef-b839-6f78a7c5c75f" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.945093 4632 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-xfvsc container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.945511 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" podUID="4e100e6e-7259-4262-be47-9c2b5be7a53a" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.945147 4632 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-tqbl9 container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.945581 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" podUID="ebf1040d-57dd-47ef-b839-6f78a7c5c75f" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.945204 4632 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-xfvsc container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:47 crc kubenswrapper[4632]: I0313 11:05:47.945809 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" podUID="4e100e6e-7259-4262-be47-9c2b5be7a53a" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:48 crc kubenswrapper[4632]: I0313 11:05:48.482175 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:48 crc kubenswrapper[4632]: I0313 11:05:48.482251 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:48 crc kubenswrapper[4632]: I0313 11:05:48.644241 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" podUID="68c5eb80-4214-42c5-a08d-de6012969621" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.58:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:48 crc kubenswrapper[4632]: I0313 11:05:48.685161 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87" podUID="3f3a462e-4d89-45b3-8611-181aca5f8558" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.60:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:48 crc kubenswrapper[4632]: I0313 11:05:48.759476 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Mar 13 11:05:48 crc kubenswrapper[4632]: I0313 11:05:48.759596 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Mar 13 11:05:48 crc kubenswrapper[4632]: I0313 11:05:48.761111 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Mar 13 11:05:48 crc kubenswrapper[4632]: I0313 11:05:48.775372 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"3da76186915cfbbbe688750a6110b1e64143d37e61c44ef62a9740eabb32c983"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Mar 13 11:05:48 crc kubenswrapper[4632]: I0313 11:05:48.775493 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-central-agent" containerID="cri-o://3da76186915cfbbbe688750a6110b1e64143d37e61c44ef62a9740eabb32c983" gracePeriod=30 Mar 13 11:05:48 crc kubenswrapper[4632]: I0313 11:05:48.776834 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" podUID="ff6d4dcb-9eb8-44fc-951e-f2aecd77a639" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.62:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:48 crc kubenswrapper[4632]: I0313 11:05:48.818194 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" podUID="9a963f9c-ac58-4e21-abfa-fca1279a192d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.64:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:48 crc kubenswrapper[4632]: I0313 11:05:48.862202 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" podUID="c8fc6f03-c43b-4ade-92a8-acc5537a4eeb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.74:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:48 crc kubenswrapper[4632]: I0313 11:05:48.903659 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq" podUID="82fe7ef6-50a5-41d4-9419-787812e16bd6" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.57:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:49 crc kubenswrapper[4632]: I0313 11:05:49.008164 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c" podUID="9040a0e0-2a56-4331-ba50-b19ff05ef0c0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.77:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:49 crc kubenswrapper[4632]: I0313 11:05:49.226091 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" podUID="f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.75:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:49 crc kubenswrapper[4632]: I0313 11:05:49.274085 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d" podUID="7b491335-6a73-46de-8098-f27ff4c6f795" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.76:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:49 crc kubenswrapper[4632]: I0313 11:05:49.370132 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" podUID="d04e9aa6-f234-4ffa-81e2-1a2407addb77" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.83:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:49 crc kubenswrapper[4632]: I0313 11:05:49.427182 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-62gpm" podUID="9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.82:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:49 crc kubenswrapper[4632]: I0313 11:05:49.440836 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 13 11:05:49 crc kubenswrapper[4632]: I0313 11:05:49.440906 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 13 11:05:49 crc kubenswrapper[4632]: I0313 11:05:49.543190 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" podUID="0a9d48f4-d68b-4ef9-826e-ed619c761405" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.85:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:49 crc kubenswrapper[4632]: I0313 11:05:49.623258 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" podUID="e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.86:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:49 crc kubenswrapper[4632]: I0313 11:05:49.760058 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-mpfnk" podUID="33445a2b-7fa8-4198-a60a-09caeb69b8ed" containerName="nmstate-handler" probeResult="failure" output="command timed out" Mar 13 11:05:49 crc kubenswrapper[4632]: I0313 11:05:49.897128 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" podUID="ee081327-4c3f-4c0a-9085-71085c6487b5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.88:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:50 crc kubenswrapper[4632]: I0313 11:05:50.105300 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" podUID="7bab78c8-7dac-48dc-a426-ccd4ae00a428" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.89:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:50 crc kubenswrapper[4632]: I0313 11:05:50.222131 4632 prober.go:107] "Probe failed" probeType="Startup" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:51 crc kubenswrapper[4632]: I0313 11:05:51.365504 4632 patch_prober.go:28] interesting pod/oauth-openshift-75bb75cfd7-8sh2x container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:51 crc kubenswrapper[4632]: I0313 11:05:51.365913 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" podUID="48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:51 crc kubenswrapper[4632]: I0313 11:05:51.757025 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-khrch" podUID="b6e936db-ec1c-447a-894d-49bd7c74c315" containerName="ovnkube-controller" probeResult="failure" output="command timed out" Mar 13 11:05:51 crc kubenswrapper[4632]: I0313 11:05:51.966200 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" podUID="a0d52d98-fe87-4bc8-890e-5c5efb1f30d6" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.81:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:51 crc kubenswrapper[4632]: I0313 11:05:51.966206 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" podUID="a0d52d98-fe87-4bc8-890e-5c5efb1f30d6" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.81:6080/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:52 crc kubenswrapper[4632]: I0313 11:05:52.048788 4632 generic.go:334] "Generic (PLEG): container finished" podID="f660255f-8f78-4876-973d-db58f2ee7020" containerID="e98f0e8253db82d7fc1c628a628a0d9ea91c85c3796f3abe0d968983b3e782e2" exitCode=0 Mar 13 11:05:52 crc kubenswrapper[4632]: I0313 11:05:52.057331 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" event={"ID":"f660255f-8f78-4876-973d-db58f2ee7020","Type":"ContainerDied","Data":"e98f0e8253db82d7fc1c628a628a0d9ea91c85c3796f3abe0d968983b3e782e2"} Mar 13 11:05:52 crc kubenswrapper[4632]: I0313 11:05:52.175128 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" podUID="3fdb377f-5a78-4687-82e1-50718514290d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.91:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:52 crc kubenswrapper[4632]: I0313 11:05:52.175290 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" podUID="3fdb377f-5a78-4687-82e1-50718514290d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.91:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:52 crc kubenswrapper[4632]: I0313 11:05:52.438364 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 13 11:05:52 crc kubenswrapper[4632]: I0313 11:05:52.439411 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 13 11:05:52 crc kubenswrapper[4632]: I0313 11:05:52.884810 4632 patch_prober.go:28] interesting pod/route-controller-manager-db6b8fbf8-pllt2 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:52 crc kubenswrapper[4632]: I0313 11:05:52.884885 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" podUID="2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:52 crc kubenswrapper[4632]: I0313 11:05:52.885167 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 11:05:52 crc kubenswrapper[4632]: I0313 11:05:52.885720 4632 patch_prober.go:28] interesting pod/route-controller-manager-db6b8fbf8-pllt2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:52 crc kubenswrapper[4632]: I0313 11:05:52.885840 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" podUID="2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:52 crc kubenswrapper[4632]: I0313 11:05:52.893544 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" podUID="09ddc697-7ac1-4896-b9e2-1ae6c59c6f47" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 11:05:52 crc kubenswrapper[4632]: I0313 11:05:52.903225 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="route-controller-manager" containerStatusID={"Type":"cri-o","ID":"71808a85287e54b9fb184ad4c73a074a1ff3d6b35824bd6122d42af589681e05"} pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" containerMessage="Container route-controller-manager failed liveness probe, will be restarted" Mar 13 11:05:52 crc kubenswrapper[4632]: I0313 11:05:52.904167 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" podUID="2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3" containerName="route-controller-manager" containerID="cri-o://71808a85287e54b9fb184ad4c73a074a1ff3d6b35824bd6122d42af589681e05" gracePeriod=30 Mar 13 11:05:53 crc kubenswrapper[4632]: I0313 11:05:53.601396 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-774lb" podUID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:53 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:53 crc kubenswrapper[4632]: > Mar 13 11:05:53 crc kubenswrapper[4632]: I0313 11:05:53.613184 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-774lb" podUID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:53 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:53 crc kubenswrapper[4632]: > Mar 13 11:05:53 crc kubenswrapper[4632]: I0313 11:05:53.688490 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="26ce3314-15f1-490c-83e5-a1c609212437" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.0.231:8080/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:53 crc kubenswrapper[4632]: I0313 11:05:53.755596 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:53 crc kubenswrapper[4632]: I0313 11:05:53.755623 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:53 crc kubenswrapper[4632]: I0313 11:05:53.759387 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack-operators/openstack-operator-index-2jqnk" podUID="7de02b7f-4e1c-4ba1-9659-c864e9080092" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 11:05:53 crc kubenswrapper[4632]: I0313 11:05:53.803662 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 13 11:05:53 crc kubenswrapper[4632]: I0313 11:05:53.803714 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/openstack-galera-0" Mar 13 11:05:53 crc kubenswrapper[4632]: I0313 11:05:53.818522 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="galera" containerStatusID={"Type":"cri-o","ID":"cdf295326a62a01129c4f9b5741f57b3d80d103e3c5a6bf64f5cc1951034264f"} pod="openstack/openstack-galera-0" containerMessage="Container galera failed liveness probe, will be restarted" Mar 13 11:05:53 crc kubenswrapper[4632]: I0313 11:05:53.897439 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" podUID="09ddc697-7ac1-4896-b9e2-1ae6c59c6f47" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.155775 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" event={"ID":"f660255f-8f78-4876-973d-db58f2ee7020","Type":"ContainerStarted","Data":"bdfa238d1dda3afead970f6c0c59d9c82cc9066974eef2637a5f643bcf655e99"} Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.167691 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.480326 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" podUID="712b2002-4fce-4983-926a-99a4b2dc7a8c" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.48:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.480369 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-6c7bf5ddc5-v6t5l" podUID="712b2002-4fce-4983-926a-99a4b2dc7a8c" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.48:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.753118 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" podUID="1542a9c8-92f6-4bc9-8231-829f649b0b8f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.67:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.753219 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-flfxh" podUID="1542a9c8-92f6-4bc9-8231-829f649b0b8f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.67:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.756894 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="1761ca69-46fd-4375-af60-22b3e77c19a2" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.756912 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="1761ca69-46fd-4375-af60-22b3e77c19a2" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.757020 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.757168 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.758536 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.754714 4632 patch_prober.go:28] interesting pod/nmstate-webhook-5f558f5558-gcngd container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.32:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.761190 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f558f5558-gcngd" podUID="9bf11778-d854-4c97-acd1-ed4822ee5f47" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.32:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.763385 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="galera" containerStatusID={"Type":"cri-o","ID":"8d6e3c3bf2cc94b4f346233606a1c5c55a2993e7644d8a78c77dc12972c98a9e"} pod="openstack/openstack-cell1-galera-0" containerMessage="Container galera failed liveness probe, will be restarted" Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.804464 4632 patch_prober.go:28] interesting pod/console-5678554f8b-n7dcv container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.45:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.804580 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-5678554f8b-n7dcv" podUID="a59bb7d3-da4a-4275-9dcb-b851215a9cd0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.45:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:54 crc kubenswrapper[4632]: I0313 11:05:54.804657 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.001350 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-gdt8x" podUID="f7f61b75-16bf-4c5a-be30-c88d155c203f" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:55 crc kubenswrapper[4632]: timeout: health rpc did not complete within 1s Mar 13 11:05:55 crc kubenswrapper[4632]: > Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.018657 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/certified-operators-7ksc5" podUID="0fa3faab-9e82-4fde-afff-3de6939a17d1" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:55 crc kubenswrapper[4632]: timeout: health rpc did not complete within 1s Mar 13 11:05:55 crc kubenswrapper[4632]: > Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.018773 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/certified-operators-7ksc5" podUID="0fa3faab-9e82-4fde-afff-3de6939a17d1" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:55 crc kubenswrapper[4632]: timeout: health rpc did not complete within 1s Mar 13 11:05:55 crc kubenswrapper[4632]: > Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.019321 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-gdt8x" podUID="f7f61b75-16bf-4c5a-be30-c88d155c203f" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:55 crc kubenswrapper[4632]: timeout: health rpc did not complete within 1s Mar 13 11:05:55 crc kubenswrapper[4632]: > Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.068562 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:05:55 crc kubenswrapper[4632]: E0313 11:05:55.075723 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.330263 4632 prober.go:107] "Probe failed" probeType="Startup" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.330323 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.330453 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-lvlxj" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.371219 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" podUID="2d221857-ee77-4165-a351-ecd5fc424970" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.84:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.412254 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.412332 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/frr-k8s-lvlxj" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.412372 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7v927j" podUID="2d221857-ee77-4165-a351-ecd5fc424970" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.84:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.495182 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-7bb4cc7c98-62bwr" podUID="277ddd7f-fd9c-4b27-9563-c904f1dffd40" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.51:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.495828 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-7bb4cc7c98-62bwr" podUID="277ddd7f-fd9c-4b27-9563-c904f1dffd40" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.51:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.756740 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="1761ca69-46fd-4375-af60-22b3e77c19a2" containerName="galera" probeResult="failure" output="command timed out" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.811712 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:05:55 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:05:55 crc kubenswrapper[4632]: > Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.875157 4632 patch_prober.go:28] interesting pod/console-5678554f8b-n7dcv container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.45:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.875189 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" podUID="b33bccd8-6f28-4ffe-9500-069a52aab5df" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.50:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.875222 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-5678554f8b-n7dcv" podUID="a59bb7d3-da4a-4275-9dcb-b851215a9cd0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.45:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.875260 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.875321 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" podUID="b33bccd8-6f28-4ffe-9500-069a52aab5df" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.50:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.875427 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.885173 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="frr-k8s-webhook-server" containerStatusID={"Type":"cri-o","ID":"6b277b3621566e90d2ea8a306394444270adbf026557398f5520284a63c356df"} pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" containerMessage="Container frr-k8s-webhook-server failed liveness probe, will be restarted" Mar 13 11:05:55 crc kubenswrapper[4632]: I0313 11:05:55.898766 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" podUID="b33bccd8-6f28-4ffe-9500-069a52aab5df" containerName="frr-k8s-webhook-server" containerID="cri-o://6b277b3621566e90d2ea8a306394444270adbf026557398f5520284a63c356df" gracePeriod=10 Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.172050 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="controller" containerStatusID={"Type":"cri-o","ID":"bfa960455f207db762d901f5af9c2b35ade8cd1c5f43d1bc1d4a40a5bfd8199d"} pod="metallb-system/frr-k8s-lvlxj" containerMessage="Container controller failed liveness probe, will be restarted" Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.172223 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="controller" containerID="cri-o://bfa960455f207db762d901f5af9c2b35ade8cd1c5f43d1bc1d4a40a5bfd8199d" gracePeriod=2 Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.201990 4632 patch_prober.go:28] interesting pod/controller-manager-7469657588-kpf64 container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.202087 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" podUID="a8ff14f9-e25c-4839-acab-a622f6f70f88" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.202156 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.202155 4632 patch_prober.go:28] interesting pod/controller-manager-7469657588-kpf64 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.202250 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" podUID="a8ff14f9-e25c-4839-acab-a622f6f70f88" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.203485 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="controller-manager" containerStatusID={"Type":"cri-o","ID":"432a739d763fd09cb52fbc4a7bbe481e0fb4c89b88f7822f73b594d3596d0d39"} pod="openshift-controller-manager/controller-manager-7469657588-kpf64" containerMessage="Container controller-manager failed liveness probe, will be restarted" Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.203532 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" podUID="a8ff14f9-e25c-4839-acab-a622f6f70f88" containerName="controller-manager" containerID="cri-o://432a739d763fd09cb52fbc4a7bbe481e0fb4c89b88f7822f73b594d3596d0d39" gracePeriod=30 Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.374155 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.709437 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" podUID="48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd" containerName="oauth-openshift" containerID="cri-o://224837f104bcdbc6545d62209161e349a9d07cdcaf5c66e47c1de75b3af4b369" gracePeriod=15 Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.892110 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-tztd9" podUID="8f51973a-596d-40dc-9b5b-b2c95a60ea0c" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.892123 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-tztd9" podUID="8f51973a-596d-40dc-9b5b-b2c95a60ea0c" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.893532 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-tztd9" Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.893595 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/speaker-tztd9" Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.900091 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="speaker" containerStatusID={"Type":"cri-o","ID":"7369028ab3380b8162926288f2a66e0780eba331066b6d04106bd606debba692"} pod="metallb-system/speaker-tztd9" containerMessage="Container speaker failed liveness probe, will be restarted" Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.900219 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/speaker-tztd9" podUID="8f51973a-596d-40dc-9b5b-b2c95a60ea0c" containerName="speaker" containerID="cri-o://7369028ab3380b8162926288f2a66e0780eba331066b6d04106bd606debba692" gracePeriod=2 Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.933106 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" podUID="b33bccd8-6f28-4ffe-9500-069a52aab5df" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.50:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.933369 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" podUID="a0d52d98-fe87-4bc8-890e-5c5efb1f30d6" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.81:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.933437 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" Mar 13 11:05:56 crc kubenswrapper[4632]: I0313 11:05:56.934464 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-tjkbb" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.017452 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-lvlxj" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.100359 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.100416 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.100459 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.100368 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.100504 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.100560 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.101247 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="console-operator" containerStatusID={"Type":"cri-o","ID":"faaf308e22c1a8d08431430b330cacf53efc9923cc70f0515be295533e608c79"} pod="openshift-console-operator/console-operator-58897d9998-sbtn5" containerMessage="Container console-operator failed liveness probe, will be restarted" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.101298 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" containerID="cri-o://faaf308e22c1a8d08431430b330cacf53efc9923cc70f0515be295533e608c79" gracePeriod=30 Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.182826 4632 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-r5v5p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.182880 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" podUID="32f62e32-732b-4646-85f0-45b8ea6544a6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.182897 4632 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-r5v5p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.182926 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.183003 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" podUID="32f62e32-732b-4646-85f0-45b8ea6544a6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.183076 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.192674 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="catalog-operator" containerStatusID={"Type":"cri-o","ID":"4f41aedb607002fa771d4b82bf1fb15a527c048ee3048ce7cd9db7dc1d8b7961"} pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" containerMessage="Container catalog-operator failed liveness probe, will be restarted" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.192751 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" podUID="32f62e32-732b-4646-85f0-45b8ea6544a6" containerName="catalog-operator" containerID="cri-o://4f41aedb607002fa771d4b82bf1fb15a527c048ee3048ce7cd9db7dc1d8b7961" gracePeriod=30 Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.222842 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" event={"ID":"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3","Type":"ContainerDied","Data":"71808a85287e54b9fb184ad4c73a074a1ff3d6b35824bd6122d42af589681e05"} Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.226359 4632 generic.go:334] "Generic (PLEG): container finished" podID="2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3" containerID="71808a85287e54b9fb184ad4c73a074a1ff3d6b35824bd6122d42af589681e05" exitCode=0 Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.279607 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.280066 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.280161 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.279924 4632 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-svhr5 container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.280253 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" podUID="353e9ca9-cb3b-4c6e-b1ca-446611a12dca" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.280323 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.279891 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.282095 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.282138 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.288381 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="router" containerStatusID={"Type":"cri-o","ID":"c0db1ffabe3d33862c8266179a821f8fd8c1a4906081849cc73b575a98544e3b"} pod="openshift-ingress/router-default-5444994796-t9vht" containerMessage="Container router failed liveness probe, will be restarted" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.288452 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" containerID="cri-o://c0db1ffabe3d33862c8266179a821f8fd8c1a4906081849cc73b575a98544e3b" gracePeriod=10 Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.290032 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="authentication-operator" containerStatusID={"Type":"cri-o","ID":"696b18c58833c0581e6bf36ae1881e00a6717c6dc6b1a5150c21fe634a2b6edb"} pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" containerMessage="Container authentication-operator failed liveness probe, will be restarted" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.290136 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" podUID="353e9ca9-cb3b-4c6e-b1ca-446611a12dca" containerName="authentication-operator" containerID="cri-o://696b18c58833c0581e6bf36ae1881e00a6717c6dc6b1a5150c21fe634a2b6edb" gracePeriod=30 Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.564913 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.565053 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.568125 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.575929 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="packageserver" containerStatusID={"Type":"cri-o","ID":"35b32e739ccce4a6f84a62ef541fb840a3cf0ce2a60fb788f618073e6f79bd60"} pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" containerMessage="Container packageserver failed liveness probe, will be restarted" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.576009 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" containerID="cri-o://35b32e739ccce4a6f84a62ef541fb840a3cf0ce2a60fb788f618073e6f79bd60" gracePeriod=30 Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.576182 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.576217 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.576328 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.696413 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-hvrrc" podUID="09ddc697-7ac1-4896-b9e2-1ae6c59c6f47" containerName="hostpath-provisioner" probeResult="failure" output="Get \"http://10.217.0.42:9898/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.947132 4632 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-tqbl9 container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.947168 4632 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-tqbl9 container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.37:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.947204 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" podUID="ebf1040d-57dd-47ef-b839-6f78a7c5c75f" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.947132 4632 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-xfvsc container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.947246 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" podUID="4e100e6e-7259-4262-be47-9c2b5be7a53a" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.947203 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tqbl9" podUID="ebf1040d-57dd-47ef-b839-6f78a7c5c75f" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.37:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.988165 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-tztd9" podUID="8f51973a-596d-40dc-9b5b-b2c95a60ea0c" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.988266 4632 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-xfvsc container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:57 crc kubenswrapper[4632]: I0313 11:05:57.988292 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-xfvsc" podUID="4e100e6e-7259-4262-be47-9c2b5be7a53a" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.33:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.146714 4632 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-r5v5p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": EOF" start-of-body= Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.147067 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" podUID="32f62e32-732b-4646-85f0-45b8ea6544a6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": EOF" Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.146749 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.147138 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.247741 4632 generic.go:334] "Generic (PLEG): container finished" podID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerID="bfa960455f207db762d901f5af9c2b35ade8cd1c5f43d1bc1d4a40a5bfd8199d" exitCode=0 Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.247829 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lvlxj" event={"ID":"85b58bb0-63f5-4c85-8759-ce28d2c7db58","Type":"ContainerDied","Data":"bfa960455f207db762d901f5af9c2b35ade8cd1c5f43d1bc1d4a40a5bfd8199d"} Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.252077 4632 generic.go:334] "Generic (PLEG): container finished" podID="b33bccd8-6f28-4ffe-9500-069a52aab5df" containerID="6b277b3621566e90d2ea8a306394444270adbf026557398f5520284a63c356df" exitCode=0 Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.252115 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" event={"ID":"b33bccd8-6f28-4ffe-9500-069a52aab5df","Type":"ContainerDied","Data":"6b277b3621566e90d2ea8a306394444270adbf026557398f5520284a63c356df"} Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.323433 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.323505 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.438930 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.438969 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.439017 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.439030 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.690500 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" podUID="68c5eb80-4214-42c5-a08d-de6012969621" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.58:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.690958 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.857177 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" podUID="68c5eb80-4214-42c5-a08d-de6012969621" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.58:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.857241 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.857190 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn" podUID="75d652c7-8521-4039-913a-fa625f89b094" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.63:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:58 crc kubenswrapper[4632]: I0313 11:05:58.939159 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l" podUID="20f92131-aca4-41ea-9144-a23bd9216f49" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.61:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.022096 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87" podUID="3f3a462e-4d89-45b3-8611-181aca5f8558" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.60:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.022162 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" podUID="ff6d4dcb-9eb8-44fc-951e-f2aecd77a639" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.62:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.104191 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-f6c87" podUID="3f3a462e-4d89-45b3-8611-181aca5f8558" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.60:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.104192 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" podUID="9a963f9c-ac58-4e21-abfa-fca1279a192d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.64:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.104504 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-cfcgn" podUID="75d652c7-8521-4039-913a-fa625f89b094" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.63:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.147756 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" podUID="c8fc6f03-c43b-4ade-92a8-acc5537a4eeb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.74:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.230109 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-qg79l" podUID="20f92131-aca4-41ea-9144-a23bd9216f49" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.61:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.230160 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq" podUID="82fe7ef6-50a5-41d4-9419-787812e16bd6" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.57:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.275398 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-lvlxj" event={"ID":"85b58bb0-63f5-4c85-8759-ce28d2c7db58","Type":"ContainerStarted","Data":"b22a68c5198c461a36773fe0b66eb17c50943b0a9d7a1785a9982b7ddc2598b3"} Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.275997 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-lvlxj" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.310481 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-58897d9998-sbtn5_ef269b18-ea84-43c2-971c-e772149acbf6/console-operator/0.log" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.310609 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" event={"ID":"ef269b18-ea84-43c2-971c-e772149acbf6","Type":"ContainerDied","Data":"faaf308e22c1a8d08431430b330cacf53efc9923cc70f0515be295533e608c79"} Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.310632 4632 generic.go:334] "Generic (PLEG): container finished" podID="ef269b18-ea84-43c2-971c-e772149acbf6" containerID="faaf308e22c1a8d08431430b330cacf53efc9923cc70f0515be295533e608c79" exitCode=1 Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.314145 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-cgh6c" podUID="ff6d4dcb-9eb8-44fc-951e-f2aecd77a639" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.62:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.314234 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c" podUID="9040a0e0-2a56-4331-ba50-b19ff05ef0c0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.77:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.315111 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-2rv7s" podUID="9a963f9c-ac58-4e21-abfa-fca1279a192d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.64:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.315363 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-wtzrw" podUID="c8fc6f03-c43b-4ade-92a8-acc5537a4eeb" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.74:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.319844 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" event={"ID":"2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3","Type":"ContainerStarted","Data":"4ec063236f8aa26a1d317386ccdd18403efd3b518cbac7f9c5b11ea9c585aba7"} Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.320180 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.324761 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" event={"ID":"b33bccd8-6f28-4ffe-9500-069a52aab5df","Type":"ContainerStarted","Data":"7b741b052f66780121aafd6b779efc40ca5030202933ec65b5ae41819bfe4649"} Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.325135 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.333687 4632 generic.go:334] "Generic (PLEG): container finished" podID="32f62e32-732b-4646-85f0-45b8ea6544a6" containerID="4f41aedb607002fa771d4b82bf1fb15a527c048ee3048ce7cd9db7dc1d8b7961" exitCode=0 Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.333887 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" event={"ID":"32f62e32-732b-4646-85f0-45b8ea6544a6","Type":"ContainerDied","Data":"4f41aedb607002fa771d4b82bf1fb15a527c048ee3048ce7cd9db7dc1d8b7961"} Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.333917 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" event={"ID":"32f62e32-732b-4646-85f0-45b8ea6544a6","Type":"ContainerStarted","Data":"2de143252a2a2cf135b056cea81707b81fee21a3cac7b6c26a55bce23c3d8eb4"} Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.334156 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.334498 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="manager" containerStatusID={"Type":"cri-o","ID":"579d286b9eb7e56fb8f1cb6d18127cc0ece5c920fbbbc7e2c67943e4800bb183"} pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" containerMessage="Container manager failed liveness probe, will be restarted" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.334536 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" podUID="68c5eb80-4214-42c5-a08d-de6012969621" containerName="manager" containerID="cri-o://579d286b9eb7e56fb8f1cb6d18127cc0ece5c920fbbbc7e2c67943e4800bb183" gracePeriod=10 Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.357151 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" podUID="f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.75:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.357224 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.358830 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="manager" containerStatusID={"Type":"cri-o","ID":"b79dde4b0109a751bfba6b9882a550b5aaf0de838fae99b2eeecdc581770755b"} pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" containerMessage="Container manager failed liveness probe, will be restarted" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.358880 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" podUID="f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda" containerName="manager" containerID="cri-o://b79dde4b0109a751bfba6b9882a550b5aaf0de838fae99b2eeecdc581770755b" gracePeriod=10 Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.399174 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" podUID="f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.75:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.399289 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.399287 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-init-865685cd99-ls9jq" podUID="82fe7ef6-50a5-41d4-9419-787812e16bd6" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.57:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.399932 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.481185 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-szd7c" podUID="9040a0e0-2a56-4331-ba50-b19ff05ef0c0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.77:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.481577 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d" podUID="7b491335-6a73-46de-8098-f27ff4c6f795" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.76:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.564155 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" podUID="d04e9aa6-f234-4ffa-81e2-1a2407addb77" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.83:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.564486 4632 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-r5v5p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.564514 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-sxw8d" podUID="7b491335-6a73-46de-8098-f27ff4c6f795" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.76:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.564578 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" podUID="32f62e32-732b-4646-85f0-45b8ea6544a6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.564670 4632 patch_prober.go:28] interesting pod/route-controller-manager-db6b8fbf8-pllt2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" start-of-body= Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.564767 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" podUID="2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.607669 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" podUID="d04e9aa6-f234-4ffa-81e2-1a2407addb77" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.83:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.607822 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.608412 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" podUID="0a9d48f4-d68b-4ef9-826e-ed619c761405" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.85:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.608518 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.693526 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" podUID="e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.86:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.693806 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" podUID="0a9d48f4-d68b-4ef9-826e-ed619c761405" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.85:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.694277 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-qkr9n" podUID="e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.86:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.696403 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-4m8kf" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.757133 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-mpfnk" podUID="33445a2b-7fa8-4198-a60a-09caeb69b8ed" containerName="nmstate-handler" probeResult="failure" output="command timed out" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.939085 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" podUID="ee081327-4c3f-4c0a-9085-71085c6487b5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.88:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:05:59 crc kubenswrapper[4632]: I0313 11:05:59.939088 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-6cd66dbd4b-nt7np" podUID="ee081327-4c3f-4c0a-9085-71085c6487b5" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.88:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.146150 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" podUID="7bab78c8-7dac-48dc-a426-ccd4ae00a428" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.89:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.146219 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" podUID="7bab78c8-7dac-48dc-a426-ccd4ae00a428" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.89:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.229096 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" podUID="e0d1d349-d63d-498b-ae15-3121f9ae73f8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.90:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.271095 4632 prober.go:107] "Probe failed" probeType="Startup" pod="metallb-system/frr-k8s-lvlxj" podUID="85b58bb0-63f5-4c85-8759-ce28d2c7db58" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.271097 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-kv8b2" podUID="e0d1d349-d63d-498b-ae15-3121f9ae73f8" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.90:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.356106 4632 generic.go:334] "Generic (PLEG): container finished" podID="a8ff14f9-e25c-4839-acab-a622f6f70f88" containerID="432a739d763fd09cb52fbc4a7bbe481e0fb4c89b88f7822f73b594d3596d0d39" exitCode=0 Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.356150 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" event={"ID":"a8ff14f9-e25c-4839-acab-a622f6f70f88","Type":"ContainerDied","Data":"432a739d763fd09cb52fbc4a7bbe481e0fb4c89b88f7822f73b594d3596d0d39"} Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.360357 4632 generic.go:334] "Generic (PLEG): container finished" podID="353e9ca9-cb3b-4c6e-b1ca-446611a12dca" containerID="696b18c58833c0581e6bf36ae1881e00a6717c6dc6b1a5150c21fe634a2b6edb" exitCode=0 Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.360423 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" event={"ID":"353e9ca9-cb3b-4c6e-b1ca-446611a12dca","Type":"ContainerDied","Data":"696b18c58833c0581e6bf36ae1881e00a6717c6dc6b1a5150c21fe634a2b6edb"} Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.367216 4632 patch_prober.go:28] interesting pod/oauth-openshift-75bb75cfd7-8sh2x container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": dial tcp 10.217.0.68:6443: connect: connection refused" start-of-body= Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.367255 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" podUID="48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": dial tcp 10.217.0.68:6443: connect: connection refused" Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.368128 4632 generic.go:334] "Generic (PLEG): container finished" podID="8f51973a-596d-40dc-9b5b-b2c95a60ea0c" containerID="7369028ab3380b8162926288f2a66e0780eba331066b6d04106bd606debba692" exitCode=137 Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.369422 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tztd9" event={"ID":"8f51973a-596d-40dc-9b5b-b2c95a60ea0c","Type":"ContainerDied","Data":"7369028ab3380b8162926288f2a66e0780eba331066b6d04106bd606debba692"} Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.370146 4632 patch_prober.go:28] interesting pod/route-controller-manager-db6b8fbf8-pllt2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" start-of-body= Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.370199 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" podUID="2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.370216 4632 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-r5v5p container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.370238 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" podUID="32f62e32-732b-4646-85f0-45b8ea6544a6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Mar 13 11:06:00 crc kubenswrapper[4632]: I0313 11:06:00.538428 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-628ss" Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.408203 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-58897d9998-sbtn5_ef269b18-ea84-43c2-971c-e772149acbf6/console-operator/0.log" Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.410121 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" event={"ID":"ef269b18-ea84-43c2-971c-e772149acbf6","Type":"ContainerStarted","Data":"6d6064b7502063ece3533259b5cf853c2915d2ee3e3d73d55543f1192104d84d"} Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.410153 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.421185 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.421237 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.426598 4632 generic.go:334] "Generic (PLEG): container finished" podID="49c520f1-fb05-48ca-8435-1985ce668451" containerID="35b32e739ccce4a6f84a62ef541fb840a3cf0ce2a60fb788f618073e6f79bd60" exitCode=0 Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.426677 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" event={"ID":"49c520f1-fb05-48ca-8435-1985ce668451","Type":"ContainerDied","Data":"35b32e739ccce4a6f84a62ef541fb840a3cf0ce2a60fb788f618073e6f79bd60"} Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.438579 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.438652 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.438582 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.438721 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.442831 4632 generic.go:334] "Generic (PLEG): container finished" podID="48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd" containerID="224837f104bcdbc6545d62209161e349a9d07cdcaf5c66e47c1de75b3af4b369" exitCode=0 Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.442969 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" event={"ID":"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd","Type":"ContainerDied","Data":"224837f104bcdbc6545d62209161e349a9d07cdcaf5c66e47c1de75b3af4b369"} Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.450437 4632 generic.go:334] "Generic (PLEG): container finished" podID="68c5eb80-4214-42c5-a08d-de6012969621" containerID="579d286b9eb7e56fb8f1cb6d18127cc0ece5c920fbbbc7e2c67943e4800bb183" exitCode=0 Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.450815 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" event={"ID":"68c5eb80-4214-42c5-a08d-de6012969621","Type":"ContainerDied","Data":"579d286b9eb7e56fb8f1cb6d18127cc0ece5c920fbbbc7e2c67943e4800bb183"} Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.884693 4632 patch_prober.go:28] interesting pod/route-controller-manager-db6b8fbf8-pllt2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" start-of-body= Mar 13 11:06:01 crc kubenswrapper[4632]: I0313 11:06:01.885125 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" podUID="2f5d4f7c-4d7b-4347-bd38-d5fd29fed3f3" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.65:8443/healthz\": dial tcp 10.217.0.65:8443: connect: connection refused" Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.043332 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="1761ca69-46fd-4375-af60-22b3e77c19a2" containerName="galera" containerID="cri-o://8d6e3c3bf2cc94b4f346233606a1c5c55a2993e7644d8a78c77dc12972c98a9e" gracePeriod=23 Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.076928 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" containerID="cri-o://cdf295326a62a01129c4f9b5741f57b3d80d103e3c5a6bf64f5cc1951034264f" gracePeriod=22 Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.134098 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-85c677895b-thbc4" podUID="3fdb377f-5a78-4687-82e1-50718514290d" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.91:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.463540 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-svhr5" event={"ID":"353e9ca9-cb3b-4c6e-b1ca-446611a12dca","Type":"ContainerStarted","Data":"1576ef1b44221dac835a1417f0e95e462cbd508e0e4f9c0f2281aeaa1ae366d8"} Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.472236 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" event={"ID":"49c520f1-fb05-48ca-8435-1985ce668451","Type":"ContainerStarted","Data":"3b3b129c9e469a5415a2dda08053f620658cbc9d22814b96f2371d28d2669ac4"} Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.473514 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.473591 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" start-of-body= Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.473618 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.476038 4632 generic.go:334] "Generic (PLEG): container finished" podID="f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda" containerID="b79dde4b0109a751bfba6b9882a550b5aaf0de838fae99b2eeecdc581770755b" exitCode=0 Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.476076 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" event={"ID":"f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda","Type":"ContainerDied","Data":"b79dde4b0109a751bfba6b9882a550b5aaf0de838fae99b2eeecdc581770755b"} Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.481724 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" event={"ID":"68c5eb80-4214-42c5-a08d-de6012969621","Type":"ContainerStarted","Data":"1147eea75904622c793225e4c48cb3f7549325cc9fa40143c5548037a2e7beba"} Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.484870 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" event={"ID":"a8ff14f9-e25c-4839-acab-a622f6f70f88","Type":"ContainerStarted","Data":"d85ac86ee31b16bfd2373d1e2caf130e6b8e409fe626bfe294b211f86639098c"} Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.485389 4632 patch_prober.go:28] interesting pod/controller-manager-7469657588-kpf64 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.485429 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" podUID="a8ff14f9-e25c-4839-acab-a622f6f70f88" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.485394 4632 patch_prober.go:28] interesting pod/console-operator-58897d9998-sbtn5 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Mar 13 11:06:02 crc kubenswrapper[4632]: I0313 11:06:02.485574 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" podUID="ef269b18-ea84-43c2-971c-e772149acbf6" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Mar 13 11:06:02 crc kubenswrapper[4632]: E0313 11:06:02.496922 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cdf295326a62a01129c4f9b5741f57b3d80d103e3c5a6bf64f5cc1951034264f" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 13 11:06:02 crc kubenswrapper[4632]: E0313 11:06:02.501287 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cdf295326a62a01129c4f9b5741f57b3d80d103e3c5a6bf64f5cc1951034264f" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 13 11:06:02 crc kubenswrapper[4632]: E0313 11:06:02.505190 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cdf295326a62a01129c4f9b5741f57b3d80d103e3c5a6bf64f5cc1951034264f" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 13 11:06:02 crc kubenswrapper[4632]: E0313 11:06:02.505243 4632 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" Mar 13 11:06:03 crc kubenswrapper[4632]: I0313 11:06:03.501618 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 11:06:03 crc kubenswrapper[4632]: I0313 11:06:03.502334 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" podUID="68c5eb80-4214-42c5-a08d-de6012969621" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.58:8081/readyz\": dial tcp 10.217.0.58:8081: connect: connection refused" Mar 13 11:06:03 crc kubenswrapper[4632]: I0313 11:06:03.501769 4632 patch_prober.go:28] interesting pod/controller-manager-7469657588-kpf64 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Mar 13 11:06:03 crc kubenswrapper[4632]: I0313 11:06:03.504004 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" podUID="a8ff14f9-e25c-4839-acab-a622f6f70f88" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Mar 13 11:06:03 crc kubenswrapper[4632]: I0313 11:06:03.501963 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" start-of-body= Mar 13 11:06:03 crc kubenswrapper[4632]: I0313 11:06:03.504060 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" Mar 13 11:06:03 crc kubenswrapper[4632]: I0313 11:06:03.590161 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openstack-operators/openstack-operator-index-2jqnk" podUID="7de02b7f-4e1c-4ba1-9659-c864e9080092" containerName="registry-server" probeResult="failure" output=< Mar 13 11:06:03 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:06:03 crc kubenswrapper[4632]: > Mar 13 11:06:03 crc kubenswrapper[4632]: E0313 11:06:03.737498 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8d6e3c3bf2cc94b4f346233606a1c5c55a2993e7644d8a78c77dc12972c98a9e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 13 11:06:03 crc kubenswrapper[4632]: E0313 11:06:03.739425 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8d6e3c3bf2cc94b4f346233606a1c5c55a2993e7644d8a78c77dc12972c98a9e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 13 11:06:03 crc kubenswrapper[4632]: E0313 11:06:03.740825 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="8d6e3c3bf2cc94b4f346233606a1c5c55a2993e7644d8a78c77dc12972c98a9e" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 13 11:06:03 crc kubenswrapper[4632]: E0313 11:06:03.740855 4632 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="1761ca69-46fd-4375-af60-22b3e77c19a2" containerName="galera" Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.120181 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5678554f8b-n7dcv" Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.357512 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="d2c1c19b-95a5-4db1-8e54-36fe83704b25" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.438719 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.438785 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.438840 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.439165 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.439239 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.439508 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.439534 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.441762 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="openshift-config-operator" containerStatusID={"Type":"cri-o","ID":"bdfa238d1dda3afead970f6c0c59d9c82cc9066974eef2637a5f643bcf655e99"} pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" containerMessage="Container openshift-config-operator failed liveness probe, will be restarted" Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.441896 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" containerID="cri-o://bdfa238d1dda3afead970f6c0c59d9c82cc9066974eef2637a5f643bcf655e99" gracePeriod=30 Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.536913 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" event={"ID":"f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda","Type":"ContainerStarted","Data":"94c387b11711608204f9f44972b019fe580f00c8f5f04ef461b483111e42908d"} Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.537685 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.538666 4632 patch_prober.go:28] interesting pod/controller-manager-7469657588-kpf64 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.538714 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" podUID="a8ff14f9-e25c-4839-acab-a622f6f70f88" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.538986 4632 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-zgxcd container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" start-of-body= Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.539031 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" podUID="49c520f1-fb05-48ca-8435-1985ce668451" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": dial tcp 10.217.0.27:5443: connect: connection refused" Mar 13 11:06:04 crc kubenswrapper[4632]: I0313 11:06:04.698171 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-lvlxj" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.202503 4632 patch_prober.go:28] interesting pod/controller-manager-7469657588-kpf64 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" start-of-body= Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.202975 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" podUID="a8ff14f9-e25c-4839-acab-a622f6f70f88" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.66:8443/healthz\": dial tcp 10.217.0.66:8443: connect: connection refused" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.563952 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-tztd9" event={"ID":"8f51973a-596d-40dc-9b5b-b2c95a60ea0c","Type":"ContainerStarted","Data":"e99ea45426a904ea957ea0ea2fbbe1c2c9717dbd8a660fe6124f64d95546a1bf"} Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.564014 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-tztd9" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.570741 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-config-operator_openshift-config-operator-7777fb866f-sk2l6_f660255f-8f78-4876-973d-db58f2ee7020/openshift-config-operator/1.log" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.576009 4632 generic.go:334] "Generic (PLEG): container finished" podID="f660255f-8f78-4876-973d-db58f2ee7020" containerID="bdfa238d1dda3afead970f6c0c59d9c82cc9066974eef2637a5f643bcf655e99" exitCode=2 Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.576099 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" event={"ID":"f660255f-8f78-4876-973d-db58f2ee7020","Type":"ContainerDied","Data":"bdfa238d1dda3afead970f6c0c59d9c82cc9066974eef2637a5f643bcf655e99"} Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.576166 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" event={"ID":"f660255f-8f78-4876-973d-db58f2ee7020","Type":"ContainerStarted","Data":"dd3e92e888d4cb87e3f4eb1d1058f6b2c0167d20d2613589ae205bd1b8bf5ea0"} Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.576515 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.592899 4632 scope.go:117] "RemoveContainer" containerID="e98f0e8253db82d7fc1c628a628a0d9ea91c85c3796f3abe0d968983b3e782e2" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.639955 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556666-pjsh7"] Mar 13 11:06:05 crc kubenswrapper[4632]: E0313 11:06:05.699170 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerName="extract-utilities" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.699212 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerName="extract-utilities" Mar 13 11:06:05 crc kubenswrapper[4632]: E0313 11:06:05.699243 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerName="extract-content" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.699250 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerName="extract-content" Mar 13 11:06:05 crc kubenswrapper[4632]: E0313 11:06:05.699263 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerName="registry-server" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.699270 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerName="registry-server" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.703395 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c07de1e-84e2-4dae-a3c3-ced19801c8c2" containerName="registry-server" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.717454 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556666-pjsh7" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.790549 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:06:05 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:06:05 crc kubenswrapper[4632]: > Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.841601 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.842017 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.841602 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:06:05 crc kubenswrapper[4632]: I0313 11:06:05.906096 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9fsz\" (UniqueName: \"kubernetes.io/projected/c793856b-f941-4c9e-b70e-36b4844e4eac-kube-api-access-f9fsz\") pod \"auto-csr-approver-29556666-pjsh7\" (UID: \"c793856b-f941-4c9e-b70e-36b4844e4eac\") " pod="openshift-infra/auto-csr-approver-29556666-pjsh7" Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.009361 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9fsz\" (UniqueName: \"kubernetes.io/projected/c793856b-f941-4c9e-b70e-36b4844e4eac-kube-api-access-f9fsz\") pod \"auto-csr-approver-29556666-pjsh7\" (UID: \"c793856b-f941-4c9e-b70e-36b4844e4eac\") " pod="openshift-infra/auto-csr-approver-29556666-pjsh7" Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.161582 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9fsz\" (UniqueName: \"kubernetes.io/projected/c793856b-f941-4c9e-b70e-36b4844e4eac-kube-api-access-f9fsz\") pod \"auto-csr-approver-29556666-pjsh7\" (UID: \"c793856b-f941-4c9e-b70e-36b4844e4eac\") " pod="openshift-infra/auto-csr-approver-29556666-pjsh7" Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.189347 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-r5v5p" Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.211821 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556666-pjsh7" Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.303331 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]backend-http ok Mar 13 11:06:06 crc kubenswrapper[4632]: [+]has-synced ok Mar 13 11:06:06 crc kubenswrapper[4632]: [-]process-running failed: reason withheld Mar 13 11:06:06 crc kubenswrapper[4632]: healthz check failed Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.303390 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.492712 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556666-pjsh7"] Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.555647 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-sbtn5" Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.589705 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-config-operator_openshift-config-operator-7777fb866f-sk2l6_f660255f-8f78-4876-973d-db58f2ee7020/openshift-config-operator/1.log" Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.622418 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" event={"ID":"48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd","Type":"ContainerStarted","Data":"40af26cca0da5fdf2bc9e8a5f824f238d6ed7789f9bf66f1ba1f3a52d86ef473"} Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.622825 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.623312 4632 patch_prober.go:28] interesting pod/oauth-openshift-75bb75cfd7-8sh2x container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": dial tcp 10.217.0.68:6443: connect: connection refused" start-of-body= Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.623347 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" podUID="48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": dial tcp 10.217.0.68:6443: connect: connection refused" Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.625330 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-zgxcd" Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.710974 4632 generic.go:334] "Generic (PLEG): container finished" podID="1761ca69-46fd-4375-af60-22b3e77c19a2" containerID="8d6e3c3bf2cc94b4f346233606a1c5c55a2993e7644d8a78c77dc12972c98a9e" exitCode=0 Mar 13 11:06:06 crc kubenswrapper[4632]: I0313 11:06:06.711465 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1761ca69-46fd-4375-af60-22b3e77c19a2","Type":"ContainerDied","Data":"8d6e3c3bf2cc94b4f346233606a1c5c55a2993e7644d8a78c77dc12972c98a9e"} Mar 13 11:06:07 crc kubenswrapper[4632]: I0313 11:06:07.603634 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" Mar 13 11:06:07 crc kubenswrapper[4632]: I0313 11:06:07.610133 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-wj9qs" Mar 13 11:06:07 crc kubenswrapper[4632]: I0313 11:06:07.720687 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="d2c1c19b-95a5-4db1-8e54-36fe83704b25" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 11:06:07 crc kubenswrapper[4632]: I0313 11:06:07.807403 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-5444994796-t9vht_7b959a85-56a5-4296-9cf3-87741e1f9c39/router/0.log" Mar 13 11:06:07 crc kubenswrapper[4632]: I0313 11:06:07.807484 4632 generic.go:334] "Generic (PLEG): container finished" podID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerID="c0db1ffabe3d33862c8266179a821f8fd8c1a4906081849cc73b575a98544e3b" exitCode=137 Mar 13 11:06:07 crc kubenswrapper[4632]: I0313 11:06:07.808505 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-t9vht" event={"ID":"7b959a85-56a5-4296-9cf3-87741e1f9c39","Type":"ContainerDied","Data":"c0db1ffabe3d33862c8266179a821f8fd8c1a4906081849cc73b575a98544e3b"} Mar 13 11:06:07 crc kubenswrapper[4632]: I0313 11:06:07.809345 4632 patch_prober.go:28] interesting pod/oauth-openshift-75bb75cfd7-8sh2x container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": dial tcp 10.217.0.68:6443: connect: connection refused" start-of-body= Mar 13 11:06:07 crc kubenswrapper[4632]: I0313 11:06:07.809391 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" podUID="48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": dial tcp 10.217.0.68:6443: connect: connection refused" Mar 13 11:06:08 crc kubenswrapper[4632]: I0313 11:06:08.184359 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" podUID="f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.75:8081/readyz\": dial tcp 10.217.0.75:8081: connect: connection refused" Mar 13 11:06:08 crc kubenswrapper[4632]: I0313 11:06:08.867416 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"1761ca69-46fd-4375-af60-22b3e77c19a2","Type":"ContainerStarted","Data":"ac8180dd063c8f549a7901dcd6b082d73376e180684216596e3f93393b60968e"} Mar 13 11:06:08 crc kubenswrapper[4632]: I0313 11:06:08.872530 4632 generic.go:334] "Generic (PLEG): container finished" podID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerID="3da76186915cfbbbe688750a6110b1e64143d37e61c44ef62a9740eabb32c983" exitCode=0 Mar 13 11:06:08 crc kubenswrapper[4632]: I0313 11:06:08.872569 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac97dc03-9537-4f95-bb79-5bb60a99089d","Type":"ContainerDied","Data":"3da76186915cfbbbe688750a6110b1e64143d37e61c44ef62a9740eabb32c983"} Mar 13 11:06:09 crc kubenswrapper[4632]: I0313 11:06:09.363194 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-bkmbn" podUID="c33d0da9-5a04-42d6-80d3-2f558b4a90b0" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.78:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.070530 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:06:10 crc kubenswrapper[4632]: E0313 11:06:10.072402 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.103145 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" podUID="7bab78c8-7dac-48dc-a426-ccd4ae00a428" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.89:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.103316 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.144360 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-jwrgq" Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.364649 4632 patch_prober.go:28] interesting pod/oauth-openshift-75bb75cfd7-8sh2x container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.68:6443/healthz\": dial tcp 10.217.0.68:6443: connect: connection refused" start-of-body= Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.364695 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" podUID="48d2bc7e-c929-42c9-b3f2-9e78c7eac8cd" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.68:6443/healthz\": dial tcp 10.217.0.68:6443: connect: connection refused" Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.439363 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.439411 4632 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-sk2l6 container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.439424 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.439449 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" podUID="f660255f-8f78-4876-973d-db58f2ee7020" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.16:8443/healthz\": dial tcp 10.217.0.16:8443: connect: connection refused" Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.714398 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="d2c1c19b-95a5-4db1-8e54-36fe83704b25" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.714675 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.725026 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cinder-scheduler" containerStatusID={"Type":"cri-o","ID":"67516bb124d863acdb93cbafff12001c1c53c2a821587b0e3e99f6135ee28e92"} pod="openstack/cinder-scheduler-0" containerMessage="Container cinder-scheduler failed liveness probe, will be restarted" Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.728663 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="d2c1c19b-95a5-4db1-8e54-36fe83704b25" containerName="cinder-scheduler" containerID="cri-o://67516bb124d863acdb93cbafff12001c1c53c2a821587b0e3e99f6135ee28e92" gracePeriod=30 Mar 13 11:06:10 crc kubenswrapper[4632]: I0313 11:06:10.996017 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-5444994796-t9vht_7b959a85-56a5-4296-9cf3-87741e1f9c39/router/0.log" Mar 13 11:06:11 crc kubenswrapper[4632]: I0313 11:06:11.000096 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-t9vht" event={"ID":"7b959a85-56a5-4296-9cf3-87741e1f9c39","Type":"ContainerStarted","Data":"e68290d07d3694966f4a24ae2bc7a4e991ff7c78cb35b4090c838101dbb38ee0"} Mar 13 11:06:11 crc kubenswrapper[4632]: I0313 11:06:11.190464 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 11:06:11 crc kubenswrapper[4632]: I0313 11:06:11.196705 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 13 11:06:11 crc kubenswrapper[4632]: I0313 11:06:11.196777 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 13 11:06:11 crc kubenswrapper[4632]: I0313 11:06:11.210601 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 11:06:11 crc kubenswrapper[4632]: I0313 11:06:11.247275 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-2jqnk" Mar 13 11:06:12 crc kubenswrapper[4632]: I0313 11:06:12.136915 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-db6b8fbf8-pllt2" Mar 13 11:06:12 crc kubenswrapper[4632]: I0313 11:06:12.219222 4632 patch_prober.go:28] interesting pod/router-default-5444994796-t9vht container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 11:06:12 crc kubenswrapper[4632]: [-]has-synced failed: reason withheld Mar 13 11:06:12 crc kubenswrapper[4632]: [+]process-running ok Mar 13 11:06:12 crc kubenswrapper[4632]: healthz check failed Mar 13 11:06:12 crc kubenswrapper[4632]: I0313 11:06:12.219580 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-t9vht" podUID="7b959a85-56a5-4296-9cf3-87741e1f9c39" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 11:06:12 crc kubenswrapper[4632]: E0313 11:06:12.503794 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cdf295326a62a01129c4f9b5741f57b3d80d103e3c5a6bf64f5cc1951034264f" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 13 11:06:12 crc kubenswrapper[4632]: E0313 11:06:12.508245 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cdf295326a62a01129c4f9b5741f57b3d80d103e3c5a6bf64f5cc1951034264f" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 13 11:06:12 crc kubenswrapper[4632]: E0313 11:06:12.513473 4632 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="cdf295326a62a01129c4f9b5741f57b3d80d103e3c5a6bf64f5cc1951034264f" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 13 11:06:12 crc kubenswrapper[4632]: E0313 11:06:12.513557 4632 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" Mar 13 11:06:13 crc kubenswrapper[4632]: I0313 11:06:13.042997 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac97dc03-9537-4f95-bb79-5bb60a99089d","Type":"ContainerStarted","Data":"b57f85647a9f24895343d16ffa69abe242612a8de43972c5cb53a02f9838e13b"} Mar 13 11:06:13 crc kubenswrapper[4632]: I0313 11:06:13.198258 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 11:06:13 crc kubenswrapper[4632]: I0313 11:06:13.258568 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556666-pjsh7"] Mar 13 11:06:13 crc kubenswrapper[4632]: W0313 11:06:13.302807 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc793856b_f941_4c9e_b70e_36b4844e4eac.slice/crio-c4d5e3a91207e5a09f63b2550c521edf15d369030662552b3e6ac28d2f654bea WatchSource:0}: Error finding container c4d5e3a91207e5a09f63b2550c521edf15d369030662552b3e6ac28d2f654bea: Status 404 returned error can't find the container with id c4d5e3a91207e5a09f63b2550c521edf15d369030662552b3e6ac28d2f654bea Mar 13 11:06:13 crc kubenswrapper[4632]: I0313 11:06:13.507669 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sk2l6" Mar 13 11:06:13 crc kubenswrapper[4632]: I0313 11:06:13.734391 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 13 11:06:13 crc kubenswrapper[4632]: I0313 11:06:13.734433 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 13 11:06:14 crc kubenswrapper[4632]: I0313 11:06:14.079872 4632 generic.go:334] "Generic (PLEG): container finished" podID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerID="cdf295326a62a01129c4f9b5741f57b3d80d103e3c5a6bf64f5cc1951034264f" exitCode=0 Mar 13 11:06:14 crc kubenswrapper[4632]: I0313 11:06:14.079978 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2cb2f546-c8c5-4ec9-aba8-d3782431de10","Type":"ContainerDied","Data":"cdf295326a62a01129c4f9b5741f57b3d80d103e3c5a6bf64f5cc1951034264f"} Mar 13 11:06:14 crc kubenswrapper[4632]: I0313 11:06:14.081652 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556666-pjsh7" event={"ID":"c793856b-f941-4c9e-b70e-36b4844e4eac","Type":"ContainerStarted","Data":"c4d5e3a91207e5a09f63b2550c521edf15d369030662552b3e6ac28d2f654bea"} Mar 13 11:06:14 crc kubenswrapper[4632]: I0313 11:06:14.083096 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 11:06:14 crc kubenswrapper[4632]: I0313 11:06:14.086235 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-t9vht" Mar 13 11:06:14 crc kubenswrapper[4632]: I0313 11:06:14.389483 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-lvlxj" Mar 13 11:06:14 crc kubenswrapper[4632]: I0313 11:06:14.810256 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-9zbh8" Mar 13 11:06:15 crc kubenswrapper[4632]: I0313 11:06:15.106576 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"2cb2f546-c8c5-4ec9-aba8-d3782431de10","Type":"ContainerStarted","Data":"75b6983c07eb146a71cf122cfc436a920ef6367277c67e8ac65ff3bc2c650e0f"} Mar 13 11:06:15 crc kubenswrapper[4632]: I0313 11:06:15.487274 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7469657588-kpf64" Mar 13 11:06:15 crc kubenswrapper[4632]: I0313 11:06:15.510667 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 11:06:15 crc kubenswrapper[4632]: I0313 11:06:15.511282 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-notification-agent" containerID="cri-o://5d34565e3f3d53e4eb4eec7fc127b7d0ef95db5c894a8b9fbc65ec70d12e4d20" gracePeriod=30 Mar 13 11:06:15 crc kubenswrapper[4632]: I0313 11:06:15.511412 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-central-agent" containerID="cri-o://b57f85647a9f24895343d16ffa69abe242612a8de43972c5cb53a02f9838e13b" gracePeriod=30 Mar 13 11:06:15 crc kubenswrapper[4632]: I0313 11:06:15.511443 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="sg-core" containerID="cri-o://7784ac325dc1b12d740a758a07a8e9e03da012db50eef1bc62b207161880f530" gracePeriod=30 Mar 13 11:06:15 crc kubenswrapper[4632]: I0313 11:06:15.511412 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="proxy-httpd" containerID="cri-o://9a2b2ece3b9e850a4d4ebe5776040511a71de7bd0fba43340538aa166e80ade2" gracePeriod=30 Mar 13 11:06:15 crc kubenswrapper[4632]: I0313 11:06:15.789830 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:06:15 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:06:15 crc kubenswrapper[4632]: > Mar 13 11:06:15 crc kubenswrapper[4632]: I0313 11:06:15.794637 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-tztd9" Mar 13 11:06:16 crc kubenswrapper[4632]: I0313 11:06:16.152847 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556666-pjsh7" event={"ID":"c793856b-f941-4c9e-b70e-36b4844e4eac","Type":"ContainerStarted","Data":"ca3b5ee01f58147d02e068e963d2c27601cb82a563a21e34b84cdc61a03e2f80"} Mar 13 11:06:16 crc kubenswrapper[4632]: I0313 11:06:16.160051 4632 generic.go:334] "Generic (PLEG): container finished" podID="d2c1c19b-95a5-4db1-8e54-36fe83704b25" containerID="67516bb124d863acdb93cbafff12001c1c53c2a821587b0e3e99f6135ee28e92" exitCode=0 Mar 13 11:06:16 crc kubenswrapper[4632]: I0313 11:06:16.160107 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d2c1c19b-95a5-4db1-8e54-36fe83704b25","Type":"ContainerDied","Data":"67516bb124d863acdb93cbafff12001c1c53c2a821587b0e3e99f6135ee28e92"} Mar 13 11:06:16 crc kubenswrapper[4632]: I0313 11:06:16.174133 4632 generic.go:334] "Generic (PLEG): container finished" podID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerID="9a2b2ece3b9e850a4d4ebe5776040511a71de7bd0fba43340538aa166e80ade2" exitCode=0 Mar 13 11:06:16 crc kubenswrapper[4632]: I0313 11:06:16.174506 4632 generic.go:334] "Generic (PLEG): container finished" podID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerID="7784ac325dc1b12d740a758a07a8e9e03da012db50eef1bc62b207161880f530" exitCode=2 Mar 13 11:06:16 crc kubenswrapper[4632]: I0313 11:06:16.174405 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac97dc03-9537-4f95-bb79-5bb60a99089d","Type":"ContainerDied","Data":"9a2b2ece3b9e850a4d4ebe5776040511a71de7bd0fba43340538aa166e80ade2"} Mar 13 11:06:16 crc kubenswrapper[4632]: I0313 11:06:16.174653 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac97dc03-9537-4f95-bb79-5bb60a99089d","Type":"ContainerDied","Data":"7784ac325dc1b12d740a758a07a8e9e03da012db50eef1bc62b207161880f530"} Mar 13 11:06:16 crc kubenswrapper[4632]: I0313 11:06:16.174884 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556666-pjsh7" podStartSLOduration=14.206123141 podStartE2EDuration="15.17249465s" podCreationTimestamp="2026-03-13 11:06:01 +0000 UTC" firstStartedPulling="2026-03-13 11:06:13.314219236 +0000 UTC m=+3747.336749369" lastFinishedPulling="2026-03-13 11:06:14.280590745 +0000 UTC m=+3748.303120878" observedRunningTime="2026-03-13 11:06:16.170610543 +0000 UTC m=+3750.193140686" watchObservedRunningTime="2026-03-13 11:06:16.17249465 +0000 UTC m=+3750.195024793" Mar 13 11:06:17 crc kubenswrapper[4632]: I0313 11:06:17.189388 4632 generic.go:334] "Generic (PLEG): container finished" podID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerID="5d34565e3f3d53e4eb4eec7fc127b7d0ef95db5c894a8b9fbc65ec70d12e4d20" exitCode=0 Mar 13 11:06:17 crc kubenswrapper[4632]: I0313 11:06:17.189478 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac97dc03-9537-4f95-bb79-5bb60a99089d","Type":"ContainerDied","Data":"5d34565e3f3d53e4eb4eec7fc127b7d0ef95db5c894a8b9fbc65ec70d12e4d20"} Mar 13 11:06:18 crc kubenswrapper[4632]: I0313 11:06:18.242427 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-6nb82" Mar 13 11:06:18 crc kubenswrapper[4632]: I0313 11:06:18.501620 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 13 11:06:18 crc kubenswrapper[4632]: I0313 11:06:18.759774 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 13 11:06:19 crc kubenswrapper[4632]: I0313 11:06:19.214637 4632 generic.go:334] "Generic (PLEG): container finished" podID="c793856b-f941-4c9e-b70e-36b4844e4eac" containerID="ca3b5ee01f58147d02e068e963d2c27601cb82a563a21e34b84cdc61a03e2f80" exitCode=0 Mar 13 11:06:19 crc kubenswrapper[4632]: I0313 11:06:19.214704 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556666-pjsh7" event={"ID":"c793856b-f941-4c9e-b70e-36b4844e4eac","Type":"ContainerDied","Data":"ca3b5ee01f58147d02e068e963d2c27601cb82a563a21e34b84cdc61a03e2f80"} Mar 13 11:06:19 crc kubenswrapper[4632]: I0313 11:06:19.219288 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"d2c1c19b-95a5-4db1-8e54-36fe83704b25","Type":"ContainerStarted","Data":"eebcdf889d3b8172d873374b42b5baa02f10654715a042e6a0eb5b0ebb82e252"} Mar 13 11:06:20 crc kubenswrapper[4632]: I0313 11:06:20.326049 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 11:06:20 crc kubenswrapper[4632]: I0313 11:06:20.376277 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-75bb75cfd7-8sh2x" Mar 13 11:06:21 crc kubenswrapper[4632]: I0313 11:06:21.042783 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556666-pjsh7" Mar 13 11:06:21 crc kubenswrapper[4632]: I0313 11:06:21.214096 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9fsz\" (UniqueName: \"kubernetes.io/projected/c793856b-f941-4c9e-b70e-36b4844e4eac-kube-api-access-f9fsz\") pod \"c793856b-f941-4c9e-b70e-36b4844e4eac\" (UID: \"c793856b-f941-4c9e-b70e-36b4844e4eac\") " Mar 13 11:06:21 crc kubenswrapper[4632]: I0313 11:06:21.246319 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c793856b-f941-4c9e-b70e-36b4844e4eac-kube-api-access-f9fsz" (OuterVolumeSpecName: "kube-api-access-f9fsz") pod "c793856b-f941-4c9e-b70e-36b4844e4eac" (UID: "c793856b-f941-4c9e-b70e-36b4844e4eac"). InnerVolumeSpecName "kube-api-access-f9fsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:06:21 crc kubenswrapper[4632]: I0313 11:06:21.248326 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556666-pjsh7" Mar 13 11:06:21 crc kubenswrapper[4632]: I0313 11:06:21.249324 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556666-pjsh7" event={"ID":"c793856b-f941-4c9e-b70e-36b4844e4eac","Type":"ContainerDied","Data":"c4d5e3a91207e5a09f63b2550c521edf15d369030662552b3e6ac28d2f654bea"} Mar 13 11:06:21 crc kubenswrapper[4632]: I0313 11:06:21.261308 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4d5e3a91207e5a09f63b2550c521edf15d369030662552b3e6ac28d2f654bea" Mar 13 11:06:21 crc kubenswrapper[4632]: I0313 11:06:21.319138 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9fsz\" (UniqueName: \"kubernetes.io/projected/c793856b-f941-4c9e-b70e-36b4844e4eac-kube-api-access-f9fsz\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:21 crc kubenswrapper[4632]: I0313 11:06:21.349629 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556660-dxrjl"] Mar 13 11:06:21 crc kubenswrapper[4632]: I0313 11:06:21.363247 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 11:06:21 crc kubenswrapper[4632]: I0313 11:06:21.364159 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556660-dxrjl"] Mar 13 11:06:21 crc kubenswrapper[4632]: E0313 11:06:21.500708 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc793856b_f941_4c9e_b70e_36b4844e4eac.slice\": RecentStats: unable to find data in memory cache]" Mar 13 11:06:22 crc kubenswrapper[4632]: I0313 11:06:22.057010 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae414ebe-e9fa-4c30-965a-e368234bbb18" path="/var/lib/kubelet/pods/ae414ebe-e9fa-4c30-965a-e368234bbb18/volumes" Mar 13 11:06:22 crc kubenswrapper[4632]: I0313 11:06:22.476861 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 13 11:06:22 crc kubenswrapper[4632]: I0313 11:06:22.476977 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 13 11:06:22 crc kubenswrapper[4632]: I0313 11:06:22.612920 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 13 11:06:23 crc kubenswrapper[4632]: I0313 11:06:23.044091 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:06:23 crc kubenswrapper[4632]: E0313 11:06:23.044674 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:06:23 crc kubenswrapper[4632]: I0313 11:06:23.285595 4632 generic.go:334] "Generic (PLEG): container finished" podID="a62e0eae-95dd-40a3-a489-80646fde4301" containerID="ec15c016ac8280363b8fb347025993466f5b7492f2d0ac470ef8fc423974c0e2" exitCode=1 Mar 13 11:06:23 crc kubenswrapper[4632]: I0313 11:06:23.285680 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" event={"ID":"a62e0eae-95dd-40a3-a489-80646fde4301","Type":"ContainerDied","Data":"ec15c016ac8280363b8fb347025993466f5b7492f2d0ac470ef8fc423974c0e2"} Mar 13 11:06:23 crc kubenswrapper[4632]: I0313 11:06:23.408370 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.396656 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.430119 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.518666 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"a62e0eae-95dd-40a3-a489-80646fde4301\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.518769 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a62e0eae-95dd-40a3-a489-80646fde4301-openstack-config\") pod \"a62e0eae-95dd-40a3-a489-80646fde4301\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.518794 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8w5vp\" (UniqueName: \"kubernetes.io/projected/a62e0eae-95dd-40a3-a489-80646fde4301-kube-api-access-8w5vp\") pod \"a62e0eae-95dd-40a3-a489-80646fde4301\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.518819 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a62e0eae-95dd-40a3-a489-80646fde4301-test-operator-ephemeral-workdir\") pod \"a62e0eae-95dd-40a3-a489-80646fde4301\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.518915 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-ssh-key\") pod \"a62e0eae-95dd-40a3-a489-80646fde4301\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.518987 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a62e0eae-95dd-40a3-a489-80646fde4301-test-operator-ephemeral-temporary\") pod \"a62e0eae-95dd-40a3-a489-80646fde4301\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.519008 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a62e0eae-95dd-40a3-a489-80646fde4301-config-data\") pod \"a62e0eae-95dd-40a3-a489-80646fde4301\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.519051 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-ca-certs\") pod \"a62e0eae-95dd-40a3-a489-80646fde4301\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.519080 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-openstack-config-secret\") pod \"a62e0eae-95dd-40a3-a489-80646fde4301\" (UID: \"a62e0eae-95dd-40a3-a489-80646fde4301\") " Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.522070 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a62e0eae-95dd-40a3-a489-80646fde4301-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "a62e0eae-95dd-40a3-a489-80646fde4301" (UID: "a62e0eae-95dd-40a3-a489-80646fde4301"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.523898 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a62e0eae-95dd-40a3-a489-80646fde4301-config-data" (OuterVolumeSpecName: "config-data") pod "a62e0eae-95dd-40a3-a489-80646fde4301" (UID: "a62e0eae-95dd-40a3-a489-80646fde4301"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.530445 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "a62e0eae-95dd-40a3-a489-80646fde4301" (UID: "a62e0eae-95dd-40a3-a489-80646fde4301"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.536446 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a62e0eae-95dd-40a3-a489-80646fde4301-kube-api-access-8w5vp" (OuterVolumeSpecName: "kube-api-access-8w5vp") pod "a62e0eae-95dd-40a3-a489-80646fde4301" (UID: "a62e0eae-95dd-40a3-a489-80646fde4301"). InnerVolumeSpecName "kube-api-access-8w5vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.576466 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "a62e0eae-95dd-40a3-a489-80646fde4301" (UID: "a62e0eae-95dd-40a3-a489-80646fde4301"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.614371 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "a62e0eae-95dd-40a3-a489-80646fde4301" (UID: "a62e0eae-95dd-40a3-a489-80646fde4301"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.621912 4632 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-ca-certs\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.625007 4632 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.625036 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8w5vp\" (UniqueName: \"kubernetes.io/projected/a62e0eae-95dd-40a3-a489-80646fde4301-kube-api-access-8w5vp\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.625051 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-ssh-key\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.625061 4632 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/a62e0eae-95dd-40a3-a489-80646fde4301-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.625071 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a62e0eae-95dd-40a3-a489-80646fde4301-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.647400 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a62e0eae-95dd-40a3-a489-80646fde4301-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "a62e0eae-95dd-40a3-a489-80646fde4301" (UID: "a62e0eae-95dd-40a3-a489-80646fde4301"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.652842 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s01-single-thread-testing"] Mar 13 11:06:25 crc kubenswrapper[4632]: E0313 11:06:25.656989 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a62e0eae-95dd-40a3-a489-80646fde4301" containerName="tempest-tests-tempest-tests-runner" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.657230 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a62e0eae-95dd-40a3-a489-80646fde4301" containerName="tempest-tests-tempest-tests-runner" Mar 13 11:06:25 crc kubenswrapper[4632]: E0313 11:06:25.657877 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c793856b-f941-4c9e-b70e-36b4844e4eac" containerName="oc" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.658021 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c793856b-f941-4c9e-b70e-36b4844e4eac" containerName="oc" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.658463 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="a62e0eae-95dd-40a3-a489-80646fde4301" containerName="tempest-tests-tempest-tests-runner" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.658570 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c793856b-f941-4c9e-b70e-36b4844e4eac" containerName="oc" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.666139 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a62e0eae-95dd-40a3-a489-80646fde4301-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "a62e0eae-95dd-40a3-a489-80646fde4301" (UID: "a62e0eae-95dd-40a3-a489-80646fde4301"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.668161 4632 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.670397 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.679529 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "a62e0eae-95dd-40a3-a489-80646fde4301" (UID: "a62e0eae-95dd-40a3-a489-80646fde4301"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.682030 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s1" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.682032 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s1" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.693083 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-thread-testing"] Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.726945 4632 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/a62e0eae-95dd-40a3-a489-80646fde4301-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.727198 4632 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.727210 4632 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/a62e0eae-95dd-40a3-a489-80646fde4301-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.727219 4632 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/a62e0eae-95dd-40a3-a489-80646fde4301-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.730868 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:06:25 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:06:25 crc kubenswrapper[4632]: > Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.730961 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.731847 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"0bbbe65ea71f36a37f33d902708fe700b15c322c14c94c121a3ca523a54d026b"} pod="openshift-marketplace/redhat-operators-mm6fq" containerMessage="Container registry-server failed startup probe, will be restarted" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.731880 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" containerID="cri-o://0bbbe65ea71f36a37f33d902708fe700b15c322c14c94c121a3ca523a54d026b" gracePeriod=30 Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.829463 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.829792 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/611401cc-04fe-4276-82fa-a896182802d4-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.829997 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/611401cc-04fe-4276-82fa-a896182802d4-config-data\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.830225 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-ca-certs\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.830415 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/611401cc-04fe-4276-82fa-a896182802d4-openstack-config\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.830499 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.830581 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-ssh-key\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.830605 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrwf2\" (UniqueName: \"kubernetes.io/projected/611401cc-04fe-4276-82fa-a896182802d4-kube-api-access-hrwf2\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.830662 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/611401cc-04fe-4276-82fa-a896182802d4-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.932647 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-ca-certs\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.932883 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/611401cc-04fe-4276-82fa-a896182802d4-openstack-config\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.932992 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.933118 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-ssh-key\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.933187 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrwf2\" (UniqueName: \"kubernetes.io/projected/611401cc-04fe-4276-82fa-a896182802d4-kube-api-access-hrwf2\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.933297 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/611401cc-04fe-4276-82fa-a896182802d4-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.933779 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.934030 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/611401cc-04fe-4276-82fa-a896182802d4-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.934169 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/611401cc-04fe-4276-82fa-a896182802d4-config-data\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.934486 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/611401cc-04fe-4276-82fa-a896182802d4-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.934525 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/611401cc-04fe-4276-82fa-a896182802d4-openstack-config\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.933722 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/611401cc-04fe-4276-82fa-a896182802d4-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.934743 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.937734 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.940602 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-ca-certs\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.942534 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-ssh-key\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.951811 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/611401cc-04fe-4276-82fa-a896182802d4-config-data\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.958671 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrwf2\" (UniqueName: \"kubernetes.io/projected/611401cc-04fe-4276-82fa-a896182802d4-kube-api-access-hrwf2\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.977217 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s01-single-thread-testing\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:25 crc kubenswrapper[4632]: I0313 11:06:25.992064 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 11:06:26 crc kubenswrapper[4632]: I0313 11:06:26.313049 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" event={"ID":"a62e0eae-95dd-40a3-a489-80646fde4301","Type":"ContainerDied","Data":"959458908fe1f2c8aa4edafce9f9395e573f668491b9554e12daf71db7b5cc6a"} Mar 13 11:06:26 crc kubenswrapper[4632]: I0313 11:06:26.313414 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="959458908fe1f2c8aa4edafce9f9395e573f668491b9554e12daf71db7b5cc6a" Mar 13 11:06:26 crc kubenswrapper[4632]: I0313 11:06:26.313106 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-multi-thread-testing" Mar 13 11:06:26 crc kubenswrapper[4632]: I0313 11:06:26.639834 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-thread-testing"] Mar 13 11:06:27 crc kubenswrapper[4632]: I0313 11:06:27.324850 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" event={"ID":"611401cc-04fe-4276-82fa-a896182802d4","Type":"ContainerStarted","Data":"95b1b1d6a519cb7b9bfef154cebb6e4b73104a8706f52af49a8997ffa20ebd91"} Mar 13 11:06:30 crc kubenswrapper[4632]: I0313 11:06:30.354242 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" event={"ID":"611401cc-04fe-4276-82fa-a896182802d4","Type":"ContainerStarted","Data":"e5092a16adcd02c327c069b34afdd26aca8018f63ed747e3778a6c696a0e6a3c"} Mar 13 11:06:30 crc kubenswrapper[4632]: I0313 11:06:30.374643 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" podStartSLOduration=5.374631369 podStartE2EDuration="5.374631369s" podCreationTimestamp="2026-03-13 11:06:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 11:06:30.373131872 +0000 UTC m=+3764.395662015" watchObservedRunningTime="2026-03-13 11:06:30.374631369 +0000 UTC m=+3764.397161502" Mar 13 11:06:35 crc kubenswrapper[4632]: I0313 11:06:35.046301 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:06:35 crc kubenswrapper[4632]: E0313 11:06:35.047058 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:06:42 crc kubenswrapper[4632]: I0313 11:06:42.907764 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.232:3000/\": dial tcp 10.217.0.232:3000: connect: connection refused" Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.539568 4632 generic.go:334] "Generic (PLEG): container finished" podID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerID="b57f85647a9f24895343d16ffa69abe242612a8de43972c5cb53a02f9838e13b" exitCode=137 Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.540169 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac97dc03-9537-4f95-bb79-5bb60a99089d","Type":"ContainerDied","Data":"b57f85647a9f24895343d16ffa69abe242612a8de43972c5cb53a02f9838e13b"} Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.540217 4632 scope.go:117] "RemoveContainer" containerID="3da76186915cfbbbe688750a6110b1e64143d37e61c44ef62a9740eabb32c983" Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.755371 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.893150 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-ceilometer-tls-certs\") pod \"ac97dc03-9537-4f95-bb79-5bb60a99089d\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.893666 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-22tlw\" (UniqueName: \"kubernetes.io/projected/ac97dc03-9537-4f95-bb79-5bb60a99089d-kube-api-access-22tlw\") pod \"ac97dc03-9537-4f95-bb79-5bb60a99089d\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.893761 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-scripts\") pod \"ac97dc03-9537-4f95-bb79-5bb60a99089d\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.893983 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac97dc03-9537-4f95-bb79-5bb60a99089d-run-httpd\") pod \"ac97dc03-9537-4f95-bb79-5bb60a99089d\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.894036 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-combined-ca-bundle\") pod \"ac97dc03-9537-4f95-bb79-5bb60a99089d\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.894238 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac97dc03-9537-4f95-bb79-5bb60a99089d-log-httpd\") pod \"ac97dc03-9537-4f95-bb79-5bb60a99089d\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.894300 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-sg-core-conf-yaml\") pod \"ac97dc03-9537-4f95-bb79-5bb60a99089d\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.894375 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-config-data\") pod \"ac97dc03-9537-4f95-bb79-5bb60a99089d\" (UID: \"ac97dc03-9537-4f95-bb79-5bb60a99089d\") " Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.896834 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac97dc03-9537-4f95-bb79-5bb60a99089d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ac97dc03-9537-4f95-bb79-5bb60a99089d" (UID: "ac97dc03-9537-4f95-bb79-5bb60a99089d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.897114 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac97dc03-9537-4f95-bb79-5bb60a99089d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ac97dc03-9537-4f95-bb79-5bb60a99089d" (UID: "ac97dc03-9537-4f95-bb79-5bb60a99089d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.910201 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac97dc03-9537-4f95-bb79-5bb60a99089d-kube-api-access-22tlw" (OuterVolumeSpecName: "kube-api-access-22tlw") pod "ac97dc03-9537-4f95-bb79-5bb60a99089d" (UID: "ac97dc03-9537-4f95-bb79-5bb60a99089d"). InnerVolumeSpecName "kube-api-access-22tlw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.936274 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-scripts" (OuterVolumeSpecName: "scripts") pod "ac97dc03-9537-4f95-bb79-5bb60a99089d" (UID: "ac97dc03-9537-4f95-bb79-5bb60a99089d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.948311 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ac97dc03-9537-4f95-bb79-5bb60a99089d" (UID: "ac97dc03-9537-4f95-bb79-5bb60a99089d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.979113 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "ac97dc03-9537-4f95-bb79-5bb60a99089d" (UID: "ac97dc03-9537-4f95-bb79-5bb60a99089d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.997211 4632 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac97dc03-9537-4f95-bb79-5bb60a99089d-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.997316 4632 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.997328 4632 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.997339 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-22tlw\" (UniqueName: \"kubernetes.io/projected/ac97dc03-9537-4f95-bb79-5bb60a99089d-kube-api-access-22tlw\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.997346 4632 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:46 crc kubenswrapper[4632]: I0313 11:06:46.997354 4632 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ac97dc03-9537-4f95-bb79-5bb60a99089d-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.033515 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac97dc03-9537-4f95-bb79-5bb60a99089d" (UID: "ac97dc03-9537-4f95-bb79-5bb60a99089d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.044891 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:06:47 crc kubenswrapper[4632]: E0313 11:06:47.045374 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.099476 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.115140 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-config-data" (OuterVolumeSpecName: "config-data") pod "ac97dc03-9537-4f95-bb79-5bb60a99089d" (UID: "ac97dc03-9537-4f95-bb79-5bb60a99089d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.201123 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac97dc03-9537-4f95-bb79-5bb60a99089d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.553555 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ac97dc03-9537-4f95-bb79-5bb60a99089d","Type":"ContainerDied","Data":"6c35a65f59ec813bb19b2b3e4862d24780f1cf59570c0c358308767506eead20"} Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.553651 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.553668 4632 scope.go:117] "RemoveContainer" containerID="b57f85647a9f24895343d16ffa69abe242612a8de43972c5cb53a02f9838e13b" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.605723 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.617870 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.621967 4632 scope.go:117] "RemoveContainer" containerID="9a2b2ece3b9e850a4d4ebe5776040511a71de7bd0fba43340538aa166e80ade2" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.653690 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 11:06:47 crc kubenswrapper[4632]: E0313 11:06:47.654167 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-notification-agent" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.654187 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-notification-agent" Mar 13 11:06:47 crc kubenswrapper[4632]: E0313 11:06:47.654217 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="proxy-httpd" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.654226 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="proxy-httpd" Mar 13 11:06:47 crc kubenswrapper[4632]: E0313 11:06:47.654255 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="sg-core" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.654264 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="sg-core" Mar 13 11:06:47 crc kubenswrapper[4632]: E0313 11:06:47.654278 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-central-agent" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.654288 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-central-agent" Mar 13 11:06:47 crc kubenswrapper[4632]: E0313 11:06:47.654302 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-central-agent" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.654310 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-central-agent" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.654525 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-central-agent" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.654543 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="proxy-httpd" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.654558 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-notification-agent" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.654586 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="sg-core" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.655021 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" containerName="ceilometer-central-agent" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.661402 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.671541 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.671998 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.672151 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.678082 4632 scope.go:117] "RemoveContainer" containerID="7784ac325dc1b12d740a758a07a8e9e03da012db50eef1bc62b207161880f530" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.682140 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.778836 4632 scope.go:117] "RemoveContainer" containerID="5d34565e3f3d53e4eb4eec7fc127b7d0ef95db5c894a8b9fbc65ec70d12e4d20" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.811501 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.811615 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btc6h\" (UniqueName: \"kubernetes.io/projected/046f071d-f091-4681-8a9b-06c7e7dc2192-kube-api-access-btc6h\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.811691 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/046f071d-f091-4681-8a9b-06c7e7dc2192-run-httpd\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.811850 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.811960 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.812072 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/046f071d-f091-4681-8a9b-06c7e7dc2192-log-httpd\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.812119 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-config-data\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.812189 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-scripts\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.914667 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.915795 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.915892 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/046f071d-f091-4681-8a9b-06c7e7dc2192-log-httpd\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.915971 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-config-data\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.916667 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-scripts\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.916871 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/046f071d-f091-4681-8a9b-06c7e7dc2192-log-httpd\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.917354 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.917438 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btc6h\" (UniqueName: \"kubernetes.io/projected/046f071d-f091-4681-8a9b-06c7e7dc2192-kube-api-access-btc6h\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.917491 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/046f071d-f091-4681-8a9b-06c7e7dc2192-run-httpd\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.918082 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/046f071d-f091-4681-8a9b-06c7e7dc2192-run-httpd\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.922489 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.923322 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-scripts\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.924059 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-config-data\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.924658 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.934162 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/046f071d-f091-4681-8a9b-06c7e7dc2192-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:47 crc kubenswrapper[4632]: I0313 11:06:47.942345 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btc6h\" (UniqueName: \"kubernetes.io/projected/046f071d-f091-4681-8a9b-06c7e7dc2192-kube-api-access-btc6h\") pod \"ceilometer-0\" (UID: \"046f071d-f091-4681-8a9b-06c7e7dc2192\") " pod="openstack/ceilometer-0" Mar 13 11:06:48 crc kubenswrapper[4632]: I0313 11:06:48.059100 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 11:06:48 crc kubenswrapper[4632]: I0313 11:06:48.062422 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac97dc03-9537-4f95-bb79-5bb60a99089d" path="/var/lib/kubelet/pods/ac97dc03-9537-4f95-bb79-5bb60a99089d/volumes" Mar 13 11:06:48 crc kubenswrapper[4632]: I0313 11:06:48.633850 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 11:06:48 crc kubenswrapper[4632]: W0313 11:06:48.643014 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod046f071d_f091_4681_8a9b_06c7e7dc2192.slice/crio-27050561388e6ddf9f76aa250d32d0fc503c15ed538344df77e2efb4d8e9e619 WatchSource:0}: Error finding container 27050561388e6ddf9f76aa250d32d0fc503c15ed538344df77e2efb4d8e9e619: Status 404 returned error can't find the container with id 27050561388e6ddf9f76aa250d32d0fc503c15ed538344df77e2efb4d8e9e619 Mar 13 11:06:49 crc kubenswrapper[4632]: I0313 11:06:49.574113 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"046f071d-f091-4681-8a9b-06c7e7dc2192","Type":"ContainerStarted","Data":"576756d63180fa10edf4bdedeab7fa16aa80d2e15a64d7a8dc5e32f747e2cd3a"} Mar 13 11:06:49 crc kubenswrapper[4632]: I0313 11:06:49.574713 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"046f071d-f091-4681-8a9b-06c7e7dc2192","Type":"ContainerStarted","Data":"02ad1c47d8e6b66a14651c47c09b8b2a59f4584678d570942aab1e3282149017"} Mar 13 11:06:49 crc kubenswrapper[4632]: I0313 11:06:49.574727 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"046f071d-f091-4681-8a9b-06c7e7dc2192","Type":"ContainerStarted","Data":"27050561388e6ddf9f76aa250d32d0fc503c15ed538344df77e2efb4d8e9e619"} Mar 13 11:06:50 crc kubenswrapper[4632]: I0313 11:06:50.584931 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"046f071d-f091-4681-8a9b-06c7e7dc2192","Type":"ContainerStarted","Data":"caef16b68b74ddf8de5706c4854903b7a02a9f270caee37fd1eb97735d2e12cf"} Mar 13 11:06:52 crc kubenswrapper[4632]: I0313 11:06:52.603707 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"046f071d-f091-4681-8a9b-06c7e7dc2192","Type":"ContainerStarted","Data":"dfdadecf41bcb95c18a4fe4bdcab7d2331d8cdd63461ec09255dc41b0025b9ee"} Mar 13 11:06:52 crc kubenswrapper[4632]: I0313 11:06:52.604221 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 11:06:52 crc kubenswrapper[4632]: I0313 11:06:52.630521 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.937397272 podStartE2EDuration="5.630503906s" podCreationTimestamp="2026-03-13 11:06:47 +0000 UTC" firstStartedPulling="2026-03-13 11:06:48.64578449 +0000 UTC m=+3782.668314623" lastFinishedPulling="2026-03-13 11:06:51.338891104 +0000 UTC m=+3785.361421257" observedRunningTime="2026-03-13 11:06:52.626818854 +0000 UTC m=+3786.649348987" watchObservedRunningTime="2026-03-13 11:06:52.630503906 +0000 UTC m=+3786.653034039" Mar 13 11:06:56 crc kubenswrapper[4632]: I0313 11:06:56.668873 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mm6fq_269ac923-f4f9-43f2-934f-8b0f26f6c4af/registry-server/1.log" Mar 13 11:06:56 crc kubenswrapper[4632]: I0313 11:06:56.675654 4632 generic.go:334] "Generic (PLEG): container finished" podID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerID="0bbbe65ea71f36a37f33d902708fe700b15c322c14c94c121a3ca523a54d026b" exitCode=137 Mar 13 11:06:56 crc kubenswrapper[4632]: I0313 11:06:56.675710 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6fq" event={"ID":"269ac923-f4f9-43f2-934f-8b0f26f6c4af","Type":"ContainerDied","Data":"0bbbe65ea71f36a37f33d902708fe700b15c322c14c94c121a3ca523a54d026b"} Mar 13 11:06:56 crc kubenswrapper[4632]: I0313 11:06:56.675772 4632 scope.go:117] "RemoveContainer" containerID="67b531d65834622b374c34e759c46150ba93cade0961705aa2b576c0c27e19d2" Mar 13 11:06:57 crc kubenswrapper[4632]: I0313 11:06:57.688344 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mm6fq_269ac923-f4f9-43f2-934f-8b0f26f6c4af/registry-server/1.log" Mar 13 11:06:57 crc kubenswrapper[4632]: I0313 11:06:57.689402 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6fq" event={"ID":"269ac923-f4f9-43f2-934f-8b0f26f6c4af","Type":"ContainerStarted","Data":"261dd98ddf33ee923d82b99030fb5e045cfb4833509d7ef9ec05e635f3f13122"} Mar 13 11:06:58 crc kubenswrapper[4632]: I0313 11:06:58.054536 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:06:58 crc kubenswrapper[4632]: E0313 11:06:58.054895 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:07:04 crc kubenswrapper[4632]: I0313 11:07:04.671139 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:07:04 crc kubenswrapper[4632]: I0313 11:07:04.671585 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:07:05 crc kubenswrapper[4632]: I0313 11:07:05.720977 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:07:05 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:07:05 crc kubenswrapper[4632]: > Mar 13 11:07:12 crc kubenswrapper[4632]: I0313 11:07:12.044982 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:07:12 crc kubenswrapper[4632]: E0313 11:07:12.045778 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:07:15 crc kubenswrapper[4632]: I0313 11:07:15.378361 4632 scope.go:117] "RemoveContainer" containerID="b3d4b9e8bcea3a6dbdeee6316ce9071df3a8c8906a4c416a00caede29a1de5ca" Mar 13 11:07:15 crc kubenswrapper[4632]: I0313 11:07:15.713488 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:07:15 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:07:15 crc kubenswrapper[4632]: > Mar 13 11:07:18 crc kubenswrapper[4632]: I0313 11:07:18.075276 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 11:07:23 crc kubenswrapper[4632]: I0313 11:07:23.044923 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:07:23 crc kubenswrapper[4632]: E0313 11:07:23.046713 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:07:25 crc kubenswrapper[4632]: I0313 11:07:25.731298 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:07:25 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:07:25 crc kubenswrapper[4632]: > Mar 13 11:07:35 crc kubenswrapper[4632]: I0313 11:07:35.728292 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" probeResult="failure" output=< Mar 13 11:07:35 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:07:35 crc kubenswrapper[4632]: > Mar 13 11:07:36 crc kubenswrapper[4632]: I0313 11:07:36.044352 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:07:36 crc kubenswrapper[4632]: E0313 11:07:36.044589 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:07:44 crc kubenswrapper[4632]: I0313 11:07:44.732070 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:07:44 crc kubenswrapper[4632]: I0313 11:07:44.803341 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:07:45 crc kubenswrapper[4632]: I0313 11:07:45.035017 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mm6fq"] Mar 13 11:07:46 crc kubenswrapper[4632]: I0313 11:07:46.203484 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mm6fq" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" containerID="cri-o://261dd98ddf33ee923d82b99030fb5e045cfb4833509d7ef9ec05e635f3f13122" gracePeriod=2 Mar 13 11:07:47 crc kubenswrapper[4632]: I0313 11:07:47.216873 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mm6fq_269ac923-f4f9-43f2-934f-8b0f26f6c4af/registry-server/1.log" Mar 13 11:07:47 crc kubenswrapper[4632]: I0313 11:07:47.218536 4632 generic.go:334] "Generic (PLEG): container finished" podID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerID="261dd98ddf33ee923d82b99030fb5e045cfb4833509d7ef9ec05e635f3f13122" exitCode=0 Mar 13 11:07:47 crc kubenswrapper[4632]: I0313 11:07:47.218574 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6fq" event={"ID":"269ac923-f4f9-43f2-934f-8b0f26f6c4af","Type":"ContainerDied","Data":"261dd98ddf33ee923d82b99030fb5e045cfb4833509d7ef9ec05e635f3f13122"} Mar 13 11:07:47 crc kubenswrapper[4632]: I0313 11:07:47.218624 4632 scope.go:117] "RemoveContainer" containerID="0bbbe65ea71f36a37f33d902708fe700b15c322c14c94c121a3ca523a54d026b" Mar 13 11:07:47 crc kubenswrapper[4632]: I0313 11:07:47.732008 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:07:47 crc kubenswrapper[4632]: I0313 11:07:47.780923 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/269ac923-f4f9-43f2-934f-8b0f26f6c4af-utilities\") pod \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\" (UID: \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\") " Mar 13 11:07:47 crc kubenswrapper[4632]: I0313 11:07:47.781002 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9tfk7\" (UniqueName: \"kubernetes.io/projected/269ac923-f4f9-43f2-934f-8b0f26f6c4af-kube-api-access-9tfk7\") pod \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\" (UID: \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\") " Mar 13 11:07:47 crc kubenswrapper[4632]: I0313 11:07:47.781037 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/269ac923-f4f9-43f2-934f-8b0f26f6c4af-catalog-content\") pod \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\" (UID: \"269ac923-f4f9-43f2-934f-8b0f26f6c4af\") " Mar 13 11:07:47 crc kubenswrapper[4632]: I0313 11:07:47.782519 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/269ac923-f4f9-43f2-934f-8b0f26f6c4af-utilities" (OuterVolumeSpecName: "utilities") pod "269ac923-f4f9-43f2-934f-8b0f26f6c4af" (UID: "269ac923-f4f9-43f2-934f-8b0f26f6c4af"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:07:47 crc kubenswrapper[4632]: I0313 11:07:47.802566 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/269ac923-f4f9-43f2-934f-8b0f26f6c4af-kube-api-access-9tfk7" (OuterVolumeSpecName: "kube-api-access-9tfk7") pod "269ac923-f4f9-43f2-934f-8b0f26f6c4af" (UID: "269ac923-f4f9-43f2-934f-8b0f26f6c4af"). InnerVolumeSpecName "kube-api-access-9tfk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:07:47 crc kubenswrapper[4632]: I0313 11:07:47.883591 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/269ac923-f4f9-43f2-934f-8b0f26f6c4af-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:07:47 crc kubenswrapper[4632]: I0313 11:07:47.883625 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9tfk7\" (UniqueName: \"kubernetes.io/projected/269ac923-f4f9-43f2-934f-8b0f26f6c4af-kube-api-access-9tfk7\") on node \"crc\" DevicePath \"\"" Mar 13 11:07:47 crc kubenswrapper[4632]: I0313 11:07:47.910317 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/269ac923-f4f9-43f2-934f-8b0f26f6c4af-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "269ac923-f4f9-43f2-934f-8b0f26f6c4af" (UID: "269ac923-f4f9-43f2-934f-8b0f26f6c4af"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:07:47 crc kubenswrapper[4632]: I0313 11:07:47.985605 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/269ac923-f4f9-43f2-934f-8b0f26f6c4af-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:07:48 crc kubenswrapper[4632]: I0313 11:07:48.058560 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:07:48 crc kubenswrapper[4632]: E0313 11:07:48.059278 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:07:48 crc kubenswrapper[4632]: I0313 11:07:48.264295 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mm6fq" event={"ID":"269ac923-f4f9-43f2-934f-8b0f26f6c4af","Type":"ContainerDied","Data":"c53342fac7b10f1bef54be90bcf0e83cc2e423f561f4ea27cefd68e4947d5bb4"} Mar 13 11:07:48 crc kubenswrapper[4632]: I0313 11:07:48.266049 4632 scope.go:117] "RemoveContainer" containerID="261dd98ddf33ee923d82b99030fb5e045cfb4833509d7ef9ec05e635f3f13122" Mar 13 11:07:48 crc kubenswrapper[4632]: I0313 11:07:48.264621 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mm6fq" Mar 13 11:07:48 crc kubenswrapper[4632]: I0313 11:07:48.314251 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mm6fq"] Mar 13 11:07:48 crc kubenswrapper[4632]: I0313 11:07:48.323443 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mm6fq"] Mar 13 11:07:48 crc kubenswrapper[4632]: I0313 11:07:48.331446 4632 scope.go:117] "RemoveContainer" containerID="d03ceb30503b22a2ad94cc53347c3b0ae54c134bb2b9db1bd0c47dcfc27a8ece" Mar 13 11:07:48 crc kubenswrapper[4632]: I0313 11:07:48.360190 4632 scope.go:117] "RemoveContainer" containerID="93ddcb9911b3bbd33b20a1520c077d1ce20ed42dceb52f18631471d802d7e139" Mar 13 11:07:50 crc kubenswrapper[4632]: I0313 11:07:50.061709 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" path="/var/lib/kubelet/pods/269ac923-f4f9-43f2-934f-8b0f26f6c4af/volumes" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.014171 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-59586ff4c9-s4xn7"] Mar 13 11:07:52 crc kubenswrapper[4632]: E0313 11:07:52.024304 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.024374 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" Mar 13 11:07:52 crc kubenswrapper[4632]: E0313 11:07:52.024435 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.024442 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" Mar 13 11:07:52 crc kubenswrapper[4632]: E0313 11:07:52.024452 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="extract-utilities" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.024460 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="extract-utilities" Mar 13 11:07:52 crc kubenswrapper[4632]: E0313 11:07:52.024486 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.024492 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" Mar 13 11:07:52 crc kubenswrapper[4632]: E0313 11:07:52.024511 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="extract-content" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.024519 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="extract-content" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.024864 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.024881 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.025326 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="269ac923-f4f9-43f2-934f-8b0f26f6c4af" containerName="registry-server" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.031506 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.058919 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-59586ff4c9-s4xn7"] Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.077066 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llcp8\" (UniqueName: \"kubernetes.io/projected/8b9495c7-c9ae-4a07-b216-a250d4cd274e-kube-api-access-llcp8\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.077369 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-config\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.077490 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-internal-tls-certs\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.077770 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-httpd-config\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.078495 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-public-tls-certs\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.078550 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-ovndb-tls-certs\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.078904 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-combined-ca-bundle\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.180641 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-httpd-config\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.180732 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-public-tls-certs\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.180759 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-ovndb-tls-certs\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.180858 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-combined-ca-bundle\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.180992 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llcp8\" (UniqueName: \"kubernetes.io/projected/8b9495c7-c9ae-4a07-b216-a250d4cd274e-kube-api-access-llcp8\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.181052 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-config\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.181077 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-internal-tls-certs\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.192895 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-ovndb-tls-certs\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.192954 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-public-tls-certs\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.193469 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-combined-ca-bundle\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.193572 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-httpd-config\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.193971 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-internal-tls-certs\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.199732 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-config\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.204021 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llcp8\" (UniqueName: \"kubernetes.io/projected/8b9495c7-c9ae-4a07-b216-a250d4cd274e-kube-api-access-llcp8\") pod \"neutron-59586ff4c9-s4xn7\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:52 crc kubenswrapper[4632]: I0313 11:07:52.355394 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:53 crc kubenswrapper[4632]: I0313 11:07:53.830914 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-59586ff4c9-s4xn7"] Mar 13 11:07:54 crc kubenswrapper[4632]: I0313 11:07:54.323411 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59586ff4c9-s4xn7" event={"ID":"8b9495c7-c9ae-4a07-b216-a250d4cd274e","Type":"ContainerStarted","Data":"94fc75b5bf96292690ce359a5d4ce65dd30bc2b06b1aeb4d309bd6e1dcd7e70c"} Mar 13 11:07:54 crc kubenswrapper[4632]: I0313 11:07:54.324115 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59586ff4c9-s4xn7" event={"ID":"8b9495c7-c9ae-4a07-b216-a250d4cd274e","Type":"ContainerStarted","Data":"baa77c1d37fb9c8cc82676bdaaab769c07869647e902c21581eef67e591e5d68"} Mar 13 11:07:55 crc kubenswrapper[4632]: I0313 11:07:55.349608 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59586ff4c9-s4xn7" event={"ID":"8b9495c7-c9ae-4a07-b216-a250d4cd274e","Type":"ContainerStarted","Data":"82e1ea3147a5e24713a581b7fd1d1be6dc38543edaf91f1fa20ce5282f06b072"} Mar 13 11:07:55 crc kubenswrapper[4632]: I0313 11:07:55.349848 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:07:55 crc kubenswrapper[4632]: I0313 11:07:55.390100 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-59586ff4c9-s4xn7" podStartSLOduration=4.390085623 podStartE2EDuration="4.390085623s" podCreationTimestamp="2026-03-13 11:07:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 11:07:55.387303574 +0000 UTC m=+3849.409833707" watchObservedRunningTime="2026-03-13 11:07:55.390085623 +0000 UTC m=+3849.412615756" Mar 13 11:07:59 crc kubenswrapper[4632]: I0313 11:07:59.045497 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:07:59 crc kubenswrapper[4632]: E0313 11:07:59.046526 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:08:00 crc kubenswrapper[4632]: I0313 11:08:00.180279 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556668-bzw4c"] Mar 13 11:08:00 crc kubenswrapper[4632]: I0313 11:08:00.182177 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556668-bzw4c" Mar 13 11:08:00 crc kubenswrapper[4632]: I0313 11:08:00.195636 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556668-bzw4c"] Mar 13 11:08:00 crc kubenswrapper[4632]: I0313 11:08:00.199832 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:08:00 crc kubenswrapper[4632]: I0313 11:08:00.199834 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:08:00 crc kubenswrapper[4632]: I0313 11:08:00.200023 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:08:00 crc kubenswrapper[4632]: I0313 11:08:00.286630 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jm97k\" (UniqueName: \"kubernetes.io/projected/4d08697f-ce87-4d33-823a-9bf5d2d0d801-kube-api-access-jm97k\") pod \"auto-csr-approver-29556668-bzw4c\" (UID: \"4d08697f-ce87-4d33-823a-9bf5d2d0d801\") " pod="openshift-infra/auto-csr-approver-29556668-bzw4c" Mar 13 11:08:00 crc kubenswrapper[4632]: I0313 11:08:00.388735 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jm97k\" (UniqueName: \"kubernetes.io/projected/4d08697f-ce87-4d33-823a-9bf5d2d0d801-kube-api-access-jm97k\") pod \"auto-csr-approver-29556668-bzw4c\" (UID: \"4d08697f-ce87-4d33-823a-9bf5d2d0d801\") " pod="openshift-infra/auto-csr-approver-29556668-bzw4c" Mar 13 11:08:00 crc kubenswrapper[4632]: I0313 11:08:00.427825 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jm97k\" (UniqueName: \"kubernetes.io/projected/4d08697f-ce87-4d33-823a-9bf5d2d0d801-kube-api-access-jm97k\") pod \"auto-csr-approver-29556668-bzw4c\" (UID: \"4d08697f-ce87-4d33-823a-9bf5d2d0d801\") " pod="openshift-infra/auto-csr-approver-29556668-bzw4c" Mar 13 11:08:00 crc kubenswrapper[4632]: I0313 11:08:00.513865 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556668-bzw4c" Mar 13 11:08:01 crc kubenswrapper[4632]: I0313 11:08:01.001572 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556668-bzw4c"] Mar 13 11:08:01 crc kubenswrapper[4632]: I0313 11:08:01.008320 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 11:08:01 crc kubenswrapper[4632]: I0313 11:08:01.408911 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556668-bzw4c" event={"ID":"4d08697f-ce87-4d33-823a-9bf5d2d0d801","Type":"ContainerStarted","Data":"f023d21fce137ea242abc1719cf5d0fd4ebe997c19eab16e9956070cc8c44339"} Mar 13 11:08:03 crc kubenswrapper[4632]: I0313 11:08:03.438166 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556668-bzw4c" event={"ID":"4d08697f-ce87-4d33-823a-9bf5d2d0d801","Type":"ContainerStarted","Data":"cdc6d57bdf1672f0d8614c97a39233ec4a346ac1edaab1837f60116110b310ef"} Mar 13 11:08:03 crc kubenswrapper[4632]: I0313 11:08:03.471779 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556668-bzw4c" podStartSLOduration=2.415084248 podStartE2EDuration="3.471759343s" podCreationTimestamp="2026-03-13 11:08:00 +0000 UTC" firstStartedPulling="2026-03-13 11:08:01.004722876 +0000 UTC m=+3855.027253009" lastFinishedPulling="2026-03-13 11:08:02.061397971 +0000 UTC m=+3856.083928104" observedRunningTime="2026-03-13 11:08:03.457736996 +0000 UTC m=+3857.480267199" watchObservedRunningTime="2026-03-13 11:08:03.471759343 +0000 UTC m=+3857.494289476" Mar 13 11:08:04 crc kubenswrapper[4632]: I0313 11:08:04.455986 4632 generic.go:334] "Generic (PLEG): container finished" podID="4d08697f-ce87-4d33-823a-9bf5d2d0d801" containerID="cdc6d57bdf1672f0d8614c97a39233ec4a346ac1edaab1837f60116110b310ef" exitCode=0 Mar 13 11:08:04 crc kubenswrapper[4632]: I0313 11:08:04.456297 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556668-bzw4c" event={"ID":"4d08697f-ce87-4d33-823a-9bf5d2d0d801","Type":"ContainerDied","Data":"cdc6d57bdf1672f0d8614c97a39233ec4a346ac1edaab1837f60116110b310ef"} Mar 13 11:08:05 crc kubenswrapper[4632]: I0313 11:08:05.890756 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556668-bzw4c" Mar 13 11:08:06 crc kubenswrapper[4632]: I0313 11:08:06.007494 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jm97k\" (UniqueName: \"kubernetes.io/projected/4d08697f-ce87-4d33-823a-9bf5d2d0d801-kube-api-access-jm97k\") pod \"4d08697f-ce87-4d33-823a-9bf5d2d0d801\" (UID: \"4d08697f-ce87-4d33-823a-9bf5d2d0d801\") " Mar 13 11:08:06 crc kubenswrapper[4632]: I0313 11:08:06.016270 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d08697f-ce87-4d33-823a-9bf5d2d0d801-kube-api-access-jm97k" (OuterVolumeSpecName: "kube-api-access-jm97k") pod "4d08697f-ce87-4d33-823a-9bf5d2d0d801" (UID: "4d08697f-ce87-4d33-823a-9bf5d2d0d801"). InnerVolumeSpecName "kube-api-access-jm97k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:08:06 crc kubenswrapper[4632]: I0313 11:08:06.109875 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jm97k\" (UniqueName: \"kubernetes.io/projected/4d08697f-ce87-4d33-823a-9bf5d2d0d801-kube-api-access-jm97k\") on node \"crc\" DevicePath \"\"" Mar 13 11:08:06 crc kubenswrapper[4632]: I0313 11:08:06.481404 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556668-bzw4c" event={"ID":"4d08697f-ce87-4d33-823a-9bf5d2d0d801","Type":"ContainerDied","Data":"f023d21fce137ea242abc1719cf5d0fd4ebe997c19eab16e9956070cc8c44339"} Mar 13 11:08:06 crc kubenswrapper[4632]: I0313 11:08:06.481504 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556668-bzw4c" Mar 13 11:08:06 crc kubenswrapper[4632]: I0313 11:08:06.482498 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f023d21fce137ea242abc1719cf5d0fd4ebe997c19eab16e9956070cc8c44339" Mar 13 11:08:06 crc kubenswrapper[4632]: I0313 11:08:06.593988 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556662-pw9tk"] Mar 13 11:08:06 crc kubenswrapper[4632]: I0313 11:08:06.601177 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556662-pw9tk"] Mar 13 11:08:08 crc kubenswrapper[4632]: I0313 11:08:08.060738 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="62b0f696-9e5c-4535-a181-fa2f4b645711" path="/var/lib/kubelet/pods/62b0f696-9e5c-4535-a181-fa2f4b645711/volumes" Mar 13 11:08:11 crc kubenswrapper[4632]: I0313 11:08:11.044560 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:08:11 crc kubenswrapper[4632]: E0313 11:08:11.045490 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:08:15 crc kubenswrapper[4632]: I0313 11:08:15.658751 4632 scope.go:117] "RemoveContainer" containerID="69d080c6683237a330690584133c6005521df29f2dcf4c21ed9a518e4de4e991" Mar 13 11:08:22 crc kubenswrapper[4632]: I0313 11:08:22.368142 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 11:08:22 crc kubenswrapper[4632]: I0313 11:08:22.475209 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6588559b77-6f4bf"] Mar 13 11:08:22 crc kubenswrapper[4632]: I0313 11:08:22.476069 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6588559b77-6f4bf" podUID="79498b99-6b5c-4a95-8558-5d615fc7abba" containerName="neutron-api" containerID="cri-o://a37056b823559676b78bbdad36e07fb68a02ab13bf670546d16508926857a154" gracePeriod=30 Mar 13 11:08:22 crc kubenswrapper[4632]: I0313 11:08:22.476235 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6588559b77-6f4bf" podUID="79498b99-6b5c-4a95-8558-5d615fc7abba" containerName="neutron-httpd" containerID="cri-o://2e4dbe726a115e20d5697b52cbd987856c78465356a65ffaf180382482e42ad0" gracePeriod=30 Mar 13 11:08:23 crc kubenswrapper[4632]: I0313 11:08:23.677970 4632 generic.go:334] "Generic (PLEG): container finished" podID="79498b99-6b5c-4a95-8558-5d615fc7abba" containerID="2e4dbe726a115e20d5697b52cbd987856c78465356a65ffaf180382482e42ad0" exitCode=0 Mar 13 11:08:23 crc kubenswrapper[4632]: I0313 11:08:23.678079 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6588559b77-6f4bf" event={"ID":"79498b99-6b5c-4a95-8558-5d615fc7abba","Type":"ContainerDied","Data":"2e4dbe726a115e20d5697b52cbd987856c78465356a65ffaf180382482e42ad0"} Mar 13 11:08:25 crc kubenswrapper[4632]: I0313 11:08:25.043787 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:08:25 crc kubenswrapper[4632]: E0313 11:08:25.044554 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:08:25 crc kubenswrapper[4632]: I0313 11:08:25.115149 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-6588559b77-6f4bf" podUID="79498b99-6b5c-4a95-8558-5d615fc7abba" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.174:9696/\": dial tcp 10.217.0.174:9696: connect: connection refused" Mar 13 11:08:34 crc kubenswrapper[4632]: I0313 11:08:34.812638 4632 generic.go:334] "Generic (PLEG): container finished" podID="79498b99-6b5c-4a95-8558-5d615fc7abba" containerID="a37056b823559676b78bbdad36e07fb68a02ab13bf670546d16508926857a154" exitCode=0 Mar 13 11:08:34 crc kubenswrapper[4632]: I0313 11:08:34.813214 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6588559b77-6f4bf" event={"ID":"79498b99-6b5c-4a95-8558-5d615fc7abba","Type":"ContainerDied","Data":"a37056b823559676b78bbdad36e07fb68a02ab13bf670546d16508926857a154"} Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.153483 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6588559b77-6f4bf" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.341199 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-ovndb-tls-certs\") pod \"79498b99-6b5c-4a95-8558-5d615fc7abba\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.341254 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-public-tls-certs\") pod \"79498b99-6b5c-4a95-8558-5d615fc7abba\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.341283 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g98kg\" (UniqueName: \"kubernetes.io/projected/79498b99-6b5c-4a95-8558-5d615fc7abba-kube-api-access-g98kg\") pod \"79498b99-6b5c-4a95-8558-5d615fc7abba\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.341340 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-config\") pod \"79498b99-6b5c-4a95-8558-5d615fc7abba\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.341407 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-httpd-config\") pod \"79498b99-6b5c-4a95-8558-5d615fc7abba\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.341472 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-combined-ca-bundle\") pod \"79498b99-6b5c-4a95-8558-5d615fc7abba\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.341572 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-internal-tls-certs\") pod \"79498b99-6b5c-4a95-8558-5d615fc7abba\" (UID: \"79498b99-6b5c-4a95-8558-5d615fc7abba\") " Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.355744 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79498b99-6b5c-4a95-8558-5d615fc7abba-kube-api-access-g98kg" (OuterVolumeSpecName: "kube-api-access-g98kg") pod "79498b99-6b5c-4a95-8558-5d615fc7abba" (UID: "79498b99-6b5c-4a95-8558-5d615fc7abba"). InnerVolumeSpecName "kube-api-access-g98kg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.356283 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "79498b99-6b5c-4a95-8558-5d615fc7abba" (UID: "79498b99-6b5c-4a95-8558-5d615fc7abba"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.426130 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "79498b99-6b5c-4a95-8558-5d615fc7abba" (UID: "79498b99-6b5c-4a95-8558-5d615fc7abba"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.428131 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "79498b99-6b5c-4a95-8558-5d615fc7abba" (UID: "79498b99-6b5c-4a95-8558-5d615fc7abba"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.434970 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-config" (OuterVolumeSpecName: "config") pod "79498b99-6b5c-4a95-8558-5d615fc7abba" (UID: "79498b99-6b5c-4a95-8558-5d615fc7abba"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.440095 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "79498b99-6b5c-4a95-8558-5d615fc7abba" (UID: "79498b99-6b5c-4a95-8558-5d615fc7abba"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.445879 4632 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.445927 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.445963 4632 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.446004 4632 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.446017 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g98kg\" (UniqueName: \"kubernetes.io/projected/79498b99-6b5c-4a95-8558-5d615fc7abba-kube-api-access-g98kg\") on node \"crc\" DevicePath \"\"" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.446032 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-config\") on node \"crc\" DevicePath \"\"" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.448728 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "79498b99-6b5c-4a95-8558-5d615fc7abba" (UID: "79498b99-6b5c-4a95-8558-5d615fc7abba"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.547679 4632 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/79498b99-6b5c-4a95-8558-5d615fc7abba-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.826148 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6588559b77-6f4bf" event={"ID":"79498b99-6b5c-4a95-8558-5d615fc7abba","Type":"ContainerDied","Data":"cc3aa5e44b0dc25bbbe479e7210125c65a55be4449da25f59fdeef0322a73ed3"} Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.826209 4632 scope.go:117] "RemoveContainer" containerID="2e4dbe726a115e20d5697b52cbd987856c78465356a65ffaf180382482e42ad0" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.826334 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6588559b77-6f4bf" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.877082 4632 scope.go:117] "RemoveContainer" containerID="a37056b823559676b78bbdad36e07fb68a02ab13bf670546d16508926857a154" Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.891302 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6588559b77-6f4bf"] Mar 13 11:08:35 crc kubenswrapper[4632]: I0313 11:08:35.901998 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6588559b77-6f4bf"] Mar 13 11:08:36 crc kubenswrapper[4632]: I0313 11:08:36.056002 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79498b99-6b5c-4a95-8558-5d615fc7abba" path="/var/lib/kubelet/pods/79498b99-6b5c-4a95-8558-5d615fc7abba/volumes" Mar 13 11:08:38 crc kubenswrapper[4632]: I0313 11:08:38.058175 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:08:38 crc kubenswrapper[4632]: E0313 11:08:38.059129 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:08:50 crc kubenswrapper[4632]: I0313 11:08:50.046425 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:08:50 crc kubenswrapper[4632]: E0313 11:08:50.047434 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:09:01 crc kubenswrapper[4632]: I0313 11:09:01.044724 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:09:01 crc kubenswrapper[4632]: E0313 11:09:01.046037 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:09:15 crc kubenswrapper[4632]: I0313 11:09:15.044419 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:09:15 crc kubenswrapper[4632]: E0313 11:09:15.045653 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:09:27 crc kubenswrapper[4632]: I0313 11:09:27.044444 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:09:27 crc kubenswrapper[4632]: E0313 11:09:27.045019 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:09:38 crc kubenswrapper[4632]: I0313 11:09:38.051423 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:09:38 crc kubenswrapper[4632]: E0313 11:09:38.052465 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:09:52 crc kubenswrapper[4632]: I0313 11:09:52.044878 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:09:52 crc kubenswrapper[4632]: E0313 11:09:52.046274 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.151532 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556670-fqvf6"] Mar 13 11:10:00 crc kubenswrapper[4632]: E0313 11:10:00.152381 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79498b99-6b5c-4a95-8558-5d615fc7abba" containerName="neutron-httpd" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.152394 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="79498b99-6b5c-4a95-8558-5d615fc7abba" containerName="neutron-httpd" Mar 13 11:10:00 crc kubenswrapper[4632]: E0313 11:10:00.152409 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79498b99-6b5c-4a95-8558-5d615fc7abba" containerName="neutron-api" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.152427 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="79498b99-6b5c-4a95-8558-5d615fc7abba" containerName="neutron-api" Mar 13 11:10:00 crc kubenswrapper[4632]: E0313 11:10:00.152458 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d08697f-ce87-4d33-823a-9bf5d2d0d801" containerName="oc" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.152466 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d08697f-ce87-4d33-823a-9bf5d2d0d801" containerName="oc" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.152637 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="79498b99-6b5c-4a95-8558-5d615fc7abba" containerName="neutron-api" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.152656 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d08697f-ce87-4d33-823a-9bf5d2d0d801" containerName="oc" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.152671 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="79498b99-6b5c-4a95-8558-5d615fc7abba" containerName="neutron-httpd" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.153323 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556670-fqvf6" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.155362 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.155674 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.156503 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.168686 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556670-fqvf6"] Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.290489 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgtnj\" (UniqueName: \"kubernetes.io/projected/565a5983-3957-42c2-b7d4-47d26e00aec8-kube-api-access-bgtnj\") pod \"auto-csr-approver-29556670-fqvf6\" (UID: \"565a5983-3957-42c2-b7d4-47d26e00aec8\") " pod="openshift-infra/auto-csr-approver-29556670-fqvf6" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.393258 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgtnj\" (UniqueName: \"kubernetes.io/projected/565a5983-3957-42c2-b7d4-47d26e00aec8-kube-api-access-bgtnj\") pod \"auto-csr-approver-29556670-fqvf6\" (UID: \"565a5983-3957-42c2-b7d4-47d26e00aec8\") " pod="openshift-infra/auto-csr-approver-29556670-fqvf6" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.420762 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgtnj\" (UniqueName: \"kubernetes.io/projected/565a5983-3957-42c2-b7d4-47d26e00aec8-kube-api-access-bgtnj\") pod \"auto-csr-approver-29556670-fqvf6\" (UID: \"565a5983-3957-42c2-b7d4-47d26e00aec8\") " pod="openshift-infra/auto-csr-approver-29556670-fqvf6" Mar 13 11:10:00 crc kubenswrapper[4632]: I0313 11:10:00.474517 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556670-fqvf6" Mar 13 11:10:01 crc kubenswrapper[4632]: I0313 11:10:00.999858 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556670-fqvf6"] Mar 13 11:10:01 crc kubenswrapper[4632]: I0313 11:10:01.764580 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556670-fqvf6" event={"ID":"565a5983-3957-42c2-b7d4-47d26e00aec8","Type":"ContainerStarted","Data":"c51ebc91b28459b756985f47316bb878f71da3f694f51215dc5d353dc6155f3f"} Mar 13 11:10:02 crc kubenswrapper[4632]: I0313 11:10:02.776911 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556670-fqvf6" event={"ID":"565a5983-3957-42c2-b7d4-47d26e00aec8","Type":"ContainerStarted","Data":"c0218119e7ac388fadab5a0e90f8eec2d8161ed6d34eaec2b46cb615f7e41508"} Mar 13 11:10:02 crc kubenswrapper[4632]: I0313 11:10:02.804183 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556670-fqvf6" podStartSLOduration=1.695292517 podStartE2EDuration="2.804157199s" podCreationTimestamp="2026-03-13 11:10:00 +0000 UTC" firstStartedPulling="2026-03-13 11:10:01.012255691 +0000 UTC m=+3975.034785834" lastFinishedPulling="2026-03-13 11:10:02.121120373 +0000 UTC m=+3976.143650516" observedRunningTime="2026-03-13 11:10:02.793553699 +0000 UTC m=+3976.816083832" watchObservedRunningTime="2026-03-13 11:10:02.804157199 +0000 UTC m=+3976.826687332" Mar 13 11:10:03 crc kubenswrapper[4632]: I0313 11:10:03.792513 4632 generic.go:334] "Generic (PLEG): container finished" podID="565a5983-3957-42c2-b7d4-47d26e00aec8" containerID="c0218119e7ac388fadab5a0e90f8eec2d8161ed6d34eaec2b46cb615f7e41508" exitCode=0 Mar 13 11:10:03 crc kubenswrapper[4632]: I0313 11:10:03.792622 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556670-fqvf6" event={"ID":"565a5983-3957-42c2-b7d4-47d26e00aec8","Type":"ContainerDied","Data":"c0218119e7ac388fadab5a0e90f8eec2d8161ed6d34eaec2b46cb615f7e41508"} Mar 13 11:10:05 crc kubenswrapper[4632]: I0313 11:10:05.254909 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556670-fqvf6" Mar 13 11:10:05 crc kubenswrapper[4632]: I0313 11:10:05.286318 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgtnj\" (UniqueName: \"kubernetes.io/projected/565a5983-3957-42c2-b7d4-47d26e00aec8-kube-api-access-bgtnj\") pod \"565a5983-3957-42c2-b7d4-47d26e00aec8\" (UID: \"565a5983-3957-42c2-b7d4-47d26e00aec8\") " Mar 13 11:10:05 crc kubenswrapper[4632]: I0313 11:10:05.293895 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/565a5983-3957-42c2-b7d4-47d26e00aec8-kube-api-access-bgtnj" (OuterVolumeSpecName: "kube-api-access-bgtnj") pod "565a5983-3957-42c2-b7d4-47d26e00aec8" (UID: "565a5983-3957-42c2-b7d4-47d26e00aec8"). InnerVolumeSpecName "kube-api-access-bgtnj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:10:05 crc kubenswrapper[4632]: I0313 11:10:05.388674 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgtnj\" (UniqueName: \"kubernetes.io/projected/565a5983-3957-42c2-b7d4-47d26e00aec8-kube-api-access-bgtnj\") on node \"crc\" DevicePath \"\"" Mar 13 11:10:05 crc kubenswrapper[4632]: I0313 11:10:05.819159 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556670-fqvf6" event={"ID":"565a5983-3957-42c2-b7d4-47d26e00aec8","Type":"ContainerDied","Data":"c51ebc91b28459b756985f47316bb878f71da3f694f51215dc5d353dc6155f3f"} Mar 13 11:10:05 crc kubenswrapper[4632]: I0313 11:10:05.819216 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c51ebc91b28459b756985f47316bb878f71da3f694f51215dc5d353dc6155f3f" Mar 13 11:10:05 crc kubenswrapper[4632]: I0313 11:10:05.819292 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556670-fqvf6" Mar 13 11:10:05 crc kubenswrapper[4632]: I0313 11:10:05.865243 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556664-vgmtg"] Mar 13 11:10:05 crc kubenswrapper[4632]: I0313 11:10:05.874688 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556664-vgmtg"] Mar 13 11:10:06 crc kubenswrapper[4632]: I0313 11:10:06.077799 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ecca46c-1e06-43be-bacc-eae4a1a474b7" path="/var/lib/kubelet/pods/6ecca46c-1e06-43be-bacc-eae4a1a474b7/volumes" Mar 13 11:10:07 crc kubenswrapper[4632]: I0313 11:10:07.044180 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:10:07 crc kubenswrapper[4632]: E0313 11:10:07.044760 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:10:15 crc kubenswrapper[4632]: I0313 11:10:15.927543 4632 scope.go:117] "RemoveContainer" containerID="5e0a7ac81434eac7eff8520645fc1fc30caa50af82d06bce9d4415863d0b9aa2" Mar 13 11:10:20 crc kubenswrapper[4632]: I0313 11:10:20.045974 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:10:20 crc kubenswrapper[4632]: E0313 11:10:20.046652 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:10:31 crc kubenswrapper[4632]: I0313 11:10:31.044478 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:10:31 crc kubenswrapper[4632]: E0313 11:10:31.045132 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:10:42 crc kubenswrapper[4632]: I0313 11:10:42.047283 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:10:43 crc kubenswrapper[4632]: I0313 11:10:43.320216 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"908510815c251e300c3555d9b7458818dfba317c0679f487df71c717e5c832f9"} Mar 13 11:11:16 crc kubenswrapper[4632]: I0313 11:11:16.030034 4632 scope.go:117] "RemoveContainer" containerID="c4e602b48052cce5414a1759fd3d99f56ebde469321edc3b351de56e308a589e" Mar 13 11:11:16 crc kubenswrapper[4632]: I0313 11:11:16.098915 4632 scope.go:117] "RemoveContainer" containerID="e7435d3cb1416970cf6b4162802419aa1bcf01c76e855132a1393d7b353e8c78" Mar 13 11:11:16 crc kubenswrapper[4632]: I0313 11:11:16.137372 4632 scope.go:117] "RemoveContainer" containerID="c772f3e31ea57b90bb99ca7ef746eba3e104a41a227d8c675887f2261f06ab48" Mar 13 11:12:00 crc kubenswrapper[4632]: I0313 11:12:00.151480 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556672-sd9zg"] Mar 13 11:12:00 crc kubenswrapper[4632]: E0313 11:12:00.152258 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="565a5983-3957-42c2-b7d4-47d26e00aec8" containerName="oc" Mar 13 11:12:00 crc kubenswrapper[4632]: I0313 11:12:00.152270 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="565a5983-3957-42c2-b7d4-47d26e00aec8" containerName="oc" Mar 13 11:12:00 crc kubenswrapper[4632]: I0313 11:12:00.152489 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="565a5983-3957-42c2-b7d4-47d26e00aec8" containerName="oc" Mar 13 11:12:00 crc kubenswrapper[4632]: I0313 11:12:00.153082 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556672-sd9zg" Mar 13 11:12:00 crc kubenswrapper[4632]: I0313 11:12:00.155693 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:12:00 crc kubenswrapper[4632]: I0313 11:12:00.155860 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:12:00 crc kubenswrapper[4632]: I0313 11:12:00.161771 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:12:00 crc kubenswrapper[4632]: I0313 11:12:00.181031 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556672-sd9zg"] Mar 13 11:12:00 crc kubenswrapper[4632]: I0313 11:12:00.238198 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jtfs\" (UniqueName: \"kubernetes.io/projected/bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8-kube-api-access-5jtfs\") pod \"auto-csr-approver-29556672-sd9zg\" (UID: \"bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8\") " pod="openshift-infra/auto-csr-approver-29556672-sd9zg" Mar 13 11:12:00 crc kubenswrapper[4632]: I0313 11:12:00.340347 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jtfs\" (UniqueName: \"kubernetes.io/projected/bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8-kube-api-access-5jtfs\") pod \"auto-csr-approver-29556672-sd9zg\" (UID: \"bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8\") " pod="openshift-infra/auto-csr-approver-29556672-sd9zg" Mar 13 11:12:00 crc kubenswrapper[4632]: I0313 11:12:00.360904 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jtfs\" (UniqueName: \"kubernetes.io/projected/bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8-kube-api-access-5jtfs\") pod \"auto-csr-approver-29556672-sd9zg\" (UID: \"bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8\") " pod="openshift-infra/auto-csr-approver-29556672-sd9zg" Mar 13 11:12:00 crc kubenswrapper[4632]: I0313 11:12:00.485175 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556672-sd9zg" Mar 13 11:12:01 crc kubenswrapper[4632]: I0313 11:12:01.170987 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556672-sd9zg"] Mar 13 11:12:01 crc kubenswrapper[4632]: I0313 11:12:01.342864 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556672-sd9zg" event={"ID":"bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8","Type":"ContainerStarted","Data":"f2dda7430b6fcdef405e402c434627e812eece3b35eaae4c925ef5572f23bfe1"} Mar 13 11:12:03 crc kubenswrapper[4632]: I0313 11:12:03.363799 4632 generic.go:334] "Generic (PLEG): container finished" podID="bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8" containerID="5f687cba4c29fe06e8932802cf25f9e44ab270587540acfd69f26e45b584a52b" exitCode=0 Mar 13 11:12:03 crc kubenswrapper[4632]: I0313 11:12:03.363844 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556672-sd9zg" event={"ID":"bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8","Type":"ContainerDied","Data":"5f687cba4c29fe06e8932802cf25f9e44ab270587540acfd69f26e45b584a52b"} Mar 13 11:12:04 crc kubenswrapper[4632]: I0313 11:12:04.766053 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556672-sd9zg" Mar 13 11:12:04 crc kubenswrapper[4632]: I0313 11:12:04.848559 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5jtfs\" (UniqueName: \"kubernetes.io/projected/bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8-kube-api-access-5jtfs\") pod \"bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8\" (UID: \"bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8\") " Mar 13 11:12:04 crc kubenswrapper[4632]: I0313 11:12:04.856000 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8-kube-api-access-5jtfs" (OuterVolumeSpecName: "kube-api-access-5jtfs") pod "bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8" (UID: "bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8"). InnerVolumeSpecName "kube-api-access-5jtfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:12:04 crc kubenswrapper[4632]: I0313 11:12:04.951349 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5jtfs\" (UniqueName: \"kubernetes.io/projected/bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8-kube-api-access-5jtfs\") on node \"crc\" DevicePath \"\"" Mar 13 11:12:05 crc kubenswrapper[4632]: I0313 11:12:05.388125 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556672-sd9zg" event={"ID":"bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8","Type":"ContainerDied","Data":"f2dda7430b6fcdef405e402c434627e812eece3b35eaae4c925ef5572f23bfe1"} Mar 13 11:12:05 crc kubenswrapper[4632]: I0313 11:12:05.388468 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f2dda7430b6fcdef405e402c434627e812eece3b35eaae4c925ef5572f23bfe1" Mar 13 11:12:05 crc kubenswrapper[4632]: I0313 11:12:05.388276 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556672-sd9zg" Mar 13 11:12:05 crc kubenswrapper[4632]: I0313 11:12:05.842913 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556666-pjsh7"] Mar 13 11:12:05 crc kubenswrapper[4632]: I0313 11:12:05.855726 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556666-pjsh7"] Mar 13 11:12:06 crc kubenswrapper[4632]: I0313 11:12:06.059076 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c793856b-f941-4c9e-b70e-36b4844e4eac" path="/var/lib/kubelet/pods/c793856b-f941-4c9e-b70e-36b4844e4eac/volumes" Mar 13 11:12:16 crc kubenswrapper[4632]: I0313 11:12:16.201883 4632 scope.go:117] "RemoveContainer" containerID="ca3b5ee01f58147d02e068e963d2c27601cb82a563a21e34b84cdc61a03e2f80" Mar 13 11:13:10 crc kubenswrapper[4632]: I0313 11:13:10.460797 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:13:10 crc kubenswrapper[4632]: I0313 11:13:10.461458 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:13:40 crc kubenswrapper[4632]: I0313 11:13:40.461663 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:13:40 crc kubenswrapper[4632]: I0313 11:13:40.462353 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:14:00 crc kubenswrapper[4632]: I0313 11:14:00.146652 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556674-j6lkl"] Mar 13 11:14:00 crc kubenswrapper[4632]: E0313 11:14:00.147643 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8" containerName="oc" Mar 13 11:14:00 crc kubenswrapper[4632]: I0313 11:14:00.147659 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8" containerName="oc" Mar 13 11:14:00 crc kubenswrapper[4632]: I0313 11:14:00.147931 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8" containerName="oc" Mar 13 11:14:00 crc kubenswrapper[4632]: I0313 11:14:00.148688 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556674-j6lkl" Mar 13 11:14:00 crc kubenswrapper[4632]: I0313 11:14:00.151389 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:14:00 crc kubenswrapper[4632]: I0313 11:14:00.151803 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:14:00 crc kubenswrapper[4632]: I0313 11:14:00.151989 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:14:00 crc kubenswrapper[4632]: I0313 11:14:00.202787 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556674-j6lkl"] Mar 13 11:14:00 crc kubenswrapper[4632]: I0313 11:14:00.250314 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dm8b\" (UniqueName: \"kubernetes.io/projected/2607e7bb-5f81-48cf-945a-6dee68b60040-kube-api-access-8dm8b\") pod \"auto-csr-approver-29556674-j6lkl\" (UID: \"2607e7bb-5f81-48cf-945a-6dee68b60040\") " pod="openshift-infra/auto-csr-approver-29556674-j6lkl" Mar 13 11:14:00 crc kubenswrapper[4632]: I0313 11:14:00.352821 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dm8b\" (UniqueName: \"kubernetes.io/projected/2607e7bb-5f81-48cf-945a-6dee68b60040-kube-api-access-8dm8b\") pod \"auto-csr-approver-29556674-j6lkl\" (UID: \"2607e7bb-5f81-48cf-945a-6dee68b60040\") " pod="openshift-infra/auto-csr-approver-29556674-j6lkl" Mar 13 11:14:00 crc kubenswrapper[4632]: I0313 11:14:00.378197 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dm8b\" (UniqueName: \"kubernetes.io/projected/2607e7bb-5f81-48cf-945a-6dee68b60040-kube-api-access-8dm8b\") pod \"auto-csr-approver-29556674-j6lkl\" (UID: \"2607e7bb-5f81-48cf-945a-6dee68b60040\") " pod="openshift-infra/auto-csr-approver-29556674-j6lkl" Mar 13 11:14:00 crc kubenswrapper[4632]: I0313 11:14:00.467814 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556674-j6lkl" Mar 13 11:14:00 crc kubenswrapper[4632]: I0313 11:14:00.978198 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556674-j6lkl"] Mar 13 11:14:00 crc kubenswrapper[4632]: W0313 11:14:00.993126 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2607e7bb_5f81_48cf_945a_6dee68b60040.slice/crio-64817e1f3ef5c3a97ad60dd28f54df1da870ccd9ca8dc11cc50f926ec7e8a21a WatchSource:0}: Error finding container 64817e1f3ef5c3a97ad60dd28f54df1da870ccd9ca8dc11cc50f926ec7e8a21a: Status 404 returned error can't find the container with id 64817e1f3ef5c3a97ad60dd28f54df1da870ccd9ca8dc11cc50f926ec7e8a21a Mar 13 11:14:00 crc kubenswrapper[4632]: I0313 11:14:00.994751 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 11:14:01 crc kubenswrapper[4632]: I0313 11:14:01.270801 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556674-j6lkl" event={"ID":"2607e7bb-5f81-48cf-945a-6dee68b60040","Type":"ContainerStarted","Data":"64817e1f3ef5c3a97ad60dd28f54df1da870ccd9ca8dc11cc50f926ec7e8a21a"} Mar 13 11:14:02 crc kubenswrapper[4632]: I0313 11:14:02.280712 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556674-j6lkl" event={"ID":"2607e7bb-5f81-48cf-945a-6dee68b60040","Type":"ContainerStarted","Data":"54b5c748d72d1466e81773f53e5a61ff5e546e63d80706f8982cd195e971c601"} Mar 13 11:14:02 crc kubenswrapper[4632]: I0313 11:14:02.309493 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556674-j6lkl" podStartSLOduration=1.408531875 podStartE2EDuration="2.309470612s" podCreationTimestamp="2026-03-13 11:14:00 +0000 UTC" firstStartedPulling="2026-03-13 11:14:00.994491349 +0000 UTC m=+4215.017021482" lastFinishedPulling="2026-03-13 11:14:01.895430076 +0000 UTC m=+4215.917960219" observedRunningTime="2026-03-13 11:14:02.29923714 +0000 UTC m=+4216.321767303" watchObservedRunningTime="2026-03-13 11:14:02.309470612 +0000 UTC m=+4216.332000745" Mar 13 11:14:03 crc kubenswrapper[4632]: I0313 11:14:03.300111 4632 generic.go:334] "Generic (PLEG): container finished" podID="2607e7bb-5f81-48cf-945a-6dee68b60040" containerID="54b5c748d72d1466e81773f53e5a61ff5e546e63d80706f8982cd195e971c601" exitCode=0 Mar 13 11:14:03 crc kubenswrapper[4632]: I0313 11:14:03.300288 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556674-j6lkl" event={"ID":"2607e7bb-5f81-48cf-945a-6dee68b60040","Type":"ContainerDied","Data":"54b5c748d72d1466e81773f53e5a61ff5e546e63d80706f8982cd195e971c601"} Mar 13 11:14:04 crc kubenswrapper[4632]: I0313 11:14:04.694312 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556674-j6lkl" Mar 13 11:14:04 crc kubenswrapper[4632]: I0313 11:14:04.729206 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dm8b\" (UniqueName: \"kubernetes.io/projected/2607e7bb-5f81-48cf-945a-6dee68b60040-kube-api-access-8dm8b\") pod \"2607e7bb-5f81-48cf-945a-6dee68b60040\" (UID: \"2607e7bb-5f81-48cf-945a-6dee68b60040\") " Mar 13 11:14:04 crc kubenswrapper[4632]: I0313 11:14:04.735190 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2607e7bb-5f81-48cf-945a-6dee68b60040-kube-api-access-8dm8b" (OuterVolumeSpecName: "kube-api-access-8dm8b") pod "2607e7bb-5f81-48cf-945a-6dee68b60040" (UID: "2607e7bb-5f81-48cf-945a-6dee68b60040"). InnerVolumeSpecName "kube-api-access-8dm8b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:14:04 crc kubenswrapper[4632]: I0313 11:14:04.831538 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dm8b\" (UniqueName: \"kubernetes.io/projected/2607e7bb-5f81-48cf-945a-6dee68b60040-kube-api-access-8dm8b\") on node \"crc\" DevicePath \"\"" Mar 13 11:14:05 crc kubenswrapper[4632]: I0313 11:14:05.320031 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556674-j6lkl" event={"ID":"2607e7bb-5f81-48cf-945a-6dee68b60040","Type":"ContainerDied","Data":"64817e1f3ef5c3a97ad60dd28f54df1da870ccd9ca8dc11cc50f926ec7e8a21a"} Mar 13 11:14:05 crc kubenswrapper[4632]: I0313 11:14:05.320084 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64817e1f3ef5c3a97ad60dd28f54df1da870ccd9ca8dc11cc50f926ec7e8a21a" Mar 13 11:14:05 crc kubenswrapper[4632]: I0313 11:14:05.320153 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556674-j6lkl" Mar 13 11:14:05 crc kubenswrapper[4632]: I0313 11:14:05.419066 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556668-bzw4c"] Mar 13 11:14:05 crc kubenswrapper[4632]: I0313 11:14:05.439381 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556668-bzw4c"] Mar 13 11:14:06 crc kubenswrapper[4632]: I0313 11:14:06.062692 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d08697f-ce87-4d33-823a-9bf5d2d0d801" path="/var/lib/kubelet/pods/4d08697f-ce87-4d33-823a-9bf5d2d0d801/volumes" Mar 13 11:14:10 crc kubenswrapper[4632]: I0313 11:14:10.460894 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:14:10 crc kubenswrapper[4632]: I0313 11:14:10.461381 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:14:10 crc kubenswrapper[4632]: I0313 11:14:10.461441 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 11:14:10 crc kubenswrapper[4632]: I0313 11:14:10.462265 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"908510815c251e300c3555d9b7458818dfba317c0679f487df71c717e5c832f9"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 11:14:10 crc kubenswrapper[4632]: I0313 11:14:10.462329 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://908510815c251e300c3555d9b7458818dfba317c0679f487df71c717e5c832f9" gracePeriod=600 Mar 13 11:14:11 crc kubenswrapper[4632]: I0313 11:14:11.381216 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="908510815c251e300c3555d9b7458818dfba317c0679f487df71c717e5c832f9" exitCode=0 Mar 13 11:14:11 crc kubenswrapper[4632]: I0313 11:14:11.381377 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"908510815c251e300c3555d9b7458818dfba317c0679f487df71c717e5c832f9"} Mar 13 11:14:11 crc kubenswrapper[4632]: I0313 11:14:11.381649 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a"} Mar 13 11:14:11 crc kubenswrapper[4632]: I0313 11:14:11.381673 4632 scope.go:117] "RemoveContainer" containerID="8a4edabf825a9fe82d4af3664ef24831617306062afee5f4494f20215f557582" Mar 13 11:14:16 crc kubenswrapper[4632]: I0313 11:14:16.327415 4632 scope.go:117] "RemoveContainer" containerID="cdc6d57bdf1672f0d8614c97a39233ec4a346ac1edaab1837f60116110b310ef" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.171126 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8"] Mar 13 11:15:00 crc kubenswrapper[4632]: E0313 11:15:00.172046 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2607e7bb-5f81-48cf-945a-6dee68b60040" containerName="oc" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.172062 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2607e7bb-5f81-48cf-945a-6dee68b60040" containerName="oc" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.172338 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="2607e7bb-5f81-48cf-945a-6dee68b60040" containerName="oc" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.173162 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.175474 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.175985 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.205249 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8"] Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.292814 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9481bb7b-d00a-4ee1-b711-7b90d97907c1-config-volume\") pod \"collect-profiles-29556675-n64w8\" (UID: \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.292879 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dhhd\" (UniqueName: \"kubernetes.io/projected/9481bb7b-d00a-4ee1-b711-7b90d97907c1-kube-api-access-6dhhd\") pod \"collect-profiles-29556675-n64w8\" (UID: \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.293271 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9481bb7b-d00a-4ee1-b711-7b90d97907c1-secret-volume\") pod \"collect-profiles-29556675-n64w8\" (UID: \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.395337 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9481bb7b-d00a-4ee1-b711-7b90d97907c1-secret-volume\") pod \"collect-profiles-29556675-n64w8\" (UID: \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.395482 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9481bb7b-d00a-4ee1-b711-7b90d97907c1-config-volume\") pod \"collect-profiles-29556675-n64w8\" (UID: \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.395532 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dhhd\" (UniqueName: \"kubernetes.io/projected/9481bb7b-d00a-4ee1-b711-7b90d97907c1-kube-api-access-6dhhd\") pod \"collect-profiles-29556675-n64w8\" (UID: \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.396715 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9481bb7b-d00a-4ee1-b711-7b90d97907c1-config-volume\") pod \"collect-profiles-29556675-n64w8\" (UID: \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.402990 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9481bb7b-d00a-4ee1-b711-7b90d97907c1-secret-volume\") pod \"collect-profiles-29556675-n64w8\" (UID: \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.416469 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dhhd\" (UniqueName: \"kubernetes.io/projected/9481bb7b-d00a-4ee1-b711-7b90d97907c1-kube-api-access-6dhhd\") pod \"collect-profiles-29556675-n64w8\" (UID: \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" Mar 13 11:15:00 crc kubenswrapper[4632]: I0313 11:15:00.492017 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" Mar 13 11:15:01 crc kubenswrapper[4632]: I0313 11:15:00.999850 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8"] Mar 13 11:15:01 crc kubenswrapper[4632]: I0313 11:15:01.124106 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" event={"ID":"9481bb7b-d00a-4ee1-b711-7b90d97907c1","Type":"ContainerStarted","Data":"bb607a5fcbbc599f445e9f5e35641c5968cf1efefa8cb36e1b8cbe3421faa466"} Mar 13 11:15:02 crc kubenswrapper[4632]: I0313 11:15:02.136801 4632 generic.go:334] "Generic (PLEG): container finished" podID="9481bb7b-d00a-4ee1-b711-7b90d97907c1" containerID="95360e41112a84b3ea4b235c3e7fd03654d6110fccc446520298cff419091ae2" exitCode=0 Mar 13 11:15:02 crc kubenswrapper[4632]: I0313 11:15:02.136879 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" event={"ID":"9481bb7b-d00a-4ee1-b711-7b90d97907c1","Type":"ContainerDied","Data":"95360e41112a84b3ea4b235c3e7fd03654d6110fccc446520298cff419091ae2"} Mar 13 11:15:03 crc kubenswrapper[4632]: I0313 11:15:03.512605 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" Mar 13 11:15:03 crc kubenswrapper[4632]: I0313 11:15:03.558819 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9481bb7b-d00a-4ee1-b711-7b90d97907c1-secret-volume\") pod \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\" (UID: \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\") " Mar 13 11:15:03 crc kubenswrapper[4632]: I0313 11:15:03.559135 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dhhd\" (UniqueName: \"kubernetes.io/projected/9481bb7b-d00a-4ee1-b711-7b90d97907c1-kube-api-access-6dhhd\") pod \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\" (UID: \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\") " Mar 13 11:15:03 crc kubenswrapper[4632]: I0313 11:15:03.559244 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9481bb7b-d00a-4ee1-b711-7b90d97907c1-config-volume\") pod \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\" (UID: \"9481bb7b-d00a-4ee1-b711-7b90d97907c1\") " Mar 13 11:15:03 crc kubenswrapper[4632]: I0313 11:15:03.560419 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9481bb7b-d00a-4ee1-b711-7b90d97907c1-config-volume" (OuterVolumeSpecName: "config-volume") pod "9481bb7b-d00a-4ee1-b711-7b90d97907c1" (UID: "9481bb7b-d00a-4ee1-b711-7b90d97907c1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 11:15:03 crc kubenswrapper[4632]: I0313 11:15:03.568145 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9481bb7b-d00a-4ee1-b711-7b90d97907c1-kube-api-access-6dhhd" (OuterVolumeSpecName: "kube-api-access-6dhhd") pod "9481bb7b-d00a-4ee1-b711-7b90d97907c1" (UID: "9481bb7b-d00a-4ee1-b711-7b90d97907c1"). InnerVolumeSpecName "kube-api-access-6dhhd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:15:03 crc kubenswrapper[4632]: I0313 11:15:03.570993 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9481bb7b-d00a-4ee1-b711-7b90d97907c1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9481bb7b-d00a-4ee1-b711-7b90d97907c1" (UID: "9481bb7b-d00a-4ee1-b711-7b90d97907c1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:15:03 crc kubenswrapper[4632]: I0313 11:15:03.662337 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dhhd\" (UniqueName: \"kubernetes.io/projected/9481bb7b-d00a-4ee1-b711-7b90d97907c1-kube-api-access-6dhhd\") on node \"crc\" DevicePath \"\"" Mar 13 11:15:03 crc kubenswrapper[4632]: I0313 11:15:03.662633 4632 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9481bb7b-d00a-4ee1-b711-7b90d97907c1-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 11:15:03 crc kubenswrapper[4632]: I0313 11:15:03.662745 4632 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9481bb7b-d00a-4ee1-b711-7b90d97907c1-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 11:15:04 crc kubenswrapper[4632]: I0313 11:15:04.181497 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" event={"ID":"9481bb7b-d00a-4ee1-b711-7b90d97907c1","Type":"ContainerDied","Data":"bb607a5fcbbc599f445e9f5e35641c5968cf1efefa8cb36e1b8cbe3421faa466"} Mar 13 11:15:04 crc kubenswrapper[4632]: I0313 11:15:04.181897 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb607a5fcbbc599f445e9f5e35641c5968cf1efefa8cb36e1b8cbe3421faa466" Mar 13 11:15:04 crc kubenswrapper[4632]: I0313 11:15:04.181607 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8" Mar 13 11:15:04 crc kubenswrapper[4632]: I0313 11:15:04.618291 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7"] Mar 13 11:15:04 crc kubenswrapper[4632]: I0313 11:15:04.629686 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556630-kpbz7"] Mar 13 11:15:06 crc kubenswrapper[4632]: I0313 11:15:06.056326 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e912f5a7-eb85-4d19-9703-6cd7ff46c810" path="/var/lib/kubelet/pods/e912f5a7-eb85-4d19-9703-6cd7ff46c810/volumes" Mar 13 11:15:16 crc kubenswrapper[4632]: I0313 11:15:16.875549 4632 scope.go:117] "RemoveContainer" containerID="020084ff22e9c174abe1865969844a8ece77dd4c848ac5f03af6af51bccf8643" Mar 13 11:15:28 crc kubenswrapper[4632]: I0313 11:15:28.907947 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fjcb8"] Mar 13 11:15:28 crc kubenswrapper[4632]: E0313 11:15:28.908719 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9481bb7b-d00a-4ee1-b711-7b90d97907c1" containerName="collect-profiles" Mar 13 11:15:28 crc kubenswrapper[4632]: I0313 11:15:28.908730 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="9481bb7b-d00a-4ee1-b711-7b90d97907c1" containerName="collect-profiles" Mar 13 11:15:28 crc kubenswrapper[4632]: I0313 11:15:28.908994 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="9481bb7b-d00a-4ee1-b711-7b90d97907c1" containerName="collect-profiles" Mar 13 11:15:28 crc kubenswrapper[4632]: I0313 11:15:28.910261 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:28 crc kubenswrapper[4632]: I0313 11:15:28.932394 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjcb8"] Mar 13 11:15:28 crc kubenswrapper[4632]: I0313 11:15:28.998462 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b80498f-6567-4384-8312-3eec23afb96f-utilities\") pod \"redhat-marketplace-fjcb8\" (UID: \"6b80498f-6567-4384-8312-3eec23afb96f\") " pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:28 crc kubenswrapper[4632]: I0313 11:15:28.998531 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4v48\" (UniqueName: \"kubernetes.io/projected/6b80498f-6567-4384-8312-3eec23afb96f-kube-api-access-q4v48\") pod \"redhat-marketplace-fjcb8\" (UID: \"6b80498f-6567-4384-8312-3eec23afb96f\") " pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:28 crc kubenswrapper[4632]: I0313 11:15:28.998551 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b80498f-6567-4384-8312-3eec23afb96f-catalog-content\") pod \"redhat-marketplace-fjcb8\" (UID: \"6b80498f-6567-4384-8312-3eec23afb96f\") " pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:29 crc kubenswrapper[4632]: I0313 11:15:29.100524 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b80498f-6567-4384-8312-3eec23afb96f-utilities\") pod \"redhat-marketplace-fjcb8\" (UID: \"6b80498f-6567-4384-8312-3eec23afb96f\") " pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:29 crc kubenswrapper[4632]: I0313 11:15:29.100998 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4v48\" (UniqueName: \"kubernetes.io/projected/6b80498f-6567-4384-8312-3eec23afb96f-kube-api-access-q4v48\") pod \"redhat-marketplace-fjcb8\" (UID: \"6b80498f-6567-4384-8312-3eec23afb96f\") " pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:29 crc kubenswrapper[4632]: I0313 11:15:29.101030 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b80498f-6567-4384-8312-3eec23afb96f-catalog-content\") pod \"redhat-marketplace-fjcb8\" (UID: \"6b80498f-6567-4384-8312-3eec23afb96f\") " pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:29 crc kubenswrapper[4632]: I0313 11:15:29.101405 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b80498f-6567-4384-8312-3eec23afb96f-utilities\") pod \"redhat-marketplace-fjcb8\" (UID: \"6b80498f-6567-4384-8312-3eec23afb96f\") " pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:29 crc kubenswrapper[4632]: I0313 11:15:29.101667 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b80498f-6567-4384-8312-3eec23afb96f-catalog-content\") pod \"redhat-marketplace-fjcb8\" (UID: \"6b80498f-6567-4384-8312-3eec23afb96f\") " pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:29 crc kubenswrapper[4632]: I0313 11:15:29.125463 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4v48\" (UniqueName: \"kubernetes.io/projected/6b80498f-6567-4384-8312-3eec23afb96f-kube-api-access-q4v48\") pod \"redhat-marketplace-fjcb8\" (UID: \"6b80498f-6567-4384-8312-3eec23afb96f\") " pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:29 crc kubenswrapper[4632]: I0313 11:15:29.227435 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:29 crc kubenswrapper[4632]: I0313 11:15:29.730328 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjcb8"] Mar 13 11:15:30 crc kubenswrapper[4632]: I0313 11:15:30.467626 4632 generic.go:334] "Generic (PLEG): container finished" podID="6b80498f-6567-4384-8312-3eec23afb96f" containerID="e4c2cf5dbf98373fe62cd97ac31b0a9d4d0cfebb456d4b4a21937b1cdd3aa690" exitCode=0 Mar 13 11:15:30 crc kubenswrapper[4632]: I0313 11:15:30.467738 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjcb8" event={"ID":"6b80498f-6567-4384-8312-3eec23afb96f","Type":"ContainerDied","Data":"e4c2cf5dbf98373fe62cd97ac31b0a9d4d0cfebb456d4b4a21937b1cdd3aa690"} Mar 13 11:15:30 crc kubenswrapper[4632]: I0313 11:15:30.468072 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjcb8" event={"ID":"6b80498f-6567-4384-8312-3eec23afb96f","Type":"ContainerStarted","Data":"d0a5afdd3947f0bc783bf017602e397a2d152c6ec52ad447d9ddbb2f47b390d3"} Mar 13 11:15:31 crc kubenswrapper[4632]: I0313 11:15:31.494340 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjcb8" event={"ID":"6b80498f-6567-4384-8312-3eec23afb96f","Type":"ContainerStarted","Data":"b5f67696ef58459c1f88ef05dc6509bd9cb7f8b1c9e52e4e87ca38283e2b4063"} Mar 13 11:15:32 crc kubenswrapper[4632]: I0313 11:15:32.504415 4632 generic.go:334] "Generic (PLEG): container finished" podID="6b80498f-6567-4384-8312-3eec23afb96f" containerID="b5f67696ef58459c1f88ef05dc6509bd9cb7f8b1c9e52e4e87ca38283e2b4063" exitCode=0 Mar 13 11:15:32 crc kubenswrapper[4632]: I0313 11:15:32.504636 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjcb8" event={"ID":"6b80498f-6567-4384-8312-3eec23afb96f","Type":"ContainerDied","Data":"b5f67696ef58459c1f88ef05dc6509bd9cb7f8b1c9e52e4e87ca38283e2b4063"} Mar 13 11:15:33 crc kubenswrapper[4632]: I0313 11:15:33.515777 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjcb8" event={"ID":"6b80498f-6567-4384-8312-3eec23afb96f","Type":"ContainerStarted","Data":"cae11f82bcd03a8e15a974a2d286987da044e8a582b503a6c527777766946f93"} Mar 13 11:15:33 crc kubenswrapper[4632]: I0313 11:15:33.541754 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fjcb8" podStartSLOduration=3.001395962 podStartE2EDuration="5.541726037s" podCreationTimestamp="2026-03-13 11:15:28 +0000 UTC" firstStartedPulling="2026-03-13 11:15:30.470684129 +0000 UTC m=+4304.493214262" lastFinishedPulling="2026-03-13 11:15:33.011014204 +0000 UTC m=+4307.033544337" observedRunningTime="2026-03-13 11:15:33.536003915 +0000 UTC m=+4307.558534048" watchObservedRunningTime="2026-03-13 11:15:33.541726037 +0000 UTC m=+4307.564256170" Mar 13 11:15:39 crc kubenswrapper[4632]: I0313 11:15:39.228689 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:39 crc kubenswrapper[4632]: I0313 11:15:39.229184 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:39 crc kubenswrapper[4632]: I0313 11:15:39.367784 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:40 crc kubenswrapper[4632]: I0313 11:15:40.304072 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:40 crc kubenswrapper[4632]: I0313 11:15:40.360319 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjcb8"] Mar 13 11:15:41 crc kubenswrapper[4632]: I0313 11:15:41.593733 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fjcb8" podUID="6b80498f-6567-4384-8312-3eec23afb96f" containerName="registry-server" containerID="cri-o://cae11f82bcd03a8e15a974a2d286987da044e8a582b503a6c527777766946f93" gracePeriod=2 Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.469378 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.604285 4632 generic.go:334] "Generic (PLEG): container finished" podID="6b80498f-6567-4384-8312-3eec23afb96f" containerID="cae11f82bcd03a8e15a974a2d286987da044e8a582b503a6c527777766946f93" exitCode=0 Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.604352 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fjcb8" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.604376 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjcb8" event={"ID":"6b80498f-6567-4384-8312-3eec23afb96f","Type":"ContainerDied","Data":"cae11f82bcd03a8e15a974a2d286987da044e8a582b503a6c527777766946f93"} Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.604775 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fjcb8" event={"ID":"6b80498f-6567-4384-8312-3eec23afb96f","Type":"ContainerDied","Data":"d0a5afdd3947f0bc783bf017602e397a2d152c6ec52ad447d9ddbb2f47b390d3"} Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.604805 4632 scope.go:117] "RemoveContainer" containerID="cae11f82bcd03a8e15a974a2d286987da044e8a582b503a6c527777766946f93" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.625380 4632 scope.go:117] "RemoveContainer" containerID="b5f67696ef58459c1f88ef05dc6509bd9cb7f8b1c9e52e4e87ca38283e2b4063" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.646651 4632 scope.go:117] "RemoveContainer" containerID="e4c2cf5dbf98373fe62cd97ac31b0a9d4d0cfebb456d4b4a21937b1cdd3aa690" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.669191 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b80498f-6567-4384-8312-3eec23afb96f-catalog-content\") pod \"6b80498f-6567-4384-8312-3eec23afb96f\" (UID: \"6b80498f-6567-4384-8312-3eec23afb96f\") " Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.669268 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q4v48\" (UniqueName: \"kubernetes.io/projected/6b80498f-6567-4384-8312-3eec23afb96f-kube-api-access-q4v48\") pod \"6b80498f-6567-4384-8312-3eec23afb96f\" (UID: \"6b80498f-6567-4384-8312-3eec23afb96f\") " Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.669305 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b80498f-6567-4384-8312-3eec23afb96f-utilities\") pod \"6b80498f-6567-4384-8312-3eec23afb96f\" (UID: \"6b80498f-6567-4384-8312-3eec23afb96f\") " Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.671207 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b80498f-6567-4384-8312-3eec23afb96f-utilities" (OuterVolumeSpecName: "utilities") pod "6b80498f-6567-4384-8312-3eec23afb96f" (UID: "6b80498f-6567-4384-8312-3eec23afb96f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.678072 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b80498f-6567-4384-8312-3eec23afb96f-kube-api-access-q4v48" (OuterVolumeSpecName: "kube-api-access-q4v48") pod "6b80498f-6567-4384-8312-3eec23afb96f" (UID: "6b80498f-6567-4384-8312-3eec23afb96f"). InnerVolumeSpecName "kube-api-access-q4v48". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.704981 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b80498f-6567-4384-8312-3eec23afb96f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b80498f-6567-4384-8312-3eec23afb96f" (UID: "6b80498f-6567-4384-8312-3eec23afb96f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.745662 4632 scope.go:117] "RemoveContainer" containerID="cae11f82bcd03a8e15a974a2d286987da044e8a582b503a6c527777766946f93" Mar 13 11:15:42 crc kubenswrapper[4632]: E0313 11:15:42.747020 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cae11f82bcd03a8e15a974a2d286987da044e8a582b503a6c527777766946f93\": container with ID starting with cae11f82bcd03a8e15a974a2d286987da044e8a582b503a6c527777766946f93 not found: ID does not exist" containerID="cae11f82bcd03a8e15a974a2d286987da044e8a582b503a6c527777766946f93" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.747082 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cae11f82bcd03a8e15a974a2d286987da044e8a582b503a6c527777766946f93"} err="failed to get container status \"cae11f82bcd03a8e15a974a2d286987da044e8a582b503a6c527777766946f93\": rpc error: code = NotFound desc = could not find container \"cae11f82bcd03a8e15a974a2d286987da044e8a582b503a6c527777766946f93\": container with ID starting with cae11f82bcd03a8e15a974a2d286987da044e8a582b503a6c527777766946f93 not found: ID does not exist" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.747105 4632 scope.go:117] "RemoveContainer" containerID="b5f67696ef58459c1f88ef05dc6509bd9cb7f8b1c9e52e4e87ca38283e2b4063" Mar 13 11:15:42 crc kubenswrapper[4632]: E0313 11:15:42.747749 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5f67696ef58459c1f88ef05dc6509bd9cb7f8b1c9e52e4e87ca38283e2b4063\": container with ID starting with b5f67696ef58459c1f88ef05dc6509bd9cb7f8b1c9e52e4e87ca38283e2b4063 not found: ID does not exist" containerID="b5f67696ef58459c1f88ef05dc6509bd9cb7f8b1c9e52e4e87ca38283e2b4063" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.747807 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5f67696ef58459c1f88ef05dc6509bd9cb7f8b1c9e52e4e87ca38283e2b4063"} err="failed to get container status \"b5f67696ef58459c1f88ef05dc6509bd9cb7f8b1c9e52e4e87ca38283e2b4063\": rpc error: code = NotFound desc = could not find container \"b5f67696ef58459c1f88ef05dc6509bd9cb7f8b1c9e52e4e87ca38283e2b4063\": container with ID starting with b5f67696ef58459c1f88ef05dc6509bd9cb7f8b1c9e52e4e87ca38283e2b4063 not found: ID does not exist" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.747839 4632 scope.go:117] "RemoveContainer" containerID="e4c2cf5dbf98373fe62cd97ac31b0a9d4d0cfebb456d4b4a21937b1cdd3aa690" Mar 13 11:15:42 crc kubenswrapper[4632]: E0313 11:15:42.748539 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4c2cf5dbf98373fe62cd97ac31b0a9d4d0cfebb456d4b4a21937b1cdd3aa690\": container with ID starting with e4c2cf5dbf98373fe62cd97ac31b0a9d4d0cfebb456d4b4a21937b1cdd3aa690 not found: ID does not exist" containerID="e4c2cf5dbf98373fe62cd97ac31b0a9d4d0cfebb456d4b4a21937b1cdd3aa690" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.748592 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4c2cf5dbf98373fe62cd97ac31b0a9d4d0cfebb456d4b4a21937b1cdd3aa690"} err="failed to get container status \"e4c2cf5dbf98373fe62cd97ac31b0a9d4d0cfebb456d4b4a21937b1cdd3aa690\": rpc error: code = NotFound desc = could not find container \"e4c2cf5dbf98373fe62cd97ac31b0a9d4d0cfebb456d4b4a21937b1cdd3aa690\": container with ID starting with e4c2cf5dbf98373fe62cd97ac31b0a9d4d0cfebb456d4b4a21937b1cdd3aa690 not found: ID does not exist" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.771175 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b80498f-6567-4384-8312-3eec23afb96f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.771209 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q4v48\" (UniqueName: \"kubernetes.io/projected/6b80498f-6567-4384-8312-3eec23afb96f-kube-api-access-q4v48\") on node \"crc\" DevicePath \"\"" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.771221 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b80498f-6567-4384-8312-3eec23afb96f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.939500 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjcb8"] Mar 13 11:15:42 crc kubenswrapper[4632]: I0313 11:15:42.947400 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fjcb8"] Mar 13 11:15:44 crc kubenswrapper[4632]: I0313 11:15:44.060785 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b80498f-6567-4384-8312-3eec23afb96f" path="/var/lib/kubelet/pods/6b80498f-6567-4384-8312-3eec23afb96f/volumes" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.035565 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mkx4r"] Mar 13 11:15:51 crc kubenswrapper[4632]: E0313 11:15:51.038583 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b80498f-6567-4384-8312-3eec23afb96f" containerName="extract-content" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.038684 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b80498f-6567-4384-8312-3eec23afb96f" containerName="extract-content" Mar 13 11:15:51 crc kubenswrapper[4632]: E0313 11:15:51.038776 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b80498f-6567-4384-8312-3eec23afb96f" containerName="extract-utilities" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.038869 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b80498f-6567-4384-8312-3eec23afb96f" containerName="extract-utilities" Mar 13 11:15:51 crc kubenswrapper[4632]: E0313 11:15:51.039064 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b80498f-6567-4384-8312-3eec23afb96f" containerName="registry-server" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.039080 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b80498f-6567-4384-8312-3eec23afb96f" containerName="registry-server" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.039564 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b80498f-6567-4384-8312-3eec23afb96f" containerName="registry-server" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.041423 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.080806 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mkx4r"] Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.154817 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43baafa6-f011-4a57-a843-9ff515c2d27c-catalog-content\") pod \"certified-operators-mkx4r\" (UID: \"43baafa6-f011-4a57-a843-9ff515c2d27c\") " pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.155226 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gj6h\" (UniqueName: \"kubernetes.io/projected/43baafa6-f011-4a57-a843-9ff515c2d27c-kube-api-access-5gj6h\") pod \"certified-operators-mkx4r\" (UID: \"43baafa6-f011-4a57-a843-9ff515c2d27c\") " pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.155317 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43baafa6-f011-4a57-a843-9ff515c2d27c-utilities\") pod \"certified-operators-mkx4r\" (UID: \"43baafa6-f011-4a57-a843-9ff515c2d27c\") " pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.257494 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43baafa6-f011-4a57-a843-9ff515c2d27c-catalog-content\") pod \"certified-operators-mkx4r\" (UID: \"43baafa6-f011-4a57-a843-9ff515c2d27c\") " pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.257562 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gj6h\" (UniqueName: \"kubernetes.io/projected/43baafa6-f011-4a57-a843-9ff515c2d27c-kube-api-access-5gj6h\") pod \"certified-operators-mkx4r\" (UID: \"43baafa6-f011-4a57-a843-9ff515c2d27c\") " pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.257605 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43baafa6-f011-4a57-a843-9ff515c2d27c-utilities\") pod \"certified-operators-mkx4r\" (UID: \"43baafa6-f011-4a57-a843-9ff515c2d27c\") " pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.258145 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43baafa6-f011-4a57-a843-9ff515c2d27c-utilities\") pod \"certified-operators-mkx4r\" (UID: \"43baafa6-f011-4a57-a843-9ff515c2d27c\") " pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.258204 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43baafa6-f011-4a57-a843-9ff515c2d27c-catalog-content\") pod \"certified-operators-mkx4r\" (UID: \"43baafa6-f011-4a57-a843-9ff515c2d27c\") " pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.278329 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gj6h\" (UniqueName: \"kubernetes.io/projected/43baafa6-f011-4a57-a843-9ff515c2d27c-kube-api-access-5gj6h\") pod \"certified-operators-mkx4r\" (UID: \"43baafa6-f011-4a57-a843-9ff515c2d27c\") " pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.375595 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:15:51 crc kubenswrapper[4632]: I0313 11:15:51.944068 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mkx4r"] Mar 13 11:15:52 crc kubenswrapper[4632]: I0313 11:15:52.717575 4632 generic.go:334] "Generic (PLEG): container finished" podID="43baafa6-f011-4a57-a843-9ff515c2d27c" containerID="0a8143ecbbf23166aa2635bf482dd844c2de6e91412ef57334b5958839b8171f" exitCode=0 Mar 13 11:15:52 crc kubenswrapper[4632]: I0313 11:15:52.717632 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkx4r" event={"ID":"43baafa6-f011-4a57-a843-9ff515c2d27c","Type":"ContainerDied","Data":"0a8143ecbbf23166aa2635bf482dd844c2de6e91412ef57334b5958839b8171f"} Mar 13 11:15:52 crc kubenswrapper[4632]: I0313 11:15:52.718601 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkx4r" event={"ID":"43baafa6-f011-4a57-a843-9ff515c2d27c","Type":"ContainerStarted","Data":"2d6c4a914964242a2013f8a205bc41526ff8fae16e636ee7d8cd1ac5a0677874"} Mar 13 11:15:54 crc kubenswrapper[4632]: I0313 11:15:54.739525 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkx4r" event={"ID":"43baafa6-f011-4a57-a843-9ff515c2d27c","Type":"ContainerStarted","Data":"eeb3c7988fe06aac21b41b63a9bb8fc4ac2e8f50f48b82caf52b00ff344abc9c"} Mar 13 11:15:56 crc kubenswrapper[4632]: I0313 11:15:56.761779 4632 generic.go:334] "Generic (PLEG): container finished" podID="43baafa6-f011-4a57-a843-9ff515c2d27c" containerID="eeb3c7988fe06aac21b41b63a9bb8fc4ac2e8f50f48b82caf52b00ff344abc9c" exitCode=0 Mar 13 11:15:56 crc kubenswrapper[4632]: I0313 11:15:56.761888 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkx4r" event={"ID":"43baafa6-f011-4a57-a843-9ff515c2d27c","Type":"ContainerDied","Data":"eeb3c7988fe06aac21b41b63a9bb8fc4ac2e8f50f48b82caf52b00ff344abc9c"} Mar 13 11:15:57 crc kubenswrapper[4632]: I0313 11:15:57.775592 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkx4r" event={"ID":"43baafa6-f011-4a57-a843-9ff515c2d27c","Type":"ContainerStarted","Data":"5068b71db4a67bf4fba229474a62974c266237a6aa5af74f3af86b3451d483dc"} Mar 13 11:15:57 crc kubenswrapper[4632]: I0313 11:15:57.798344 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mkx4r" podStartSLOduration=2.334089454 podStartE2EDuration="6.798327702s" podCreationTimestamp="2026-03-13 11:15:51 +0000 UTC" firstStartedPulling="2026-03-13 11:15:52.72134613 +0000 UTC m=+4326.743876273" lastFinishedPulling="2026-03-13 11:15:57.185584388 +0000 UTC m=+4331.208114521" observedRunningTime="2026-03-13 11:15:57.793474231 +0000 UTC m=+4331.816004374" watchObservedRunningTime="2026-03-13 11:15:57.798327702 +0000 UTC m=+4331.820857835" Mar 13 11:16:00 crc kubenswrapper[4632]: I0313 11:16:00.142032 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556676-8qz49"] Mar 13 11:16:00 crc kubenswrapper[4632]: I0313 11:16:00.144405 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556676-8qz49" Mar 13 11:16:00 crc kubenswrapper[4632]: I0313 11:16:00.147013 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:16:00 crc kubenswrapper[4632]: I0313 11:16:00.147720 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:16:00 crc kubenswrapper[4632]: I0313 11:16:00.148228 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:16:00 crc kubenswrapper[4632]: I0313 11:16:00.156549 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556676-8qz49"] Mar 13 11:16:00 crc kubenswrapper[4632]: I0313 11:16:00.295139 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xznxs\" (UniqueName: \"kubernetes.io/projected/f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911-kube-api-access-xznxs\") pod \"auto-csr-approver-29556676-8qz49\" (UID: \"f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911\") " pod="openshift-infra/auto-csr-approver-29556676-8qz49" Mar 13 11:16:00 crc kubenswrapper[4632]: I0313 11:16:00.396504 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xznxs\" (UniqueName: \"kubernetes.io/projected/f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911-kube-api-access-xznxs\") pod \"auto-csr-approver-29556676-8qz49\" (UID: \"f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911\") " pod="openshift-infra/auto-csr-approver-29556676-8qz49" Mar 13 11:16:00 crc kubenswrapper[4632]: I0313 11:16:00.418274 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xznxs\" (UniqueName: \"kubernetes.io/projected/f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911-kube-api-access-xznxs\") pod \"auto-csr-approver-29556676-8qz49\" (UID: \"f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911\") " pod="openshift-infra/auto-csr-approver-29556676-8qz49" Mar 13 11:16:00 crc kubenswrapper[4632]: I0313 11:16:00.464705 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556676-8qz49" Mar 13 11:16:01 crc kubenswrapper[4632]: I0313 11:16:01.036198 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556676-8qz49"] Mar 13 11:16:01 crc kubenswrapper[4632]: W0313 11:16:01.038089 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf97cdaa5_f90d_4cd0_9b62_7bb0c3b41911.slice/crio-4d598b65671a365f65cf6a6e23cbf30862f2afbcc820e1c2e386503e7ec46797 WatchSource:0}: Error finding container 4d598b65671a365f65cf6a6e23cbf30862f2afbcc820e1c2e386503e7ec46797: Status 404 returned error can't find the container with id 4d598b65671a365f65cf6a6e23cbf30862f2afbcc820e1c2e386503e7ec46797 Mar 13 11:16:01 crc kubenswrapper[4632]: I0313 11:16:01.376604 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:16:01 crc kubenswrapper[4632]: I0313 11:16:01.376709 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:16:01 crc kubenswrapper[4632]: I0313 11:16:01.820074 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556676-8qz49" event={"ID":"f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911","Type":"ContainerStarted","Data":"4d598b65671a365f65cf6a6e23cbf30862f2afbcc820e1c2e386503e7ec46797"} Mar 13 11:16:02 crc kubenswrapper[4632]: I0313 11:16:02.449072 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-mkx4r" podUID="43baafa6-f011-4a57-a843-9ff515c2d27c" containerName="registry-server" probeResult="failure" output=< Mar 13 11:16:02 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:16:02 crc kubenswrapper[4632]: > Mar 13 11:16:02 crc kubenswrapper[4632]: I0313 11:16:02.832561 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556676-8qz49" event={"ID":"f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911","Type":"ContainerStarted","Data":"1913657413fffb5d7b6f0c5a32e25db59682a49e52d39a0adc600808b4a0def3"} Mar 13 11:16:02 crc kubenswrapper[4632]: I0313 11:16:02.857335 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556676-8qz49" podStartSLOduration=2.003172086 podStartE2EDuration="2.857307759s" podCreationTimestamp="2026-03-13 11:16:00 +0000 UTC" firstStartedPulling="2026-03-13 11:16:01.039824941 +0000 UTC m=+4335.062355074" lastFinishedPulling="2026-03-13 11:16:01.893960614 +0000 UTC m=+4335.916490747" observedRunningTime="2026-03-13 11:16:02.849153578 +0000 UTC m=+4336.871683721" watchObservedRunningTime="2026-03-13 11:16:02.857307759 +0000 UTC m=+4336.879837902" Mar 13 11:16:03 crc kubenswrapper[4632]: I0313 11:16:03.746965 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-92vd7"] Mar 13 11:16:03 crc kubenswrapper[4632]: I0313 11:16:03.752532 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:03 crc kubenswrapper[4632]: I0313 11:16:03.781207 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-92vd7"] Mar 13 11:16:03 crc kubenswrapper[4632]: I0313 11:16:03.845583 4632 generic.go:334] "Generic (PLEG): container finished" podID="f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911" containerID="1913657413fffb5d7b6f0c5a32e25db59682a49e52d39a0adc600808b4a0def3" exitCode=0 Mar 13 11:16:03 crc kubenswrapper[4632]: I0313 11:16:03.845638 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556676-8qz49" event={"ID":"f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911","Type":"ContainerDied","Data":"1913657413fffb5d7b6f0c5a32e25db59682a49e52d39a0adc600808b4a0def3"} Mar 13 11:16:03 crc kubenswrapper[4632]: I0313 11:16:03.873632 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-utilities\") pod \"community-operators-92vd7\" (UID: \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\") " pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:03 crc kubenswrapper[4632]: I0313 11:16:03.873716 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-catalog-content\") pod \"community-operators-92vd7\" (UID: \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\") " pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:03 crc kubenswrapper[4632]: I0313 11:16:03.873763 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbxvh\" (UniqueName: \"kubernetes.io/projected/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-kube-api-access-fbxvh\") pod \"community-operators-92vd7\" (UID: \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\") " pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:03 crc kubenswrapper[4632]: I0313 11:16:03.974968 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-catalog-content\") pod \"community-operators-92vd7\" (UID: \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\") " pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:03 crc kubenswrapper[4632]: I0313 11:16:03.975044 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbxvh\" (UniqueName: \"kubernetes.io/projected/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-kube-api-access-fbxvh\") pod \"community-operators-92vd7\" (UID: \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\") " pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:03 crc kubenswrapper[4632]: I0313 11:16:03.975142 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-utilities\") pod \"community-operators-92vd7\" (UID: \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\") " pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:03 crc kubenswrapper[4632]: I0313 11:16:03.975498 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-catalog-content\") pod \"community-operators-92vd7\" (UID: \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\") " pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:03 crc kubenswrapper[4632]: I0313 11:16:03.975572 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-utilities\") pod \"community-operators-92vd7\" (UID: \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\") " pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:03 crc kubenswrapper[4632]: I0313 11:16:03.997813 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbxvh\" (UniqueName: \"kubernetes.io/projected/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-kube-api-access-fbxvh\") pod \"community-operators-92vd7\" (UID: \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\") " pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:04 crc kubenswrapper[4632]: I0313 11:16:04.081149 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:04 crc kubenswrapper[4632]: I0313 11:16:04.614338 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-92vd7"] Mar 13 11:16:04 crc kubenswrapper[4632]: I0313 11:16:04.854290 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92vd7" event={"ID":"c7c05421-6c9d-4a5b-b77e-278ff5610dbb","Type":"ContainerStarted","Data":"2292a7de70f1a922a86c5aac2e3477eea585fc2a4541f7fdab875f746948a354"} Mar 13 11:16:04 crc kubenswrapper[4632]: I0313 11:16:04.854578 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92vd7" event={"ID":"c7c05421-6c9d-4a5b-b77e-278ff5610dbb","Type":"ContainerStarted","Data":"b93e7fe938f5ed8d7dd19d8e5f95e035e0510382f4433390808953a862971f44"} Mar 13 11:16:05 crc kubenswrapper[4632]: I0313 11:16:05.246751 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556676-8qz49" Mar 13 11:16:05 crc kubenswrapper[4632]: I0313 11:16:05.300129 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xznxs\" (UniqueName: \"kubernetes.io/projected/f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911-kube-api-access-xznxs\") pod \"f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911\" (UID: \"f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911\") " Mar 13 11:16:05 crc kubenswrapper[4632]: I0313 11:16:05.308764 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911-kube-api-access-xznxs" (OuterVolumeSpecName: "kube-api-access-xznxs") pod "f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911" (UID: "f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911"). InnerVolumeSpecName "kube-api-access-xznxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:16:05 crc kubenswrapper[4632]: I0313 11:16:05.402703 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xznxs\" (UniqueName: \"kubernetes.io/projected/f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911-kube-api-access-xznxs\") on node \"crc\" DevicePath \"\"" Mar 13 11:16:05 crc kubenswrapper[4632]: I0313 11:16:05.866794 4632 generic.go:334] "Generic (PLEG): container finished" podID="c7c05421-6c9d-4a5b-b77e-278ff5610dbb" containerID="2292a7de70f1a922a86c5aac2e3477eea585fc2a4541f7fdab875f746948a354" exitCode=0 Mar 13 11:16:05 crc kubenswrapper[4632]: I0313 11:16:05.868288 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92vd7" event={"ID":"c7c05421-6c9d-4a5b-b77e-278ff5610dbb","Type":"ContainerDied","Data":"2292a7de70f1a922a86c5aac2e3477eea585fc2a4541f7fdab875f746948a354"} Mar 13 11:16:05 crc kubenswrapper[4632]: I0313 11:16:05.872453 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556676-8qz49" event={"ID":"f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911","Type":"ContainerDied","Data":"4d598b65671a365f65cf6a6e23cbf30862f2afbcc820e1c2e386503e7ec46797"} Mar 13 11:16:05 crc kubenswrapper[4632]: I0313 11:16:05.872513 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4d598b65671a365f65cf6a6e23cbf30862f2afbcc820e1c2e386503e7ec46797" Mar 13 11:16:05 crc kubenswrapper[4632]: I0313 11:16:05.872617 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556676-8qz49" Mar 13 11:16:05 crc kubenswrapper[4632]: I0313 11:16:05.948229 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556670-fqvf6"] Mar 13 11:16:05 crc kubenswrapper[4632]: I0313 11:16:05.957555 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556670-fqvf6"] Mar 13 11:16:06 crc kubenswrapper[4632]: I0313 11:16:06.061354 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="565a5983-3957-42c2-b7d4-47d26e00aec8" path="/var/lib/kubelet/pods/565a5983-3957-42c2-b7d4-47d26e00aec8/volumes" Mar 13 11:16:06 crc kubenswrapper[4632]: I0313 11:16:06.883204 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92vd7" event={"ID":"c7c05421-6c9d-4a5b-b77e-278ff5610dbb","Type":"ContainerStarted","Data":"863e77cb00d2ce8145dde64bc045043b8b85335c157a29f91e5e18fe7f269448"} Mar 13 11:16:08 crc kubenswrapper[4632]: I0313 11:16:08.904496 4632 generic.go:334] "Generic (PLEG): container finished" podID="c7c05421-6c9d-4a5b-b77e-278ff5610dbb" containerID="863e77cb00d2ce8145dde64bc045043b8b85335c157a29f91e5e18fe7f269448" exitCode=0 Mar 13 11:16:08 crc kubenswrapper[4632]: I0313 11:16:08.904589 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92vd7" event={"ID":"c7c05421-6c9d-4a5b-b77e-278ff5610dbb","Type":"ContainerDied","Data":"863e77cb00d2ce8145dde64bc045043b8b85335c157a29f91e5e18fe7f269448"} Mar 13 11:16:09 crc kubenswrapper[4632]: I0313 11:16:09.916231 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92vd7" event={"ID":"c7c05421-6c9d-4a5b-b77e-278ff5610dbb","Type":"ContainerStarted","Data":"4716208f34ca87ca08b3d99ecbc5950cf92ba98ce62d017396c4d72582dffff5"} Mar 13 11:16:09 crc kubenswrapper[4632]: I0313 11:16:09.940092 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-92vd7" podStartSLOduration=3.262218417 podStartE2EDuration="6.940073442s" podCreationTimestamp="2026-03-13 11:16:03 +0000 UTC" firstStartedPulling="2026-03-13 11:16:05.868924052 +0000 UTC m=+4339.891454205" lastFinishedPulling="2026-03-13 11:16:09.546779077 +0000 UTC m=+4343.569309230" observedRunningTime="2026-03-13 11:16:09.938754169 +0000 UTC m=+4343.961284312" watchObservedRunningTime="2026-03-13 11:16:09.940073442 +0000 UTC m=+4343.962603575" Mar 13 11:16:10 crc kubenswrapper[4632]: I0313 11:16:10.461163 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:16:10 crc kubenswrapper[4632]: I0313 11:16:10.461216 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:16:12 crc kubenswrapper[4632]: I0313 11:16:12.455830 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-mkx4r" podUID="43baafa6-f011-4a57-a843-9ff515c2d27c" containerName="registry-server" probeResult="failure" output=< Mar 13 11:16:12 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:16:12 crc kubenswrapper[4632]: > Mar 13 11:16:14 crc kubenswrapper[4632]: I0313 11:16:14.081867 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:14 crc kubenswrapper[4632]: I0313 11:16:14.083094 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:15 crc kubenswrapper[4632]: I0313 11:16:15.144266 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-92vd7" podUID="c7c05421-6c9d-4a5b-b77e-278ff5610dbb" containerName="registry-server" probeResult="failure" output=< Mar 13 11:16:15 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:16:15 crc kubenswrapper[4632]: > Mar 13 11:16:16 crc kubenswrapper[4632]: I0313 11:16:16.988738 4632 scope.go:117] "RemoveContainer" containerID="c0218119e7ac388fadab5a0e90f8eec2d8161ed6d34eaec2b46cb615f7e41508" Mar 13 11:16:21 crc kubenswrapper[4632]: I0313 11:16:21.606664 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:16:21 crc kubenswrapper[4632]: I0313 11:16:21.664970 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:16:22 crc kubenswrapper[4632]: I0313 11:16:22.241770 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mkx4r"] Mar 13 11:16:23 crc kubenswrapper[4632]: I0313 11:16:23.041891 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mkx4r" podUID="43baafa6-f011-4a57-a843-9ff515c2d27c" containerName="registry-server" containerID="cri-o://5068b71db4a67bf4fba229474a62974c266237a6aa5af74f3af86b3451d483dc" gracePeriod=2 Mar 13 11:16:23 crc kubenswrapper[4632]: I0313 11:16:23.780494 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:16:23 crc kubenswrapper[4632]: I0313 11:16:23.799235 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43baafa6-f011-4a57-a843-9ff515c2d27c-utilities\") pod \"43baafa6-f011-4a57-a843-9ff515c2d27c\" (UID: \"43baafa6-f011-4a57-a843-9ff515c2d27c\") " Mar 13 11:16:23 crc kubenswrapper[4632]: I0313 11:16:23.799288 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43baafa6-f011-4a57-a843-9ff515c2d27c-catalog-content\") pod \"43baafa6-f011-4a57-a843-9ff515c2d27c\" (UID: \"43baafa6-f011-4a57-a843-9ff515c2d27c\") " Mar 13 11:16:23 crc kubenswrapper[4632]: I0313 11:16:23.799496 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gj6h\" (UniqueName: \"kubernetes.io/projected/43baafa6-f011-4a57-a843-9ff515c2d27c-kube-api-access-5gj6h\") pod \"43baafa6-f011-4a57-a843-9ff515c2d27c\" (UID: \"43baafa6-f011-4a57-a843-9ff515c2d27c\") " Mar 13 11:16:23 crc kubenswrapper[4632]: I0313 11:16:23.801083 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43baafa6-f011-4a57-a843-9ff515c2d27c-utilities" (OuterVolumeSpecName: "utilities") pod "43baafa6-f011-4a57-a843-9ff515c2d27c" (UID: "43baafa6-f011-4a57-a843-9ff515c2d27c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:16:23 crc kubenswrapper[4632]: I0313 11:16:23.836266 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43baafa6-f011-4a57-a843-9ff515c2d27c-kube-api-access-5gj6h" (OuterVolumeSpecName: "kube-api-access-5gj6h") pod "43baafa6-f011-4a57-a843-9ff515c2d27c" (UID: "43baafa6-f011-4a57-a843-9ff515c2d27c"). InnerVolumeSpecName "kube-api-access-5gj6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:16:23 crc kubenswrapper[4632]: I0313 11:16:23.903918 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/43baafa6-f011-4a57-a843-9ff515c2d27c-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:16:23 crc kubenswrapper[4632]: I0313 11:16:23.903997 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gj6h\" (UniqueName: \"kubernetes.io/projected/43baafa6-f011-4a57-a843-9ff515c2d27c-kube-api-access-5gj6h\") on node \"crc\" DevicePath \"\"" Mar 13 11:16:23 crc kubenswrapper[4632]: I0313 11:16:23.916770 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/43baafa6-f011-4a57-a843-9ff515c2d27c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "43baafa6-f011-4a57-a843-9ff515c2d27c" (UID: "43baafa6-f011-4a57-a843-9ff515c2d27c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.006252 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/43baafa6-f011-4a57-a843-9ff515c2d27c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.051509 4632 generic.go:334] "Generic (PLEG): container finished" podID="43baafa6-f011-4a57-a843-9ff515c2d27c" containerID="5068b71db4a67bf4fba229474a62974c266237a6aa5af74f3af86b3451d483dc" exitCode=0 Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.051600 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mkx4r" Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.054280 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkx4r" event={"ID":"43baafa6-f011-4a57-a843-9ff515c2d27c","Type":"ContainerDied","Data":"5068b71db4a67bf4fba229474a62974c266237a6aa5af74f3af86b3451d483dc"} Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.054328 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mkx4r" event={"ID":"43baafa6-f011-4a57-a843-9ff515c2d27c","Type":"ContainerDied","Data":"2d6c4a914964242a2013f8a205bc41526ff8fae16e636ee7d8cd1ac5a0677874"} Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.054349 4632 scope.go:117] "RemoveContainer" containerID="5068b71db4a67bf4fba229474a62974c266237a6aa5af74f3af86b3451d483dc" Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.106701 4632 scope.go:117] "RemoveContainer" containerID="eeb3c7988fe06aac21b41b63a9bb8fc4ac2e8f50f48b82caf52b00ff344abc9c" Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.110590 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mkx4r"] Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.116650 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mkx4r"] Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.142605 4632 scope.go:117] "RemoveContainer" containerID="0a8143ecbbf23166aa2635bf482dd844c2de6e91412ef57334b5958839b8171f" Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.148060 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.207352 4632 scope.go:117] "RemoveContainer" containerID="5068b71db4a67bf4fba229474a62974c266237a6aa5af74f3af86b3451d483dc" Mar 13 11:16:24 crc kubenswrapper[4632]: E0313 11:16:24.208178 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5068b71db4a67bf4fba229474a62974c266237a6aa5af74f3af86b3451d483dc\": container with ID starting with 5068b71db4a67bf4fba229474a62974c266237a6aa5af74f3af86b3451d483dc not found: ID does not exist" containerID="5068b71db4a67bf4fba229474a62974c266237a6aa5af74f3af86b3451d483dc" Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.208363 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5068b71db4a67bf4fba229474a62974c266237a6aa5af74f3af86b3451d483dc"} err="failed to get container status \"5068b71db4a67bf4fba229474a62974c266237a6aa5af74f3af86b3451d483dc\": rpc error: code = NotFound desc = could not find container \"5068b71db4a67bf4fba229474a62974c266237a6aa5af74f3af86b3451d483dc\": container with ID starting with 5068b71db4a67bf4fba229474a62974c266237a6aa5af74f3af86b3451d483dc not found: ID does not exist" Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.208444 4632 scope.go:117] "RemoveContainer" containerID="eeb3c7988fe06aac21b41b63a9bb8fc4ac2e8f50f48b82caf52b00ff344abc9c" Mar 13 11:16:24 crc kubenswrapper[4632]: E0313 11:16:24.210229 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeb3c7988fe06aac21b41b63a9bb8fc4ac2e8f50f48b82caf52b00ff344abc9c\": container with ID starting with eeb3c7988fe06aac21b41b63a9bb8fc4ac2e8f50f48b82caf52b00ff344abc9c not found: ID does not exist" containerID="eeb3c7988fe06aac21b41b63a9bb8fc4ac2e8f50f48b82caf52b00ff344abc9c" Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.210346 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeb3c7988fe06aac21b41b63a9bb8fc4ac2e8f50f48b82caf52b00ff344abc9c"} err="failed to get container status \"eeb3c7988fe06aac21b41b63a9bb8fc4ac2e8f50f48b82caf52b00ff344abc9c\": rpc error: code = NotFound desc = could not find container \"eeb3c7988fe06aac21b41b63a9bb8fc4ac2e8f50f48b82caf52b00ff344abc9c\": container with ID starting with eeb3c7988fe06aac21b41b63a9bb8fc4ac2e8f50f48b82caf52b00ff344abc9c not found: ID does not exist" Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.210443 4632 scope.go:117] "RemoveContainer" containerID="0a8143ecbbf23166aa2635bf482dd844c2de6e91412ef57334b5958839b8171f" Mar 13 11:16:24 crc kubenswrapper[4632]: E0313 11:16:24.210781 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a8143ecbbf23166aa2635bf482dd844c2de6e91412ef57334b5958839b8171f\": container with ID starting with 0a8143ecbbf23166aa2635bf482dd844c2de6e91412ef57334b5958839b8171f not found: ID does not exist" containerID="0a8143ecbbf23166aa2635bf482dd844c2de6e91412ef57334b5958839b8171f" Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.210869 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a8143ecbbf23166aa2635bf482dd844c2de6e91412ef57334b5958839b8171f"} err="failed to get container status \"0a8143ecbbf23166aa2635bf482dd844c2de6e91412ef57334b5958839b8171f\": rpc error: code = NotFound desc = could not find container \"0a8143ecbbf23166aa2635bf482dd844c2de6e91412ef57334b5958839b8171f\": container with ID starting with 0a8143ecbbf23166aa2635bf482dd844c2de6e91412ef57334b5958839b8171f not found: ID does not exist" Mar 13 11:16:24 crc kubenswrapper[4632]: I0313 11:16:24.241766 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:26 crc kubenswrapper[4632]: I0313 11:16:26.075046 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43baafa6-f011-4a57-a843-9ff515c2d27c" path="/var/lib/kubelet/pods/43baafa6-f011-4a57-a843-9ff515c2d27c/volumes" Mar 13 11:16:26 crc kubenswrapper[4632]: I0313 11:16:26.444005 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-92vd7"] Mar 13 11:16:26 crc kubenswrapper[4632]: I0313 11:16:26.444501 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-92vd7" podUID="c7c05421-6c9d-4a5b-b77e-278ff5610dbb" containerName="registry-server" containerID="cri-o://4716208f34ca87ca08b3d99ecbc5950cf92ba98ce62d017396c4d72582dffff5" gracePeriod=2 Mar 13 11:16:26 crc kubenswrapper[4632]: I0313 11:16:26.996098 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.093821 4632 generic.go:334] "Generic (PLEG): container finished" podID="c7c05421-6c9d-4a5b-b77e-278ff5610dbb" containerID="4716208f34ca87ca08b3d99ecbc5950cf92ba98ce62d017396c4d72582dffff5" exitCode=0 Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.093872 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-92vd7" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.093872 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92vd7" event={"ID":"c7c05421-6c9d-4a5b-b77e-278ff5610dbb","Type":"ContainerDied","Data":"4716208f34ca87ca08b3d99ecbc5950cf92ba98ce62d017396c4d72582dffff5"} Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.094047 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-92vd7" event={"ID":"c7c05421-6c9d-4a5b-b77e-278ff5610dbb","Type":"ContainerDied","Data":"b93e7fe938f5ed8d7dd19d8e5f95e035e0510382f4433390808953a862971f44"} Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.094075 4632 scope.go:117] "RemoveContainer" containerID="4716208f34ca87ca08b3d99ecbc5950cf92ba98ce62d017396c4d72582dffff5" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.117309 4632 scope.go:117] "RemoveContainer" containerID="863e77cb00d2ce8145dde64bc045043b8b85335c157a29f91e5e18fe7f269448" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.134629 4632 scope.go:117] "RemoveContainer" containerID="2292a7de70f1a922a86c5aac2e3477eea585fc2a4541f7fdab875f746948a354" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.180260 4632 scope.go:117] "RemoveContainer" containerID="4716208f34ca87ca08b3d99ecbc5950cf92ba98ce62d017396c4d72582dffff5" Mar 13 11:16:27 crc kubenswrapper[4632]: E0313 11:16:27.180782 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4716208f34ca87ca08b3d99ecbc5950cf92ba98ce62d017396c4d72582dffff5\": container with ID starting with 4716208f34ca87ca08b3d99ecbc5950cf92ba98ce62d017396c4d72582dffff5 not found: ID does not exist" containerID="4716208f34ca87ca08b3d99ecbc5950cf92ba98ce62d017396c4d72582dffff5" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.180832 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4716208f34ca87ca08b3d99ecbc5950cf92ba98ce62d017396c4d72582dffff5"} err="failed to get container status \"4716208f34ca87ca08b3d99ecbc5950cf92ba98ce62d017396c4d72582dffff5\": rpc error: code = NotFound desc = could not find container \"4716208f34ca87ca08b3d99ecbc5950cf92ba98ce62d017396c4d72582dffff5\": container with ID starting with 4716208f34ca87ca08b3d99ecbc5950cf92ba98ce62d017396c4d72582dffff5 not found: ID does not exist" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.180863 4632 scope.go:117] "RemoveContainer" containerID="863e77cb00d2ce8145dde64bc045043b8b85335c157a29f91e5e18fe7f269448" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.181655 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbxvh\" (UniqueName: \"kubernetes.io/projected/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-kube-api-access-fbxvh\") pod \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\" (UID: \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\") " Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.181875 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-utilities\") pod \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\" (UID: \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\") " Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.182011 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-catalog-content\") pod \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\" (UID: \"c7c05421-6c9d-4a5b-b77e-278ff5610dbb\") " Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.183697 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-utilities" (OuterVolumeSpecName: "utilities") pod "c7c05421-6c9d-4a5b-b77e-278ff5610dbb" (UID: "c7c05421-6c9d-4a5b-b77e-278ff5610dbb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:16:27 crc kubenswrapper[4632]: E0313 11:16:27.184001 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"863e77cb00d2ce8145dde64bc045043b8b85335c157a29f91e5e18fe7f269448\": container with ID starting with 863e77cb00d2ce8145dde64bc045043b8b85335c157a29f91e5e18fe7f269448 not found: ID does not exist" containerID="863e77cb00d2ce8145dde64bc045043b8b85335c157a29f91e5e18fe7f269448" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.184033 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"863e77cb00d2ce8145dde64bc045043b8b85335c157a29f91e5e18fe7f269448"} err="failed to get container status \"863e77cb00d2ce8145dde64bc045043b8b85335c157a29f91e5e18fe7f269448\": rpc error: code = NotFound desc = could not find container \"863e77cb00d2ce8145dde64bc045043b8b85335c157a29f91e5e18fe7f269448\": container with ID starting with 863e77cb00d2ce8145dde64bc045043b8b85335c157a29f91e5e18fe7f269448 not found: ID does not exist" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.184060 4632 scope.go:117] "RemoveContainer" containerID="2292a7de70f1a922a86c5aac2e3477eea585fc2a4541f7fdab875f746948a354" Mar 13 11:16:27 crc kubenswrapper[4632]: E0313 11:16:27.184573 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2292a7de70f1a922a86c5aac2e3477eea585fc2a4541f7fdab875f746948a354\": container with ID starting with 2292a7de70f1a922a86c5aac2e3477eea585fc2a4541f7fdab875f746948a354 not found: ID does not exist" containerID="2292a7de70f1a922a86c5aac2e3477eea585fc2a4541f7fdab875f746948a354" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.184654 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2292a7de70f1a922a86c5aac2e3477eea585fc2a4541f7fdab875f746948a354"} err="failed to get container status \"2292a7de70f1a922a86c5aac2e3477eea585fc2a4541f7fdab875f746948a354\": rpc error: code = NotFound desc = could not find container \"2292a7de70f1a922a86c5aac2e3477eea585fc2a4541f7fdab875f746948a354\": container with ID starting with 2292a7de70f1a922a86c5aac2e3477eea585fc2a4541f7fdab875f746948a354 not found: ID does not exist" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.191291 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-kube-api-access-fbxvh" (OuterVolumeSpecName: "kube-api-access-fbxvh") pod "c7c05421-6c9d-4a5b-b77e-278ff5610dbb" (UID: "c7c05421-6c9d-4a5b-b77e-278ff5610dbb"). InnerVolumeSpecName "kube-api-access-fbxvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.237182 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c7c05421-6c9d-4a5b-b77e-278ff5610dbb" (UID: "c7c05421-6c9d-4a5b-b77e-278ff5610dbb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.285101 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbxvh\" (UniqueName: \"kubernetes.io/projected/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-kube-api-access-fbxvh\") on node \"crc\" DevicePath \"\"" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.285173 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.285186 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7c05421-6c9d-4a5b-b77e-278ff5610dbb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.450439 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-92vd7"] Mar 13 11:16:27 crc kubenswrapper[4632]: I0313 11:16:27.466870 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-92vd7"] Mar 13 11:16:28 crc kubenswrapper[4632]: I0313 11:16:28.058380 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c7c05421-6c9d-4a5b-b77e-278ff5610dbb" path="/var/lib/kubelet/pods/c7c05421-6c9d-4a5b-b77e-278ff5610dbb/volumes" Mar 13 11:16:40 crc kubenswrapper[4632]: I0313 11:16:40.460858 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:16:40 crc kubenswrapper[4632]: I0313 11:16:40.462112 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:17:10 crc kubenswrapper[4632]: I0313 11:17:10.460620 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:17:10 crc kubenswrapper[4632]: I0313 11:17:10.461199 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:17:10 crc kubenswrapper[4632]: I0313 11:17:10.461247 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 11:17:10 crc kubenswrapper[4632]: I0313 11:17:10.462212 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 11:17:10 crc kubenswrapper[4632]: I0313 11:17:10.462272 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" gracePeriod=600 Mar 13 11:17:10 crc kubenswrapper[4632]: I0313 11:17:10.678270 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" exitCode=0 Mar 13 11:17:10 crc kubenswrapper[4632]: I0313 11:17:10.678330 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a"} Mar 13 11:17:10 crc kubenswrapper[4632]: I0313 11:17:10.678403 4632 scope.go:117] "RemoveContainer" containerID="908510815c251e300c3555d9b7458818dfba317c0679f487df71c717e5c832f9" Mar 13 11:17:11 crc kubenswrapper[4632]: E0313 11:17:11.273265 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:17:11 crc kubenswrapper[4632]: I0313 11:17:11.701645 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:17:11 crc kubenswrapper[4632]: E0313 11:17:11.701892 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:17:25 crc kubenswrapper[4632]: I0313 11:17:25.044688 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:17:25 crc kubenswrapper[4632]: E0313 11:17:25.045541 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:17:36 crc kubenswrapper[4632]: I0313 11:17:36.044471 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:17:36 crc kubenswrapper[4632]: E0313 11:17:36.045252 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.719889 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gcqr2"] Mar 13 11:17:38 crc kubenswrapper[4632]: E0313 11:17:38.720719 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7c05421-6c9d-4a5b-b77e-278ff5610dbb" containerName="extract-utilities" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.720736 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7c05421-6c9d-4a5b-b77e-278ff5610dbb" containerName="extract-utilities" Mar 13 11:17:38 crc kubenswrapper[4632]: E0313 11:17:38.720755 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43baafa6-f011-4a57-a843-9ff515c2d27c" containerName="extract-content" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.720763 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="43baafa6-f011-4a57-a843-9ff515c2d27c" containerName="extract-content" Mar 13 11:17:38 crc kubenswrapper[4632]: E0313 11:17:38.720790 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911" containerName="oc" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.720799 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911" containerName="oc" Mar 13 11:17:38 crc kubenswrapper[4632]: E0313 11:17:38.720817 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7c05421-6c9d-4a5b-b77e-278ff5610dbb" containerName="extract-content" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.720825 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7c05421-6c9d-4a5b-b77e-278ff5610dbb" containerName="extract-content" Mar 13 11:17:38 crc kubenswrapper[4632]: E0313 11:17:38.720838 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43baafa6-f011-4a57-a843-9ff515c2d27c" containerName="registry-server" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.720846 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="43baafa6-f011-4a57-a843-9ff515c2d27c" containerName="registry-server" Mar 13 11:17:38 crc kubenswrapper[4632]: E0313 11:17:38.720862 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7c05421-6c9d-4a5b-b77e-278ff5610dbb" containerName="registry-server" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.720870 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7c05421-6c9d-4a5b-b77e-278ff5610dbb" containerName="registry-server" Mar 13 11:17:38 crc kubenswrapper[4632]: E0313 11:17:38.720885 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43baafa6-f011-4a57-a843-9ff515c2d27c" containerName="extract-utilities" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.720893 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="43baafa6-f011-4a57-a843-9ff515c2d27c" containerName="extract-utilities" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.721210 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7c05421-6c9d-4a5b-b77e-278ff5610dbb" containerName="registry-server" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.721232 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911" containerName="oc" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.721258 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="43baafa6-f011-4a57-a843-9ff515c2d27c" containerName="registry-server" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.723579 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.761430 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gcqr2"] Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.902073 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27c3c199-cc34-438a-ac59-4555ee7c5a1d-catalog-content\") pod \"redhat-operators-gcqr2\" (UID: \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\") " pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.902135 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x27fk\" (UniqueName: \"kubernetes.io/projected/27c3c199-cc34-438a-ac59-4555ee7c5a1d-kube-api-access-x27fk\") pod \"redhat-operators-gcqr2\" (UID: \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\") " pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:17:38 crc kubenswrapper[4632]: I0313 11:17:38.902162 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27c3c199-cc34-438a-ac59-4555ee7c5a1d-utilities\") pod \"redhat-operators-gcqr2\" (UID: \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\") " pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:17:39 crc kubenswrapper[4632]: I0313 11:17:39.004379 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x27fk\" (UniqueName: \"kubernetes.io/projected/27c3c199-cc34-438a-ac59-4555ee7c5a1d-kube-api-access-x27fk\") pod \"redhat-operators-gcqr2\" (UID: \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\") " pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:17:39 crc kubenswrapper[4632]: I0313 11:17:39.004448 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27c3c199-cc34-438a-ac59-4555ee7c5a1d-utilities\") pod \"redhat-operators-gcqr2\" (UID: \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\") " pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:17:39 crc kubenswrapper[4632]: I0313 11:17:39.004607 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27c3c199-cc34-438a-ac59-4555ee7c5a1d-catalog-content\") pod \"redhat-operators-gcqr2\" (UID: \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\") " pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:17:39 crc kubenswrapper[4632]: I0313 11:17:39.004982 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27c3c199-cc34-438a-ac59-4555ee7c5a1d-utilities\") pod \"redhat-operators-gcqr2\" (UID: \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\") " pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:17:39 crc kubenswrapper[4632]: I0313 11:17:39.005049 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27c3c199-cc34-438a-ac59-4555ee7c5a1d-catalog-content\") pod \"redhat-operators-gcqr2\" (UID: \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\") " pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:17:39 crc kubenswrapper[4632]: I0313 11:17:39.027877 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x27fk\" (UniqueName: \"kubernetes.io/projected/27c3c199-cc34-438a-ac59-4555ee7c5a1d-kube-api-access-x27fk\") pod \"redhat-operators-gcqr2\" (UID: \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\") " pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:17:39 crc kubenswrapper[4632]: I0313 11:17:39.055411 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:17:39 crc kubenswrapper[4632]: I0313 11:17:39.707862 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gcqr2"] Mar 13 11:17:40 crc kubenswrapper[4632]: I0313 11:17:40.065657 4632 generic.go:334] "Generic (PLEG): container finished" podID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerID="8fc46f8a96264e119d9565fb1071d38b9eca7de4da0002eb50dc0db108fea360" exitCode=0 Mar 13 11:17:40 crc kubenswrapper[4632]: I0313 11:17:40.065987 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gcqr2" event={"ID":"27c3c199-cc34-438a-ac59-4555ee7c5a1d","Type":"ContainerDied","Data":"8fc46f8a96264e119d9565fb1071d38b9eca7de4da0002eb50dc0db108fea360"} Mar 13 11:17:40 crc kubenswrapper[4632]: I0313 11:17:40.066047 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gcqr2" event={"ID":"27c3c199-cc34-438a-ac59-4555ee7c5a1d","Type":"ContainerStarted","Data":"507ec2a71889920853d24af4b68882a43ebcc431eef7fce060f7fd2fe2c9a9ee"} Mar 13 11:17:42 crc kubenswrapper[4632]: I0313 11:17:42.123464 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gcqr2" event={"ID":"27c3c199-cc34-438a-ac59-4555ee7c5a1d","Type":"ContainerStarted","Data":"f9fc93bd191406b3e302e5390ddb104f245922f0c823cd3aa32e935c54c7057f"} Mar 13 11:17:47 crc kubenswrapper[4632]: I0313 11:17:47.045101 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:17:47 crc kubenswrapper[4632]: E0313 11:17:47.046157 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:17:47 crc kubenswrapper[4632]: I0313 11:17:47.180408 4632 generic.go:334] "Generic (PLEG): container finished" podID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerID="f9fc93bd191406b3e302e5390ddb104f245922f0c823cd3aa32e935c54c7057f" exitCode=0 Mar 13 11:17:47 crc kubenswrapper[4632]: I0313 11:17:47.180490 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gcqr2" event={"ID":"27c3c199-cc34-438a-ac59-4555ee7c5a1d","Type":"ContainerDied","Data":"f9fc93bd191406b3e302e5390ddb104f245922f0c823cd3aa32e935c54c7057f"} Mar 13 11:17:48 crc kubenswrapper[4632]: I0313 11:17:48.193261 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gcqr2" event={"ID":"27c3c199-cc34-438a-ac59-4555ee7c5a1d","Type":"ContainerStarted","Data":"098f5138f07f45ec1bf0ad95c6647f82c5303be4fdf81f24941c5f5d332834fa"} Mar 13 11:17:48 crc kubenswrapper[4632]: I0313 11:17:48.267497 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gcqr2" podStartSLOduration=2.674656681 podStartE2EDuration="10.267479921s" podCreationTimestamp="2026-03-13 11:17:38 +0000 UTC" firstStartedPulling="2026-03-13 11:17:40.071423264 +0000 UTC m=+4434.093953397" lastFinishedPulling="2026-03-13 11:17:47.664246504 +0000 UTC m=+4441.686776637" observedRunningTime="2026-03-13 11:17:48.265262626 +0000 UTC m=+4442.287792779" watchObservedRunningTime="2026-03-13 11:17:48.267479921 +0000 UTC m=+4442.290010054" Mar 13 11:17:49 crc kubenswrapper[4632]: I0313 11:17:49.056533 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:17:49 crc kubenswrapper[4632]: I0313 11:17:49.057116 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:17:50 crc kubenswrapper[4632]: I0313 11:17:50.129759 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gcqr2" podUID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerName="registry-server" probeResult="failure" output=< Mar 13 11:17:50 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:17:50 crc kubenswrapper[4632]: > Mar 13 11:18:00 crc kubenswrapper[4632]: I0313 11:18:00.114076 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gcqr2" podUID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerName="registry-server" probeResult="failure" output=< Mar 13 11:18:00 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:18:00 crc kubenswrapper[4632]: > Mar 13 11:18:00 crc kubenswrapper[4632]: I0313 11:18:00.164443 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556678-kxcj9"] Mar 13 11:18:00 crc kubenswrapper[4632]: I0313 11:18:00.166383 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556678-kxcj9" Mar 13 11:18:00 crc kubenswrapper[4632]: I0313 11:18:00.169665 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:18:00 crc kubenswrapper[4632]: I0313 11:18:00.169726 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:18:00 crc kubenswrapper[4632]: I0313 11:18:00.170117 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:18:00 crc kubenswrapper[4632]: I0313 11:18:00.185991 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556678-kxcj9"] Mar 13 11:18:00 crc kubenswrapper[4632]: I0313 11:18:00.225932 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd7bd\" (UniqueName: \"kubernetes.io/projected/feb87e78-c7fb-4997-869e-1e652f57ffe9-kube-api-access-zd7bd\") pod \"auto-csr-approver-29556678-kxcj9\" (UID: \"feb87e78-c7fb-4997-869e-1e652f57ffe9\") " pod="openshift-infra/auto-csr-approver-29556678-kxcj9" Mar 13 11:18:00 crc kubenswrapper[4632]: I0313 11:18:00.328325 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd7bd\" (UniqueName: \"kubernetes.io/projected/feb87e78-c7fb-4997-869e-1e652f57ffe9-kube-api-access-zd7bd\") pod \"auto-csr-approver-29556678-kxcj9\" (UID: \"feb87e78-c7fb-4997-869e-1e652f57ffe9\") " pod="openshift-infra/auto-csr-approver-29556678-kxcj9" Mar 13 11:18:00 crc kubenswrapper[4632]: I0313 11:18:00.349595 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd7bd\" (UniqueName: \"kubernetes.io/projected/feb87e78-c7fb-4997-869e-1e652f57ffe9-kube-api-access-zd7bd\") pod \"auto-csr-approver-29556678-kxcj9\" (UID: \"feb87e78-c7fb-4997-869e-1e652f57ffe9\") " pod="openshift-infra/auto-csr-approver-29556678-kxcj9" Mar 13 11:18:00 crc kubenswrapper[4632]: I0313 11:18:00.486028 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556678-kxcj9" Mar 13 11:18:01 crc kubenswrapper[4632]: I0313 11:18:01.054410 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:18:01 crc kubenswrapper[4632]: E0313 11:18:01.055032 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:18:01 crc kubenswrapper[4632]: I0313 11:18:01.433806 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556678-kxcj9"] Mar 13 11:18:02 crc kubenswrapper[4632]: I0313 11:18:02.340333 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556678-kxcj9" event={"ID":"feb87e78-c7fb-4997-869e-1e652f57ffe9","Type":"ContainerStarted","Data":"ca5c093b3e400764932de5c65149b0fc21dfec697c6ff3f3c089cc7c7701ab47"} Mar 13 11:18:03 crc kubenswrapper[4632]: I0313 11:18:03.352051 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556678-kxcj9" event={"ID":"feb87e78-c7fb-4997-869e-1e652f57ffe9","Type":"ContainerStarted","Data":"d5beb81ff52cba139334670e53dbdbf15336383f8949e10d3fb4d56429b9cd89"} Mar 13 11:18:05 crc kubenswrapper[4632]: I0313 11:18:05.381260 4632 generic.go:334] "Generic (PLEG): container finished" podID="feb87e78-c7fb-4997-869e-1e652f57ffe9" containerID="d5beb81ff52cba139334670e53dbdbf15336383f8949e10d3fb4d56429b9cd89" exitCode=0 Mar 13 11:18:05 crc kubenswrapper[4632]: I0313 11:18:05.381330 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556678-kxcj9" event={"ID":"feb87e78-c7fb-4997-869e-1e652f57ffe9","Type":"ContainerDied","Data":"d5beb81ff52cba139334670e53dbdbf15336383f8949e10d3fb4d56429b9cd89"} Mar 13 11:18:06 crc kubenswrapper[4632]: I0313 11:18:06.814234 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556678-kxcj9" Mar 13 11:18:06 crc kubenswrapper[4632]: I0313 11:18:06.904516 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd7bd\" (UniqueName: \"kubernetes.io/projected/feb87e78-c7fb-4997-869e-1e652f57ffe9-kube-api-access-zd7bd\") pod \"feb87e78-c7fb-4997-869e-1e652f57ffe9\" (UID: \"feb87e78-c7fb-4997-869e-1e652f57ffe9\") " Mar 13 11:18:06 crc kubenswrapper[4632]: I0313 11:18:06.911387 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/feb87e78-c7fb-4997-869e-1e652f57ffe9-kube-api-access-zd7bd" (OuterVolumeSpecName: "kube-api-access-zd7bd") pod "feb87e78-c7fb-4997-869e-1e652f57ffe9" (UID: "feb87e78-c7fb-4997-869e-1e652f57ffe9"). InnerVolumeSpecName "kube-api-access-zd7bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:18:07 crc kubenswrapper[4632]: I0313 11:18:07.007215 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd7bd\" (UniqueName: \"kubernetes.io/projected/feb87e78-c7fb-4997-869e-1e652f57ffe9-kube-api-access-zd7bd\") on node \"crc\" DevicePath \"\"" Mar 13 11:18:07 crc kubenswrapper[4632]: I0313 11:18:07.407672 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556678-kxcj9" event={"ID":"feb87e78-c7fb-4997-869e-1e652f57ffe9","Type":"ContainerDied","Data":"ca5c093b3e400764932de5c65149b0fc21dfec697c6ff3f3c089cc7c7701ab47"} Mar 13 11:18:07 crc kubenswrapper[4632]: I0313 11:18:07.407740 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ca5c093b3e400764932de5c65149b0fc21dfec697c6ff3f3c089cc7c7701ab47" Mar 13 11:18:07 crc kubenswrapper[4632]: I0313 11:18:07.407825 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556678-kxcj9" Mar 13 11:18:07 crc kubenswrapper[4632]: I0313 11:18:07.491457 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556672-sd9zg"] Mar 13 11:18:07 crc kubenswrapper[4632]: I0313 11:18:07.499373 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556672-sd9zg"] Mar 13 11:18:08 crc kubenswrapper[4632]: I0313 11:18:08.067499 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8" path="/var/lib/kubelet/pods/bc83f2bf-bfdf-4426-a5a6-d1299a5b4da8/volumes" Mar 13 11:18:10 crc kubenswrapper[4632]: I0313 11:18:10.105072 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gcqr2" podUID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerName="registry-server" probeResult="failure" output=< Mar 13 11:18:10 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:18:10 crc kubenswrapper[4632]: > Mar 13 11:18:12 crc kubenswrapper[4632]: I0313 11:18:12.044843 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:18:12 crc kubenswrapper[4632]: E0313 11:18:12.046054 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:18:17 crc kubenswrapper[4632]: I0313 11:18:17.365087 4632 scope.go:117] "RemoveContainer" containerID="5f687cba4c29fe06e8932802cf25f9e44ab270587540acfd69f26e45b584a52b" Mar 13 11:18:20 crc kubenswrapper[4632]: I0313 11:18:20.142155 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gcqr2" podUID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerName="registry-server" probeResult="failure" output=< Mar 13 11:18:20 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:18:20 crc kubenswrapper[4632]: > Mar 13 11:18:23 crc kubenswrapper[4632]: I0313 11:18:23.045817 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:18:23 crc kubenswrapper[4632]: E0313 11:18:23.046655 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:18:29 crc kubenswrapper[4632]: I0313 11:18:29.128259 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:18:29 crc kubenswrapper[4632]: I0313 11:18:29.199063 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:18:29 crc kubenswrapper[4632]: I0313 11:18:29.382138 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gcqr2"] Mar 13 11:18:30 crc kubenswrapper[4632]: I0313 11:18:30.706155 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gcqr2" podUID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerName="registry-server" containerID="cri-o://098f5138f07f45ec1bf0ad95c6647f82c5303be4fdf81f24941c5f5d332834fa" gracePeriod=2 Mar 13 11:18:31 crc kubenswrapper[4632]: I0313 11:18:31.729568 4632 generic.go:334] "Generic (PLEG): container finished" podID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerID="098f5138f07f45ec1bf0ad95c6647f82c5303be4fdf81f24941c5f5d332834fa" exitCode=0 Mar 13 11:18:31 crc kubenswrapper[4632]: I0313 11:18:31.730008 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gcqr2" event={"ID":"27c3c199-cc34-438a-ac59-4555ee7c5a1d","Type":"ContainerDied","Data":"098f5138f07f45ec1bf0ad95c6647f82c5303be4fdf81f24941c5f5d332834fa"} Mar 13 11:18:31 crc kubenswrapper[4632]: I0313 11:18:31.730050 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gcqr2" event={"ID":"27c3c199-cc34-438a-ac59-4555ee7c5a1d","Type":"ContainerDied","Data":"507ec2a71889920853d24af4b68882a43ebcc431eef7fce060f7fd2fe2c9a9ee"} Mar 13 11:18:31 crc kubenswrapper[4632]: I0313 11:18:31.730063 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="507ec2a71889920853d24af4b68882a43ebcc431eef7fce060f7fd2fe2c9a9ee" Mar 13 11:18:31 crc kubenswrapper[4632]: I0313 11:18:31.788570 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:18:31 crc kubenswrapper[4632]: I0313 11:18:31.858789 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27c3c199-cc34-438a-ac59-4555ee7c5a1d-catalog-content\") pod \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\" (UID: \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\") " Mar 13 11:18:31 crc kubenswrapper[4632]: I0313 11:18:31.858840 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27c3c199-cc34-438a-ac59-4555ee7c5a1d-utilities\") pod \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\" (UID: \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\") " Mar 13 11:18:31 crc kubenswrapper[4632]: I0313 11:18:31.858930 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x27fk\" (UniqueName: \"kubernetes.io/projected/27c3c199-cc34-438a-ac59-4555ee7c5a1d-kube-api-access-x27fk\") pod \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\" (UID: \"27c3c199-cc34-438a-ac59-4555ee7c5a1d\") " Mar 13 11:18:31 crc kubenswrapper[4632]: I0313 11:18:31.859632 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27c3c199-cc34-438a-ac59-4555ee7c5a1d-utilities" (OuterVolumeSpecName: "utilities") pod "27c3c199-cc34-438a-ac59-4555ee7c5a1d" (UID: "27c3c199-cc34-438a-ac59-4555ee7c5a1d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:18:31 crc kubenswrapper[4632]: I0313 11:18:31.873339 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27c3c199-cc34-438a-ac59-4555ee7c5a1d-kube-api-access-x27fk" (OuterVolumeSpecName: "kube-api-access-x27fk") pod "27c3c199-cc34-438a-ac59-4555ee7c5a1d" (UID: "27c3c199-cc34-438a-ac59-4555ee7c5a1d"). InnerVolumeSpecName "kube-api-access-x27fk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:18:31 crc kubenswrapper[4632]: I0313 11:18:31.960599 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x27fk\" (UniqueName: \"kubernetes.io/projected/27c3c199-cc34-438a-ac59-4555ee7c5a1d-kube-api-access-x27fk\") on node \"crc\" DevicePath \"\"" Mar 13 11:18:31 crc kubenswrapper[4632]: I0313 11:18:31.960628 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27c3c199-cc34-438a-ac59-4555ee7c5a1d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:18:32 crc kubenswrapper[4632]: I0313 11:18:32.006586 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27c3c199-cc34-438a-ac59-4555ee7c5a1d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27c3c199-cc34-438a-ac59-4555ee7c5a1d" (UID: "27c3c199-cc34-438a-ac59-4555ee7c5a1d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:18:32 crc kubenswrapper[4632]: I0313 11:18:32.061985 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27c3c199-cc34-438a-ac59-4555ee7c5a1d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:18:32 crc kubenswrapper[4632]: I0313 11:18:32.738970 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gcqr2" Mar 13 11:18:32 crc kubenswrapper[4632]: I0313 11:18:32.769010 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gcqr2"] Mar 13 11:18:32 crc kubenswrapper[4632]: I0313 11:18:32.785009 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gcqr2"] Mar 13 11:18:34 crc kubenswrapper[4632]: I0313 11:18:34.073902 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" path="/var/lib/kubelet/pods/27c3c199-cc34-438a-ac59-4555ee7c5a1d/volumes" Mar 13 11:18:35 crc kubenswrapper[4632]: I0313 11:18:35.044459 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:18:35 crc kubenswrapper[4632]: E0313 11:18:35.045175 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:18:48 crc kubenswrapper[4632]: I0313 11:18:48.060092 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:18:48 crc kubenswrapper[4632]: E0313 11:18:48.061043 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:19:03 crc kubenswrapper[4632]: I0313 11:19:03.044530 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:19:03 crc kubenswrapper[4632]: E0313 11:19:03.045850 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:19:16 crc kubenswrapper[4632]: I0313 11:19:16.046606 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:19:16 crc kubenswrapper[4632]: E0313 11:19:16.047868 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:19:29 crc kubenswrapper[4632]: I0313 11:19:29.045016 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:19:29 crc kubenswrapper[4632]: E0313 11:19:29.045774 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:19:44 crc kubenswrapper[4632]: I0313 11:19:44.045196 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:19:44 crc kubenswrapper[4632]: E0313 11:19:44.046089 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:19:55 crc kubenswrapper[4632]: I0313 11:19:55.044213 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:19:55 crc kubenswrapper[4632]: E0313 11:19:55.045298 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.140511 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556680-g7sw8"] Mar 13 11:20:00 crc kubenswrapper[4632]: E0313 11:20:00.141303 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerName="extract-content" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.141315 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerName="extract-content" Mar 13 11:20:00 crc kubenswrapper[4632]: E0313 11:20:00.141328 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="feb87e78-c7fb-4997-869e-1e652f57ffe9" containerName="oc" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.141333 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="feb87e78-c7fb-4997-869e-1e652f57ffe9" containerName="oc" Mar 13 11:20:00 crc kubenswrapper[4632]: E0313 11:20:00.141341 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerName="extract-utilities" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.141347 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerName="extract-utilities" Mar 13 11:20:00 crc kubenswrapper[4632]: E0313 11:20:00.141370 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerName="registry-server" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.141376 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerName="registry-server" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.141561 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="feb87e78-c7fb-4997-869e-1e652f57ffe9" containerName="oc" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.141573 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="27c3c199-cc34-438a-ac59-4555ee7c5a1d" containerName="registry-server" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.142218 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556680-g7sw8" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.153227 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556680-g7sw8"] Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.171810 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.172075 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.172221 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.299554 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d29qr\" (UniqueName: \"kubernetes.io/projected/34a7cf92-d429-468d-9eff-e76b0302dee4-kube-api-access-d29qr\") pod \"auto-csr-approver-29556680-g7sw8\" (UID: \"34a7cf92-d429-468d-9eff-e76b0302dee4\") " pod="openshift-infra/auto-csr-approver-29556680-g7sw8" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.401793 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d29qr\" (UniqueName: \"kubernetes.io/projected/34a7cf92-d429-468d-9eff-e76b0302dee4-kube-api-access-d29qr\") pod \"auto-csr-approver-29556680-g7sw8\" (UID: \"34a7cf92-d429-468d-9eff-e76b0302dee4\") " pod="openshift-infra/auto-csr-approver-29556680-g7sw8" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.431302 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d29qr\" (UniqueName: \"kubernetes.io/projected/34a7cf92-d429-468d-9eff-e76b0302dee4-kube-api-access-d29qr\") pod \"auto-csr-approver-29556680-g7sw8\" (UID: \"34a7cf92-d429-468d-9eff-e76b0302dee4\") " pod="openshift-infra/auto-csr-approver-29556680-g7sw8" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.491430 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556680-g7sw8" Mar 13 11:20:00 crc kubenswrapper[4632]: I0313 11:20:00.992909 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556680-g7sw8"] Mar 13 11:20:01 crc kubenswrapper[4632]: I0313 11:20:01.010252 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 11:20:01 crc kubenswrapper[4632]: I0313 11:20:01.062396 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556680-g7sw8" event={"ID":"34a7cf92-d429-468d-9eff-e76b0302dee4","Type":"ContainerStarted","Data":"aad306172ca9a7df97cbb17f5fba2e3d8adc8878f978f0d8c1b464a869d49dd3"} Mar 13 11:20:03 crc kubenswrapper[4632]: I0313 11:20:03.092972 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556680-g7sw8" event={"ID":"34a7cf92-d429-468d-9eff-e76b0302dee4","Type":"ContainerStarted","Data":"a287e319f36103c4becc462637b974d036358eff92b92ae569b32780de4efe87"} Mar 13 11:20:04 crc kubenswrapper[4632]: I0313 11:20:04.104429 4632 generic.go:334] "Generic (PLEG): container finished" podID="34a7cf92-d429-468d-9eff-e76b0302dee4" containerID="a287e319f36103c4becc462637b974d036358eff92b92ae569b32780de4efe87" exitCode=0 Mar 13 11:20:04 crc kubenswrapper[4632]: I0313 11:20:04.104651 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556680-g7sw8" event={"ID":"34a7cf92-d429-468d-9eff-e76b0302dee4","Type":"ContainerDied","Data":"a287e319f36103c4becc462637b974d036358eff92b92ae569b32780de4efe87"} Mar 13 11:20:05 crc kubenswrapper[4632]: I0313 11:20:05.569672 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556680-g7sw8" Mar 13 11:20:05 crc kubenswrapper[4632]: I0313 11:20:05.738867 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d29qr\" (UniqueName: \"kubernetes.io/projected/34a7cf92-d429-468d-9eff-e76b0302dee4-kube-api-access-d29qr\") pod \"34a7cf92-d429-468d-9eff-e76b0302dee4\" (UID: \"34a7cf92-d429-468d-9eff-e76b0302dee4\") " Mar 13 11:20:05 crc kubenswrapper[4632]: I0313 11:20:05.750361 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34a7cf92-d429-468d-9eff-e76b0302dee4-kube-api-access-d29qr" (OuterVolumeSpecName: "kube-api-access-d29qr") pod "34a7cf92-d429-468d-9eff-e76b0302dee4" (UID: "34a7cf92-d429-468d-9eff-e76b0302dee4"). InnerVolumeSpecName "kube-api-access-d29qr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:20:05 crc kubenswrapper[4632]: I0313 11:20:05.841186 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d29qr\" (UniqueName: \"kubernetes.io/projected/34a7cf92-d429-468d-9eff-e76b0302dee4-kube-api-access-d29qr\") on node \"crc\" DevicePath \"\"" Mar 13 11:20:06 crc kubenswrapper[4632]: I0313 11:20:06.044506 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:20:06 crc kubenswrapper[4632]: E0313 11:20:06.044838 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:20:06 crc kubenswrapper[4632]: I0313 11:20:06.124103 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556680-g7sw8" event={"ID":"34a7cf92-d429-468d-9eff-e76b0302dee4","Type":"ContainerDied","Data":"aad306172ca9a7df97cbb17f5fba2e3d8adc8878f978f0d8c1b464a869d49dd3"} Mar 13 11:20:06 crc kubenswrapper[4632]: I0313 11:20:06.124153 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aad306172ca9a7df97cbb17f5fba2e3d8adc8878f978f0d8c1b464a869d49dd3" Mar 13 11:20:06 crc kubenswrapper[4632]: I0313 11:20:06.124217 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556680-g7sw8" Mar 13 11:20:06 crc kubenswrapper[4632]: I0313 11:20:06.641977 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556674-j6lkl"] Mar 13 11:20:06 crc kubenswrapper[4632]: I0313 11:20:06.653881 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556674-j6lkl"] Mar 13 11:20:08 crc kubenswrapper[4632]: I0313 11:20:08.062047 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2607e7bb-5f81-48cf-945a-6dee68b60040" path="/var/lib/kubelet/pods/2607e7bb-5f81-48cf-945a-6dee68b60040/volumes" Mar 13 11:20:18 crc kubenswrapper[4632]: I0313 11:20:18.176722 4632 scope.go:117] "RemoveContainer" containerID="54b5c748d72d1466e81773f53e5a61ff5e546e63d80706f8982cd195e971c601" Mar 13 11:20:20 crc kubenswrapper[4632]: I0313 11:20:20.044167 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:20:20 crc kubenswrapper[4632]: E0313 11:20:20.044678 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:20:34 crc kubenswrapper[4632]: I0313 11:20:34.044752 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:20:34 crc kubenswrapper[4632]: E0313 11:20:34.045750 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:20:49 crc kubenswrapper[4632]: I0313 11:20:49.044109 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:20:49 crc kubenswrapper[4632]: E0313 11:20:49.045079 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:21:01 crc kubenswrapper[4632]: I0313 11:21:01.044361 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:21:01 crc kubenswrapper[4632]: E0313 11:21:01.045226 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:21:13 crc kubenswrapper[4632]: I0313 11:21:13.044995 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:21:13 crc kubenswrapper[4632]: E0313 11:21:13.046251 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:21:28 crc kubenswrapper[4632]: I0313 11:21:28.058447 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:21:28 crc kubenswrapper[4632]: E0313 11:21:28.059470 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:21:42 crc kubenswrapper[4632]: I0313 11:21:42.044693 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:21:42 crc kubenswrapper[4632]: E0313 11:21:42.046868 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:21:56 crc kubenswrapper[4632]: I0313 11:21:56.044408 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:21:56 crc kubenswrapper[4632]: E0313 11:21:56.045144 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:22:00 crc kubenswrapper[4632]: I0313 11:22:00.152139 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556682-pm7lm"] Mar 13 11:22:00 crc kubenswrapper[4632]: E0313 11:22:00.152890 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a7cf92-d429-468d-9eff-e76b0302dee4" containerName="oc" Mar 13 11:22:00 crc kubenswrapper[4632]: I0313 11:22:00.152902 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a7cf92-d429-468d-9eff-e76b0302dee4" containerName="oc" Mar 13 11:22:00 crc kubenswrapper[4632]: I0313 11:22:00.153175 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="34a7cf92-d429-468d-9eff-e76b0302dee4" containerName="oc" Mar 13 11:22:00 crc kubenswrapper[4632]: I0313 11:22:00.153884 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556682-pm7lm" Mar 13 11:22:00 crc kubenswrapper[4632]: I0313 11:22:00.160317 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556682-pm7lm"] Mar 13 11:22:00 crc kubenswrapper[4632]: I0313 11:22:00.162256 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:22:00 crc kubenswrapper[4632]: I0313 11:22:00.162539 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:22:00 crc kubenswrapper[4632]: I0313 11:22:00.162865 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:22:00 crc kubenswrapper[4632]: I0313 11:22:00.254434 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgkd8\" (UniqueName: \"kubernetes.io/projected/191fb79a-448d-4181-8346-f9dec8721d81-kube-api-access-cgkd8\") pod \"auto-csr-approver-29556682-pm7lm\" (UID: \"191fb79a-448d-4181-8346-f9dec8721d81\") " pod="openshift-infra/auto-csr-approver-29556682-pm7lm" Mar 13 11:22:00 crc kubenswrapper[4632]: I0313 11:22:00.356022 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgkd8\" (UniqueName: \"kubernetes.io/projected/191fb79a-448d-4181-8346-f9dec8721d81-kube-api-access-cgkd8\") pod \"auto-csr-approver-29556682-pm7lm\" (UID: \"191fb79a-448d-4181-8346-f9dec8721d81\") " pod="openshift-infra/auto-csr-approver-29556682-pm7lm" Mar 13 11:22:01 crc kubenswrapper[4632]: I0313 11:22:01.070704 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgkd8\" (UniqueName: \"kubernetes.io/projected/191fb79a-448d-4181-8346-f9dec8721d81-kube-api-access-cgkd8\") pod \"auto-csr-approver-29556682-pm7lm\" (UID: \"191fb79a-448d-4181-8346-f9dec8721d81\") " pod="openshift-infra/auto-csr-approver-29556682-pm7lm" Mar 13 11:22:01 crc kubenswrapper[4632]: I0313 11:22:01.078117 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556682-pm7lm" Mar 13 11:22:01 crc kubenswrapper[4632]: I0313 11:22:01.648217 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556682-pm7lm"] Mar 13 11:22:02 crc kubenswrapper[4632]: I0313 11:22:02.428731 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556682-pm7lm" event={"ID":"191fb79a-448d-4181-8346-f9dec8721d81","Type":"ContainerStarted","Data":"9cac74b2414d34c5abeccf3930f3da4938ea48597909f1e379cb0c2aacbc5dbd"} Mar 13 11:22:04 crc kubenswrapper[4632]: I0313 11:22:04.450831 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556682-pm7lm" event={"ID":"191fb79a-448d-4181-8346-f9dec8721d81","Type":"ContainerStarted","Data":"6b6b471905ed6fd6c16476a8c40a8d65b889486b4d10fadb0c4b7b6cf7a150be"} Mar 13 11:22:04 crc kubenswrapper[4632]: I0313 11:22:04.475678 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556682-pm7lm" podStartSLOduration=3.423013903 podStartE2EDuration="4.475658595s" podCreationTimestamp="2026-03-13 11:22:00 +0000 UTC" firstStartedPulling="2026-03-13 11:22:01.657163303 +0000 UTC m=+4695.679693436" lastFinishedPulling="2026-03-13 11:22:02.709808005 +0000 UTC m=+4696.732338128" observedRunningTime="2026-03-13 11:22:04.474456005 +0000 UTC m=+4698.496986138" watchObservedRunningTime="2026-03-13 11:22:04.475658595 +0000 UTC m=+4698.498188728" Mar 13 11:22:06 crc kubenswrapper[4632]: I0313 11:22:06.479784 4632 generic.go:334] "Generic (PLEG): container finished" podID="191fb79a-448d-4181-8346-f9dec8721d81" containerID="6b6b471905ed6fd6c16476a8c40a8d65b889486b4d10fadb0c4b7b6cf7a150be" exitCode=0 Mar 13 11:22:06 crc kubenswrapper[4632]: I0313 11:22:06.479882 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556682-pm7lm" event={"ID":"191fb79a-448d-4181-8346-f9dec8721d81","Type":"ContainerDied","Data":"6b6b471905ed6fd6c16476a8c40a8d65b889486b4d10fadb0c4b7b6cf7a150be"} Mar 13 11:22:07 crc kubenswrapper[4632]: I0313 11:22:07.985383 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556682-pm7lm" Mar 13 11:22:08 crc kubenswrapper[4632]: I0313 11:22:08.057280 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:22:08 crc kubenswrapper[4632]: E0313 11:22:08.057783 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:22:08 crc kubenswrapper[4632]: I0313 11:22:08.117816 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgkd8\" (UniqueName: \"kubernetes.io/projected/191fb79a-448d-4181-8346-f9dec8721d81-kube-api-access-cgkd8\") pod \"191fb79a-448d-4181-8346-f9dec8721d81\" (UID: \"191fb79a-448d-4181-8346-f9dec8721d81\") " Mar 13 11:22:08 crc kubenswrapper[4632]: I0313 11:22:08.125186 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/191fb79a-448d-4181-8346-f9dec8721d81-kube-api-access-cgkd8" (OuterVolumeSpecName: "kube-api-access-cgkd8") pod "191fb79a-448d-4181-8346-f9dec8721d81" (UID: "191fb79a-448d-4181-8346-f9dec8721d81"). InnerVolumeSpecName "kube-api-access-cgkd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:22:08 crc kubenswrapper[4632]: I0313 11:22:08.221192 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgkd8\" (UniqueName: \"kubernetes.io/projected/191fb79a-448d-4181-8346-f9dec8721d81-kube-api-access-cgkd8\") on node \"crc\" DevicePath \"\"" Mar 13 11:22:08 crc kubenswrapper[4632]: I0313 11:22:08.503909 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556682-pm7lm" event={"ID":"191fb79a-448d-4181-8346-f9dec8721d81","Type":"ContainerDied","Data":"9cac74b2414d34c5abeccf3930f3da4938ea48597909f1e379cb0c2aacbc5dbd"} Mar 13 11:22:08 crc kubenswrapper[4632]: I0313 11:22:08.503962 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9cac74b2414d34c5abeccf3930f3da4938ea48597909f1e379cb0c2aacbc5dbd" Mar 13 11:22:08 crc kubenswrapper[4632]: I0313 11:22:08.504021 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556682-pm7lm" Mar 13 11:22:08 crc kubenswrapper[4632]: I0313 11:22:08.573050 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556676-8qz49"] Mar 13 11:22:08 crc kubenswrapper[4632]: I0313 11:22:08.580560 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556676-8qz49"] Mar 13 11:22:10 crc kubenswrapper[4632]: I0313 11:22:10.057873 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911" path="/var/lib/kubelet/pods/f97cdaa5-f90d-4cd0-9b62-7bb0c3b41911/volumes" Mar 13 11:22:18 crc kubenswrapper[4632]: I0313 11:22:18.303087 4632 scope.go:117] "RemoveContainer" containerID="1913657413fffb5d7b6f0c5a32e25db59682a49e52d39a0adc600808b4a0def3" Mar 13 11:22:22 crc kubenswrapper[4632]: I0313 11:22:22.044583 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:22:22 crc kubenswrapper[4632]: I0313 11:22:22.710264 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"49682358d72adf3dbebb4a70c4dbc847548d4046ae5ef96f55f2ae4dfd58b9f9"} Mar 13 11:24:00 crc kubenswrapper[4632]: I0313 11:24:00.165952 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556684-bd9st"] Mar 13 11:24:00 crc kubenswrapper[4632]: E0313 11:24:00.166819 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="191fb79a-448d-4181-8346-f9dec8721d81" containerName="oc" Mar 13 11:24:00 crc kubenswrapper[4632]: I0313 11:24:00.166830 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="191fb79a-448d-4181-8346-f9dec8721d81" containerName="oc" Mar 13 11:24:00 crc kubenswrapper[4632]: I0313 11:24:00.167191 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="191fb79a-448d-4181-8346-f9dec8721d81" containerName="oc" Mar 13 11:24:00 crc kubenswrapper[4632]: I0313 11:24:00.175372 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556684-bd9st" Mar 13 11:24:00 crc kubenswrapper[4632]: I0313 11:24:00.177378 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:24:00 crc kubenswrapper[4632]: I0313 11:24:00.182924 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556684-bd9st"] Mar 13 11:24:00 crc kubenswrapper[4632]: I0313 11:24:00.193389 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:24:00 crc kubenswrapper[4632]: I0313 11:24:00.193653 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:24:00 crc kubenswrapper[4632]: I0313 11:24:00.255424 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rh6ln\" (UniqueName: \"kubernetes.io/projected/fb45f1ce-58e0-4f55-afd6-2e14db5f24ca-kube-api-access-rh6ln\") pod \"auto-csr-approver-29556684-bd9st\" (UID: \"fb45f1ce-58e0-4f55-afd6-2e14db5f24ca\") " pod="openshift-infra/auto-csr-approver-29556684-bd9st" Mar 13 11:24:00 crc kubenswrapper[4632]: I0313 11:24:00.356889 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rh6ln\" (UniqueName: \"kubernetes.io/projected/fb45f1ce-58e0-4f55-afd6-2e14db5f24ca-kube-api-access-rh6ln\") pod \"auto-csr-approver-29556684-bd9st\" (UID: \"fb45f1ce-58e0-4f55-afd6-2e14db5f24ca\") " pod="openshift-infra/auto-csr-approver-29556684-bd9st" Mar 13 11:24:00 crc kubenswrapper[4632]: I0313 11:24:00.392883 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rh6ln\" (UniqueName: \"kubernetes.io/projected/fb45f1ce-58e0-4f55-afd6-2e14db5f24ca-kube-api-access-rh6ln\") pod \"auto-csr-approver-29556684-bd9st\" (UID: \"fb45f1ce-58e0-4f55-afd6-2e14db5f24ca\") " pod="openshift-infra/auto-csr-approver-29556684-bd9st" Mar 13 11:24:00 crc kubenswrapper[4632]: I0313 11:24:00.517721 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556684-bd9st" Mar 13 11:24:01 crc kubenswrapper[4632]: I0313 11:24:01.143205 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556684-bd9st"] Mar 13 11:24:01 crc kubenswrapper[4632]: I0313 11:24:01.905220 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556684-bd9st" event={"ID":"fb45f1ce-58e0-4f55-afd6-2e14db5f24ca","Type":"ContainerStarted","Data":"9a6bcaf29573d59c8877c5b6b24a62abc2b6f3f927437f93288a174ead4c9ab5"} Mar 13 11:24:02 crc kubenswrapper[4632]: I0313 11:24:02.918425 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556684-bd9st" event={"ID":"fb45f1ce-58e0-4f55-afd6-2e14db5f24ca","Type":"ContainerStarted","Data":"2af97c9efffc6f3dc7413dcff6c97889a640ef442506af8ad264876a675427dc"} Mar 13 11:24:02 crc kubenswrapper[4632]: I0313 11:24:02.941244 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556684-bd9st" podStartSLOduration=2.024133913 podStartE2EDuration="2.941222131s" podCreationTimestamp="2026-03-13 11:24:00 +0000 UTC" firstStartedPulling="2026-03-13 11:24:01.152431654 +0000 UTC m=+4815.174961787" lastFinishedPulling="2026-03-13 11:24:02.069519872 +0000 UTC m=+4816.092050005" observedRunningTime="2026-03-13 11:24:02.936216567 +0000 UTC m=+4816.958746710" watchObservedRunningTime="2026-03-13 11:24:02.941222131 +0000 UTC m=+4816.963752264" Mar 13 11:24:03 crc kubenswrapper[4632]: I0313 11:24:03.928606 4632 generic.go:334] "Generic (PLEG): container finished" podID="fb45f1ce-58e0-4f55-afd6-2e14db5f24ca" containerID="2af97c9efffc6f3dc7413dcff6c97889a640ef442506af8ad264876a675427dc" exitCode=0 Mar 13 11:24:03 crc kubenswrapper[4632]: I0313 11:24:03.928662 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556684-bd9st" event={"ID":"fb45f1ce-58e0-4f55-afd6-2e14db5f24ca","Type":"ContainerDied","Data":"2af97c9efffc6f3dc7413dcff6c97889a640ef442506af8ad264876a675427dc"} Mar 13 11:24:05 crc kubenswrapper[4632]: I0313 11:24:05.340962 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556684-bd9st" Mar 13 11:24:05 crc kubenswrapper[4632]: I0313 11:24:05.456744 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rh6ln\" (UniqueName: \"kubernetes.io/projected/fb45f1ce-58e0-4f55-afd6-2e14db5f24ca-kube-api-access-rh6ln\") pod \"fb45f1ce-58e0-4f55-afd6-2e14db5f24ca\" (UID: \"fb45f1ce-58e0-4f55-afd6-2e14db5f24ca\") " Mar 13 11:24:05 crc kubenswrapper[4632]: I0313 11:24:05.464123 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb45f1ce-58e0-4f55-afd6-2e14db5f24ca-kube-api-access-rh6ln" (OuterVolumeSpecName: "kube-api-access-rh6ln") pod "fb45f1ce-58e0-4f55-afd6-2e14db5f24ca" (UID: "fb45f1ce-58e0-4f55-afd6-2e14db5f24ca"). InnerVolumeSpecName "kube-api-access-rh6ln". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:24:05 crc kubenswrapper[4632]: I0313 11:24:05.559734 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rh6ln\" (UniqueName: \"kubernetes.io/projected/fb45f1ce-58e0-4f55-afd6-2e14db5f24ca-kube-api-access-rh6ln\") on node \"crc\" DevicePath \"\"" Mar 13 11:24:05 crc kubenswrapper[4632]: I0313 11:24:05.958986 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556684-bd9st" event={"ID":"fb45f1ce-58e0-4f55-afd6-2e14db5f24ca","Type":"ContainerDied","Data":"9a6bcaf29573d59c8877c5b6b24a62abc2b6f3f927437f93288a174ead4c9ab5"} Mar 13 11:24:05 crc kubenswrapper[4632]: I0313 11:24:05.959066 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9a6bcaf29573d59c8877c5b6b24a62abc2b6f3f927437f93288a174ead4c9ab5" Mar 13 11:24:05 crc kubenswrapper[4632]: I0313 11:24:05.959092 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556684-bd9st" Mar 13 11:24:06 crc kubenswrapper[4632]: I0313 11:24:06.023003 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556678-kxcj9"] Mar 13 11:24:06 crc kubenswrapper[4632]: I0313 11:24:06.033868 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556678-kxcj9"] Mar 13 11:24:06 crc kubenswrapper[4632]: I0313 11:24:06.058340 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="feb87e78-c7fb-4997-869e-1e652f57ffe9" path="/var/lib/kubelet/pods/feb87e78-c7fb-4997-869e-1e652f57ffe9/volumes" Mar 13 11:24:18 crc kubenswrapper[4632]: I0313 11:24:18.426756 4632 scope.go:117] "RemoveContainer" containerID="d5beb81ff52cba139334670e53dbdbf15336383f8949e10d3fb4d56429b9cd89" Mar 13 11:24:18 crc kubenswrapper[4632]: I0313 11:24:18.547115 4632 scope.go:117] "RemoveContainer" containerID="098f5138f07f45ec1bf0ad95c6647f82c5303be4fdf81f24941c5f5d332834fa" Mar 13 11:24:18 crc kubenswrapper[4632]: I0313 11:24:18.611740 4632 scope.go:117] "RemoveContainer" containerID="f9fc93bd191406b3e302e5390ddb104f245922f0c823cd3aa32e935c54c7057f" Mar 13 11:24:18 crc kubenswrapper[4632]: I0313 11:24:18.660655 4632 scope.go:117] "RemoveContainer" containerID="8fc46f8a96264e119d9565fb1071d38b9eca7de4da0002eb50dc0db108fea360" Mar 13 11:24:40 crc kubenswrapper[4632]: I0313 11:24:40.462471 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:24:40 crc kubenswrapper[4632]: I0313 11:24:40.464839 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:25:10 crc kubenswrapper[4632]: I0313 11:25:10.461609 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:25:10 crc kubenswrapper[4632]: I0313 11:25:10.462294 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:25:40 crc kubenswrapper[4632]: I0313 11:25:40.461202 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:25:40 crc kubenswrapper[4632]: I0313 11:25:40.461910 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:25:40 crc kubenswrapper[4632]: I0313 11:25:40.461985 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 11:25:40 crc kubenswrapper[4632]: I0313 11:25:40.464767 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"49682358d72adf3dbebb4a70c4dbc847548d4046ae5ef96f55f2ae4dfd58b9f9"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 11:25:40 crc kubenswrapper[4632]: I0313 11:25:40.464854 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://49682358d72adf3dbebb4a70c4dbc847548d4046ae5ef96f55f2ae4dfd58b9f9" gracePeriod=600 Mar 13 11:25:41 crc kubenswrapper[4632]: I0313 11:25:41.014034 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="49682358d72adf3dbebb4a70c4dbc847548d4046ae5ef96f55f2ae4dfd58b9f9" exitCode=0 Mar 13 11:25:41 crc kubenswrapper[4632]: I0313 11:25:41.014097 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"49682358d72adf3dbebb4a70c4dbc847548d4046ae5ef96f55f2ae4dfd58b9f9"} Mar 13 11:25:41 crc kubenswrapper[4632]: I0313 11:25:41.014669 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3"} Mar 13 11:25:41 crc kubenswrapper[4632]: I0313 11:25:41.014697 4632 scope.go:117] "RemoveContainer" containerID="70b8629c42fc12676389ef5c404d5d088a40865f44432caaaf2da0b0a7e69c7a" Mar 13 11:26:00 crc kubenswrapper[4632]: I0313 11:26:00.191662 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556686-v9c2p"] Mar 13 11:26:00 crc kubenswrapper[4632]: E0313 11:26:00.193929 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb45f1ce-58e0-4f55-afd6-2e14db5f24ca" containerName="oc" Mar 13 11:26:00 crc kubenswrapper[4632]: I0313 11:26:00.194079 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb45f1ce-58e0-4f55-afd6-2e14db5f24ca" containerName="oc" Mar 13 11:26:00 crc kubenswrapper[4632]: I0313 11:26:00.194445 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb45f1ce-58e0-4f55-afd6-2e14db5f24ca" containerName="oc" Mar 13 11:26:00 crc kubenswrapper[4632]: I0313 11:26:00.199706 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556686-v9c2p" Mar 13 11:26:00 crc kubenswrapper[4632]: I0313 11:26:00.205870 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:26:00 crc kubenswrapper[4632]: I0313 11:26:00.206034 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:26:00 crc kubenswrapper[4632]: I0313 11:26:00.206154 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:26:00 crc kubenswrapper[4632]: I0313 11:26:00.208163 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556686-v9c2p"] Mar 13 11:26:00 crc kubenswrapper[4632]: I0313 11:26:00.300677 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vjfs\" (UniqueName: \"kubernetes.io/projected/78358d6e-af24-4da5-8c77-9453e6228cda-kube-api-access-5vjfs\") pod \"auto-csr-approver-29556686-v9c2p\" (UID: \"78358d6e-af24-4da5-8c77-9453e6228cda\") " pod="openshift-infra/auto-csr-approver-29556686-v9c2p" Mar 13 11:26:00 crc kubenswrapper[4632]: I0313 11:26:00.402958 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vjfs\" (UniqueName: \"kubernetes.io/projected/78358d6e-af24-4da5-8c77-9453e6228cda-kube-api-access-5vjfs\") pod \"auto-csr-approver-29556686-v9c2p\" (UID: \"78358d6e-af24-4da5-8c77-9453e6228cda\") " pod="openshift-infra/auto-csr-approver-29556686-v9c2p" Mar 13 11:26:00 crc kubenswrapper[4632]: I0313 11:26:00.429849 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vjfs\" (UniqueName: \"kubernetes.io/projected/78358d6e-af24-4da5-8c77-9453e6228cda-kube-api-access-5vjfs\") pod \"auto-csr-approver-29556686-v9c2p\" (UID: \"78358d6e-af24-4da5-8c77-9453e6228cda\") " pod="openshift-infra/auto-csr-approver-29556686-v9c2p" Mar 13 11:26:00 crc kubenswrapper[4632]: I0313 11:26:00.522537 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556686-v9c2p" Mar 13 11:26:01 crc kubenswrapper[4632]: I0313 11:26:01.304197 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556686-v9c2p"] Mar 13 11:26:01 crc kubenswrapper[4632]: W0313 11:26:01.313656 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod78358d6e_af24_4da5_8c77_9453e6228cda.slice/crio-a9e0fec35a927773c559298dfd634557a6a7a9da41447ae6a5ede71fb7933809 WatchSource:0}: Error finding container a9e0fec35a927773c559298dfd634557a6a7a9da41447ae6a5ede71fb7933809: Status 404 returned error can't find the container with id a9e0fec35a927773c559298dfd634557a6a7a9da41447ae6a5ede71fb7933809 Mar 13 11:26:01 crc kubenswrapper[4632]: I0313 11:26:01.316711 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 11:26:02 crc kubenswrapper[4632]: I0313 11:26:02.274839 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556686-v9c2p" event={"ID":"78358d6e-af24-4da5-8c77-9453e6228cda","Type":"ContainerStarted","Data":"a9e0fec35a927773c559298dfd634557a6a7a9da41447ae6a5ede71fb7933809"} Mar 13 11:26:03 crc kubenswrapper[4632]: I0313 11:26:03.289230 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556686-v9c2p" event={"ID":"78358d6e-af24-4da5-8c77-9453e6228cda","Type":"ContainerStarted","Data":"0306bee576d03326b01d2c08c76cf2909394a8ec9e729a13ca12d86ebb721532"} Mar 13 11:26:03 crc kubenswrapper[4632]: I0313 11:26:03.313823 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556686-v9c2p" podStartSLOduration=2.167040802 podStartE2EDuration="3.313801869s" podCreationTimestamp="2026-03-13 11:26:00 +0000 UTC" firstStartedPulling="2026-03-13 11:26:01.316454439 +0000 UTC m=+4935.338984562" lastFinishedPulling="2026-03-13 11:26:02.463215496 +0000 UTC m=+4936.485745629" observedRunningTime="2026-03-13 11:26:03.304335836 +0000 UTC m=+4937.326865999" watchObservedRunningTime="2026-03-13 11:26:03.313801869 +0000 UTC m=+4937.336332022" Mar 13 11:26:04 crc kubenswrapper[4632]: I0313 11:26:04.298665 4632 generic.go:334] "Generic (PLEG): container finished" podID="78358d6e-af24-4da5-8c77-9453e6228cda" containerID="0306bee576d03326b01d2c08c76cf2909394a8ec9e729a13ca12d86ebb721532" exitCode=0 Mar 13 11:26:04 crc kubenswrapper[4632]: I0313 11:26:04.298795 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556686-v9c2p" event={"ID":"78358d6e-af24-4da5-8c77-9453e6228cda","Type":"ContainerDied","Data":"0306bee576d03326b01d2c08c76cf2909394a8ec9e729a13ca12d86ebb721532"} Mar 13 11:26:05 crc kubenswrapper[4632]: I0313 11:26:05.718452 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556686-v9c2p" Mar 13 11:26:05 crc kubenswrapper[4632]: I0313 11:26:05.809678 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vjfs\" (UniqueName: \"kubernetes.io/projected/78358d6e-af24-4da5-8c77-9453e6228cda-kube-api-access-5vjfs\") pod \"78358d6e-af24-4da5-8c77-9453e6228cda\" (UID: \"78358d6e-af24-4da5-8c77-9453e6228cda\") " Mar 13 11:26:05 crc kubenswrapper[4632]: I0313 11:26:05.833177 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/78358d6e-af24-4da5-8c77-9453e6228cda-kube-api-access-5vjfs" (OuterVolumeSpecName: "kube-api-access-5vjfs") pod "78358d6e-af24-4da5-8c77-9453e6228cda" (UID: "78358d6e-af24-4da5-8c77-9453e6228cda"). InnerVolumeSpecName "kube-api-access-5vjfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:26:05 crc kubenswrapper[4632]: I0313 11:26:05.912276 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vjfs\" (UniqueName: \"kubernetes.io/projected/78358d6e-af24-4da5-8c77-9453e6228cda-kube-api-access-5vjfs\") on node \"crc\" DevicePath \"\"" Mar 13 11:26:06 crc kubenswrapper[4632]: I0313 11:26:06.320098 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556686-v9c2p" event={"ID":"78358d6e-af24-4da5-8c77-9453e6228cda","Type":"ContainerDied","Data":"a9e0fec35a927773c559298dfd634557a6a7a9da41447ae6a5ede71fb7933809"} Mar 13 11:26:06 crc kubenswrapper[4632]: I0313 11:26:06.320140 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a9e0fec35a927773c559298dfd634557a6a7a9da41447ae6a5ede71fb7933809" Mar 13 11:26:06 crc kubenswrapper[4632]: I0313 11:26:06.320136 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556686-v9c2p" Mar 13 11:26:06 crc kubenswrapper[4632]: I0313 11:26:06.395164 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556680-g7sw8"] Mar 13 11:26:06 crc kubenswrapper[4632]: I0313 11:26:06.403393 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556680-g7sw8"] Mar 13 11:26:08 crc kubenswrapper[4632]: I0313 11:26:08.060696 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34a7cf92-d429-468d-9eff-e76b0302dee4" path="/var/lib/kubelet/pods/34a7cf92-d429-468d-9eff-e76b0302dee4/volumes" Mar 13 11:26:16 crc kubenswrapper[4632]: I0313 11:26:16.705058 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tctdq"] Mar 13 11:26:16 crc kubenswrapper[4632]: E0313 11:26:16.706093 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="78358d6e-af24-4da5-8c77-9453e6228cda" containerName="oc" Mar 13 11:26:16 crc kubenswrapper[4632]: I0313 11:26:16.706110 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="78358d6e-af24-4da5-8c77-9453e6228cda" containerName="oc" Mar 13 11:26:16 crc kubenswrapper[4632]: I0313 11:26:16.706362 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="78358d6e-af24-4da5-8c77-9453e6228cda" containerName="oc" Mar 13 11:26:16 crc kubenswrapper[4632]: I0313 11:26:16.709933 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:16 crc kubenswrapper[4632]: I0313 11:26:16.728330 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tctdq"] Mar 13 11:26:16 crc kubenswrapper[4632]: I0313 11:26:16.836826 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rt6sp\" (UniqueName: \"kubernetes.io/projected/743aacd5-2974-48e5-b2c8-ed448225edc3-kube-api-access-rt6sp\") pod \"redhat-marketplace-tctdq\" (UID: \"743aacd5-2974-48e5-b2c8-ed448225edc3\") " pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:16 crc kubenswrapper[4632]: I0313 11:26:16.836929 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743aacd5-2974-48e5-b2c8-ed448225edc3-utilities\") pod \"redhat-marketplace-tctdq\" (UID: \"743aacd5-2974-48e5-b2c8-ed448225edc3\") " pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:16 crc kubenswrapper[4632]: I0313 11:26:16.837662 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743aacd5-2974-48e5-b2c8-ed448225edc3-catalog-content\") pod \"redhat-marketplace-tctdq\" (UID: \"743aacd5-2974-48e5-b2c8-ed448225edc3\") " pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:16 crc kubenswrapper[4632]: I0313 11:26:16.939580 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743aacd5-2974-48e5-b2c8-ed448225edc3-utilities\") pod \"redhat-marketplace-tctdq\" (UID: \"743aacd5-2974-48e5-b2c8-ed448225edc3\") " pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:16 crc kubenswrapper[4632]: I0313 11:26:16.939648 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743aacd5-2974-48e5-b2c8-ed448225edc3-catalog-content\") pod \"redhat-marketplace-tctdq\" (UID: \"743aacd5-2974-48e5-b2c8-ed448225edc3\") " pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:16 crc kubenswrapper[4632]: I0313 11:26:16.939745 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rt6sp\" (UniqueName: \"kubernetes.io/projected/743aacd5-2974-48e5-b2c8-ed448225edc3-kube-api-access-rt6sp\") pod \"redhat-marketplace-tctdq\" (UID: \"743aacd5-2974-48e5-b2c8-ed448225edc3\") " pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:16 crc kubenswrapper[4632]: I0313 11:26:16.940245 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743aacd5-2974-48e5-b2c8-ed448225edc3-utilities\") pod \"redhat-marketplace-tctdq\" (UID: \"743aacd5-2974-48e5-b2c8-ed448225edc3\") " pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:16 crc kubenswrapper[4632]: I0313 11:26:16.940305 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743aacd5-2974-48e5-b2c8-ed448225edc3-catalog-content\") pod \"redhat-marketplace-tctdq\" (UID: \"743aacd5-2974-48e5-b2c8-ed448225edc3\") " pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:17 crc kubenswrapper[4632]: I0313 11:26:17.556711 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rt6sp\" (UniqueName: \"kubernetes.io/projected/743aacd5-2974-48e5-b2c8-ed448225edc3-kube-api-access-rt6sp\") pod \"redhat-marketplace-tctdq\" (UID: \"743aacd5-2974-48e5-b2c8-ed448225edc3\") " pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:17 crc kubenswrapper[4632]: I0313 11:26:17.633275 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:18 crc kubenswrapper[4632]: I0313 11:26:18.209013 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tctdq"] Mar 13 11:26:18 crc kubenswrapper[4632]: I0313 11:26:18.453334 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tctdq" event={"ID":"743aacd5-2974-48e5-b2c8-ed448225edc3","Type":"ContainerStarted","Data":"d989bc3bcc743f71b5d3bb9088655740c776263f3efde0646f3c95fa1bdc6620"} Mar 13 11:26:18 crc kubenswrapper[4632]: I0313 11:26:18.453708 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tctdq" event={"ID":"743aacd5-2974-48e5-b2c8-ed448225edc3","Type":"ContainerStarted","Data":"5bdb6190ec028b9a390285458bd532667460f65f6a7b1d61e8a16b8f194afa5a"} Mar 13 11:26:18 crc kubenswrapper[4632]: I0313 11:26:18.953491 4632 scope.go:117] "RemoveContainer" containerID="a287e319f36103c4becc462637b974d036358eff92b92ae569b32780de4efe87" Mar 13 11:26:19 crc kubenswrapper[4632]: I0313 11:26:19.463708 4632 generic.go:334] "Generic (PLEG): container finished" podID="743aacd5-2974-48e5-b2c8-ed448225edc3" containerID="d989bc3bcc743f71b5d3bb9088655740c776263f3efde0646f3c95fa1bdc6620" exitCode=0 Mar 13 11:26:19 crc kubenswrapper[4632]: I0313 11:26:19.463891 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tctdq" event={"ID":"743aacd5-2974-48e5-b2c8-ed448225edc3","Type":"ContainerDied","Data":"d989bc3bcc743f71b5d3bb9088655740c776263f3efde0646f3c95fa1bdc6620"} Mar 13 11:26:20 crc kubenswrapper[4632]: I0313 11:26:20.478963 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tctdq" event={"ID":"743aacd5-2974-48e5-b2c8-ed448225edc3","Type":"ContainerStarted","Data":"cde224ff7ac851fa03bc99aec7f4fa42671c6966019c97205cc33ba61daa73a2"} Mar 13 11:26:22 crc kubenswrapper[4632]: I0313 11:26:22.504008 4632 generic.go:334] "Generic (PLEG): container finished" podID="743aacd5-2974-48e5-b2c8-ed448225edc3" containerID="cde224ff7ac851fa03bc99aec7f4fa42671c6966019c97205cc33ba61daa73a2" exitCode=0 Mar 13 11:26:22 crc kubenswrapper[4632]: I0313 11:26:22.504047 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tctdq" event={"ID":"743aacd5-2974-48e5-b2c8-ed448225edc3","Type":"ContainerDied","Data":"cde224ff7ac851fa03bc99aec7f4fa42671c6966019c97205cc33ba61daa73a2"} Mar 13 11:26:24 crc kubenswrapper[4632]: I0313 11:26:24.533560 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tctdq" event={"ID":"743aacd5-2974-48e5-b2c8-ed448225edc3","Type":"ContainerStarted","Data":"caacbc7cb0001a5e23da731e23554473632bc37a4fa3b14376fde42cab9413c2"} Mar 13 11:26:24 crc kubenswrapper[4632]: I0313 11:26:24.556372 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tctdq" podStartSLOduration=4.81532544 podStartE2EDuration="8.556350258s" podCreationTimestamp="2026-03-13 11:26:16 +0000 UTC" firstStartedPulling="2026-03-13 11:26:19.715655216 +0000 UTC m=+4953.738185349" lastFinishedPulling="2026-03-13 11:26:23.456680034 +0000 UTC m=+4957.479210167" observedRunningTime="2026-03-13 11:26:24.554884191 +0000 UTC m=+4958.577414334" watchObservedRunningTime="2026-03-13 11:26:24.556350258 +0000 UTC m=+4958.578880401" Mar 13 11:26:27 crc kubenswrapper[4632]: I0313 11:26:27.633914 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:27 crc kubenswrapper[4632]: I0313 11:26:27.634683 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:28 crc kubenswrapper[4632]: I0313 11:26:28.679759 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-tctdq" podUID="743aacd5-2974-48e5-b2c8-ed448225edc3" containerName="registry-server" probeResult="failure" output=< Mar 13 11:26:28 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:26:28 crc kubenswrapper[4632]: > Mar 13 11:26:37 crc kubenswrapper[4632]: I0313 11:26:37.689313 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:37 crc kubenswrapper[4632]: I0313 11:26:37.750619 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:37 crc kubenswrapper[4632]: I0313 11:26:37.933776 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tctdq"] Mar 13 11:26:39 crc kubenswrapper[4632]: I0313 11:26:39.683831 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tctdq" podUID="743aacd5-2974-48e5-b2c8-ed448225edc3" containerName="registry-server" containerID="cri-o://caacbc7cb0001a5e23da731e23554473632bc37a4fa3b14376fde42cab9413c2" gracePeriod=2 Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.311955 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.343569 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6nl2m"] Mar 13 11:26:40 crc kubenswrapper[4632]: E0313 11:26:40.344148 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743aacd5-2974-48e5-b2c8-ed448225edc3" containerName="registry-server" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.344172 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="743aacd5-2974-48e5-b2c8-ed448225edc3" containerName="registry-server" Mar 13 11:26:40 crc kubenswrapper[4632]: E0313 11:26:40.344196 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743aacd5-2974-48e5-b2c8-ed448225edc3" containerName="extract-content" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.344203 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="743aacd5-2974-48e5-b2c8-ed448225edc3" containerName="extract-content" Mar 13 11:26:40 crc kubenswrapper[4632]: E0313 11:26:40.344241 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="743aacd5-2974-48e5-b2c8-ed448225edc3" containerName="extract-utilities" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.344249 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="743aacd5-2974-48e5-b2c8-ed448225edc3" containerName="extract-utilities" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.344499 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="743aacd5-2974-48e5-b2c8-ed448225edc3" containerName="registry-server" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.346212 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.370971 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6nl2m"] Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.450290 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743aacd5-2974-48e5-b2c8-ed448225edc3-utilities\") pod \"743aacd5-2974-48e5-b2c8-ed448225edc3\" (UID: \"743aacd5-2974-48e5-b2c8-ed448225edc3\") " Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.450587 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743aacd5-2974-48e5-b2c8-ed448225edc3-catalog-content\") pod \"743aacd5-2974-48e5-b2c8-ed448225edc3\" (UID: \"743aacd5-2974-48e5-b2c8-ed448225edc3\") " Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.450651 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rt6sp\" (UniqueName: \"kubernetes.io/projected/743aacd5-2974-48e5-b2c8-ed448225edc3-kube-api-access-rt6sp\") pod \"743aacd5-2974-48e5-b2c8-ed448225edc3\" (UID: \"743aacd5-2974-48e5-b2c8-ed448225edc3\") " Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.451191 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45f934fb-e679-4600-98e5-a67888251b13-catalog-content\") pod \"certified-operators-6nl2m\" (UID: \"45f934fb-e679-4600-98e5-a67888251b13\") " pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.451271 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45f934fb-e679-4600-98e5-a67888251b13-utilities\") pod \"certified-operators-6nl2m\" (UID: \"45f934fb-e679-4600-98e5-a67888251b13\") " pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.451307 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dn74\" (UniqueName: \"kubernetes.io/projected/45f934fb-e679-4600-98e5-a67888251b13-kube-api-access-2dn74\") pod \"certified-operators-6nl2m\" (UID: \"45f934fb-e679-4600-98e5-a67888251b13\") " pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.451332 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/743aacd5-2974-48e5-b2c8-ed448225edc3-utilities" (OuterVolumeSpecName: "utilities") pod "743aacd5-2974-48e5-b2c8-ed448225edc3" (UID: "743aacd5-2974-48e5-b2c8-ed448225edc3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.479641 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/743aacd5-2974-48e5-b2c8-ed448225edc3-kube-api-access-rt6sp" (OuterVolumeSpecName: "kube-api-access-rt6sp") pod "743aacd5-2974-48e5-b2c8-ed448225edc3" (UID: "743aacd5-2974-48e5-b2c8-ed448225edc3"). InnerVolumeSpecName "kube-api-access-rt6sp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.481994 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/743aacd5-2974-48e5-b2c8-ed448225edc3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "743aacd5-2974-48e5-b2c8-ed448225edc3" (UID: "743aacd5-2974-48e5-b2c8-ed448225edc3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.552687 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45f934fb-e679-4600-98e5-a67888251b13-catalog-content\") pod \"certified-operators-6nl2m\" (UID: \"45f934fb-e679-4600-98e5-a67888251b13\") " pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.552751 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45f934fb-e679-4600-98e5-a67888251b13-utilities\") pod \"certified-operators-6nl2m\" (UID: \"45f934fb-e679-4600-98e5-a67888251b13\") " pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.552787 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dn74\" (UniqueName: \"kubernetes.io/projected/45f934fb-e679-4600-98e5-a67888251b13-kube-api-access-2dn74\") pod \"certified-operators-6nl2m\" (UID: \"45f934fb-e679-4600-98e5-a67888251b13\") " pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.552926 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/743aacd5-2974-48e5-b2c8-ed448225edc3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.552956 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rt6sp\" (UniqueName: \"kubernetes.io/projected/743aacd5-2974-48e5-b2c8-ed448225edc3-kube-api-access-rt6sp\") on node \"crc\" DevicePath \"\"" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.552968 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/743aacd5-2974-48e5-b2c8-ed448225edc3-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.553194 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45f934fb-e679-4600-98e5-a67888251b13-catalog-content\") pod \"certified-operators-6nl2m\" (UID: \"45f934fb-e679-4600-98e5-a67888251b13\") " pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.553347 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45f934fb-e679-4600-98e5-a67888251b13-utilities\") pod \"certified-operators-6nl2m\" (UID: \"45f934fb-e679-4600-98e5-a67888251b13\") " pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.569566 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dn74\" (UniqueName: \"kubernetes.io/projected/45f934fb-e679-4600-98e5-a67888251b13-kube-api-access-2dn74\") pod \"certified-operators-6nl2m\" (UID: \"45f934fb-e679-4600-98e5-a67888251b13\") " pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.680063 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.696216 4632 generic.go:334] "Generic (PLEG): container finished" podID="743aacd5-2974-48e5-b2c8-ed448225edc3" containerID="caacbc7cb0001a5e23da731e23554473632bc37a4fa3b14376fde42cab9413c2" exitCode=0 Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.696277 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tctdq" event={"ID":"743aacd5-2974-48e5-b2c8-ed448225edc3","Type":"ContainerDied","Data":"caacbc7cb0001a5e23da731e23554473632bc37a4fa3b14376fde42cab9413c2"} Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.696378 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tctdq" event={"ID":"743aacd5-2974-48e5-b2c8-ed448225edc3","Type":"ContainerDied","Data":"5bdb6190ec028b9a390285458bd532667460f65f6a7b1d61e8a16b8f194afa5a"} Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.696299 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tctdq" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.696404 4632 scope.go:117] "RemoveContainer" containerID="caacbc7cb0001a5e23da731e23554473632bc37a4fa3b14376fde42cab9413c2" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.730437 4632 scope.go:117] "RemoveContainer" containerID="cde224ff7ac851fa03bc99aec7f4fa42671c6966019c97205cc33ba61daa73a2" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.753178 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tctdq"] Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.779896 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tctdq"] Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.783018 4632 scope.go:117] "RemoveContainer" containerID="d989bc3bcc743f71b5d3bb9088655740c776263f3efde0646f3c95fa1bdc6620" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.813193 4632 scope.go:117] "RemoveContainer" containerID="caacbc7cb0001a5e23da731e23554473632bc37a4fa3b14376fde42cab9413c2" Mar 13 11:26:40 crc kubenswrapper[4632]: E0313 11:26:40.822100 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"caacbc7cb0001a5e23da731e23554473632bc37a4fa3b14376fde42cab9413c2\": container with ID starting with caacbc7cb0001a5e23da731e23554473632bc37a4fa3b14376fde42cab9413c2 not found: ID does not exist" containerID="caacbc7cb0001a5e23da731e23554473632bc37a4fa3b14376fde42cab9413c2" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.822157 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"caacbc7cb0001a5e23da731e23554473632bc37a4fa3b14376fde42cab9413c2"} err="failed to get container status \"caacbc7cb0001a5e23da731e23554473632bc37a4fa3b14376fde42cab9413c2\": rpc error: code = NotFound desc = could not find container \"caacbc7cb0001a5e23da731e23554473632bc37a4fa3b14376fde42cab9413c2\": container with ID starting with caacbc7cb0001a5e23da731e23554473632bc37a4fa3b14376fde42cab9413c2 not found: ID does not exist" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.822185 4632 scope.go:117] "RemoveContainer" containerID="cde224ff7ac851fa03bc99aec7f4fa42671c6966019c97205cc33ba61daa73a2" Mar 13 11:26:40 crc kubenswrapper[4632]: E0313 11:26:40.826096 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cde224ff7ac851fa03bc99aec7f4fa42671c6966019c97205cc33ba61daa73a2\": container with ID starting with cde224ff7ac851fa03bc99aec7f4fa42671c6966019c97205cc33ba61daa73a2 not found: ID does not exist" containerID="cde224ff7ac851fa03bc99aec7f4fa42671c6966019c97205cc33ba61daa73a2" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.826141 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cde224ff7ac851fa03bc99aec7f4fa42671c6966019c97205cc33ba61daa73a2"} err="failed to get container status \"cde224ff7ac851fa03bc99aec7f4fa42671c6966019c97205cc33ba61daa73a2\": rpc error: code = NotFound desc = could not find container \"cde224ff7ac851fa03bc99aec7f4fa42671c6966019c97205cc33ba61daa73a2\": container with ID starting with cde224ff7ac851fa03bc99aec7f4fa42671c6966019c97205cc33ba61daa73a2 not found: ID does not exist" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.826167 4632 scope.go:117] "RemoveContainer" containerID="d989bc3bcc743f71b5d3bb9088655740c776263f3efde0646f3c95fa1bdc6620" Mar 13 11:26:40 crc kubenswrapper[4632]: E0313 11:26:40.835166 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d989bc3bcc743f71b5d3bb9088655740c776263f3efde0646f3c95fa1bdc6620\": container with ID starting with d989bc3bcc743f71b5d3bb9088655740c776263f3efde0646f3c95fa1bdc6620 not found: ID does not exist" containerID="d989bc3bcc743f71b5d3bb9088655740c776263f3efde0646f3c95fa1bdc6620" Mar 13 11:26:40 crc kubenswrapper[4632]: I0313 11:26:40.835210 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d989bc3bcc743f71b5d3bb9088655740c776263f3efde0646f3c95fa1bdc6620"} err="failed to get container status \"d989bc3bcc743f71b5d3bb9088655740c776263f3efde0646f3c95fa1bdc6620\": rpc error: code = NotFound desc = could not find container \"d989bc3bcc743f71b5d3bb9088655740c776263f3efde0646f3c95fa1bdc6620\": container with ID starting with d989bc3bcc743f71b5d3bb9088655740c776263f3efde0646f3c95fa1bdc6620 not found: ID does not exist" Mar 13 11:26:42 crc kubenswrapper[4632]: I0313 11:26:41.461905 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6nl2m"] Mar 13 11:26:42 crc kubenswrapper[4632]: I0313 11:26:41.713356 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6nl2m" event={"ID":"45f934fb-e679-4600-98e5-a67888251b13","Type":"ContainerStarted","Data":"4e3b7c19053eb4c84db186ea166af9584471f9afc53b60a314467a0d726485ee"} Mar 13 11:26:42 crc kubenswrapper[4632]: I0313 11:26:42.057690 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="743aacd5-2974-48e5-b2c8-ed448225edc3" path="/var/lib/kubelet/pods/743aacd5-2974-48e5-b2c8-ed448225edc3/volumes" Mar 13 11:26:42 crc kubenswrapper[4632]: I0313 11:26:42.723698 4632 generic.go:334] "Generic (PLEG): container finished" podID="45f934fb-e679-4600-98e5-a67888251b13" containerID="5c3a87ecc65399621b88755466dcd9fbdd959c39ae545371f0246e2e62cc5227" exitCode=0 Mar 13 11:26:42 crc kubenswrapper[4632]: I0313 11:26:42.724056 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6nl2m" event={"ID":"45f934fb-e679-4600-98e5-a67888251b13","Type":"ContainerDied","Data":"5c3a87ecc65399621b88755466dcd9fbdd959c39ae545371f0246e2e62cc5227"} Mar 13 11:26:44 crc kubenswrapper[4632]: I0313 11:26:44.742874 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6nl2m" event={"ID":"45f934fb-e679-4600-98e5-a67888251b13","Type":"ContainerStarted","Data":"6c54f8c37225b428e25e0b2eb3a551336c13c5eab39c71c35116c1eb9a44fab6"} Mar 13 11:26:46 crc kubenswrapper[4632]: I0313 11:26:46.762138 4632 generic.go:334] "Generic (PLEG): container finished" podID="45f934fb-e679-4600-98e5-a67888251b13" containerID="6c54f8c37225b428e25e0b2eb3a551336c13c5eab39c71c35116c1eb9a44fab6" exitCode=0 Mar 13 11:26:46 crc kubenswrapper[4632]: I0313 11:26:46.762207 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6nl2m" event={"ID":"45f934fb-e679-4600-98e5-a67888251b13","Type":"ContainerDied","Data":"6c54f8c37225b428e25e0b2eb3a551336c13c5eab39c71c35116c1eb9a44fab6"} Mar 13 11:26:47 crc kubenswrapper[4632]: I0313 11:26:47.773831 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6nl2m" event={"ID":"45f934fb-e679-4600-98e5-a67888251b13","Type":"ContainerStarted","Data":"1f3d3300925758d71546e25b20caa6d6766393cc646bf78998dc1871e1262799"} Mar 13 11:26:47 crc kubenswrapper[4632]: I0313 11:26:47.799897 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6nl2m" podStartSLOduration=3.330901611 podStartE2EDuration="7.799877975s" podCreationTimestamp="2026-03-13 11:26:40 +0000 UTC" firstStartedPulling="2026-03-13 11:26:42.725752978 +0000 UTC m=+4976.748283111" lastFinishedPulling="2026-03-13 11:26:47.194729332 +0000 UTC m=+4981.217259475" observedRunningTime="2026-03-13 11:26:47.789661213 +0000 UTC m=+4981.812191366" watchObservedRunningTime="2026-03-13 11:26:47.799877975 +0000 UTC m=+4981.822408108" Mar 13 11:26:50 crc kubenswrapper[4632]: I0313 11:26:50.680694 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:26:50 crc kubenswrapper[4632]: I0313 11:26:50.682284 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:26:51 crc kubenswrapper[4632]: I0313 11:26:51.766113 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-6nl2m" podUID="45f934fb-e679-4600-98e5-a67888251b13" containerName="registry-server" probeResult="failure" output=< Mar 13 11:26:51 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:26:51 crc kubenswrapper[4632]: > Mar 13 11:27:00 crc kubenswrapper[4632]: I0313 11:27:00.762069 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:27:00 crc kubenswrapper[4632]: I0313 11:27:00.832528 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:27:01 crc kubenswrapper[4632]: I0313 11:27:01.007389 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6nl2m"] Mar 13 11:27:01 crc kubenswrapper[4632]: I0313 11:27:01.912457 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6nl2m" podUID="45f934fb-e679-4600-98e5-a67888251b13" containerName="registry-server" containerID="cri-o://1f3d3300925758d71546e25b20caa6d6766393cc646bf78998dc1871e1262799" gracePeriod=2 Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.550722 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.699361 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2dn74\" (UniqueName: \"kubernetes.io/projected/45f934fb-e679-4600-98e5-a67888251b13-kube-api-access-2dn74\") pod \"45f934fb-e679-4600-98e5-a67888251b13\" (UID: \"45f934fb-e679-4600-98e5-a67888251b13\") " Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.699671 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45f934fb-e679-4600-98e5-a67888251b13-catalog-content\") pod \"45f934fb-e679-4600-98e5-a67888251b13\" (UID: \"45f934fb-e679-4600-98e5-a67888251b13\") " Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.700299 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45f934fb-e679-4600-98e5-a67888251b13-utilities\") pod \"45f934fb-e679-4600-98e5-a67888251b13\" (UID: \"45f934fb-e679-4600-98e5-a67888251b13\") " Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.701385 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45f934fb-e679-4600-98e5-a67888251b13-utilities" (OuterVolumeSpecName: "utilities") pod "45f934fb-e679-4600-98e5-a67888251b13" (UID: "45f934fb-e679-4600-98e5-a67888251b13"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.702308 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/45f934fb-e679-4600-98e5-a67888251b13-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.706993 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/45f934fb-e679-4600-98e5-a67888251b13-kube-api-access-2dn74" (OuterVolumeSpecName: "kube-api-access-2dn74") pod "45f934fb-e679-4600-98e5-a67888251b13" (UID: "45f934fb-e679-4600-98e5-a67888251b13"). InnerVolumeSpecName "kube-api-access-2dn74". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.764542 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/45f934fb-e679-4600-98e5-a67888251b13-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "45f934fb-e679-4600-98e5-a67888251b13" (UID: "45f934fb-e679-4600-98e5-a67888251b13"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.803712 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2dn74\" (UniqueName: \"kubernetes.io/projected/45f934fb-e679-4600-98e5-a67888251b13-kube-api-access-2dn74\") on node \"crc\" DevicePath \"\"" Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.803750 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/45f934fb-e679-4600-98e5-a67888251b13-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.936421 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6nl2m" Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.937074 4632 generic.go:334] "Generic (PLEG): container finished" podID="45f934fb-e679-4600-98e5-a67888251b13" containerID="1f3d3300925758d71546e25b20caa6d6766393cc646bf78998dc1871e1262799" exitCode=0 Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.937145 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6nl2m" event={"ID":"45f934fb-e679-4600-98e5-a67888251b13","Type":"ContainerDied","Data":"1f3d3300925758d71546e25b20caa6d6766393cc646bf78998dc1871e1262799"} Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.937187 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6nl2m" event={"ID":"45f934fb-e679-4600-98e5-a67888251b13","Type":"ContainerDied","Data":"4e3b7c19053eb4c84db186ea166af9584471f9afc53b60a314467a0d726485ee"} Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.937218 4632 scope.go:117] "RemoveContainer" containerID="1f3d3300925758d71546e25b20caa6d6766393cc646bf78998dc1871e1262799" Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.977857 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6nl2m"] Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.982463 4632 scope.go:117] "RemoveContainer" containerID="6c54f8c37225b428e25e0b2eb3a551336c13c5eab39c71c35116c1eb9a44fab6" Mar 13 11:27:02 crc kubenswrapper[4632]: I0313 11:27:02.987760 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6nl2m"] Mar 13 11:27:03 crc kubenswrapper[4632]: I0313 11:27:03.010960 4632 scope.go:117] "RemoveContainer" containerID="5c3a87ecc65399621b88755466dcd9fbdd959c39ae545371f0246e2e62cc5227" Mar 13 11:27:03 crc kubenswrapper[4632]: I0313 11:27:03.045130 4632 scope.go:117] "RemoveContainer" containerID="1f3d3300925758d71546e25b20caa6d6766393cc646bf78998dc1871e1262799" Mar 13 11:27:03 crc kubenswrapper[4632]: E0313 11:27:03.046242 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f3d3300925758d71546e25b20caa6d6766393cc646bf78998dc1871e1262799\": container with ID starting with 1f3d3300925758d71546e25b20caa6d6766393cc646bf78998dc1871e1262799 not found: ID does not exist" containerID="1f3d3300925758d71546e25b20caa6d6766393cc646bf78998dc1871e1262799" Mar 13 11:27:03 crc kubenswrapper[4632]: I0313 11:27:03.046370 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f3d3300925758d71546e25b20caa6d6766393cc646bf78998dc1871e1262799"} err="failed to get container status \"1f3d3300925758d71546e25b20caa6d6766393cc646bf78998dc1871e1262799\": rpc error: code = NotFound desc = could not find container \"1f3d3300925758d71546e25b20caa6d6766393cc646bf78998dc1871e1262799\": container with ID starting with 1f3d3300925758d71546e25b20caa6d6766393cc646bf78998dc1871e1262799 not found: ID does not exist" Mar 13 11:27:03 crc kubenswrapper[4632]: I0313 11:27:03.046452 4632 scope.go:117] "RemoveContainer" containerID="6c54f8c37225b428e25e0b2eb3a551336c13c5eab39c71c35116c1eb9a44fab6" Mar 13 11:27:03 crc kubenswrapper[4632]: E0313 11:27:03.046972 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c54f8c37225b428e25e0b2eb3a551336c13c5eab39c71c35116c1eb9a44fab6\": container with ID starting with 6c54f8c37225b428e25e0b2eb3a551336c13c5eab39c71c35116c1eb9a44fab6 not found: ID does not exist" containerID="6c54f8c37225b428e25e0b2eb3a551336c13c5eab39c71c35116c1eb9a44fab6" Mar 13 11:27:03 crc kubenswrapper[4632]: I0313 11:27:03.047071 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c54f8c37225b428e25e0b2eb3a551336c13c5eab39c71c35116c1eb9a44fab6"} err="failed to get container status \"6c54f8c37225b428e25e0b2eb3a551336c13c5eab39c71c35116c1eb9a44fab6\": rpc error: code = NotFound desc = could not find container \"6c54f8c37225b428e25e0b2eb3a551336c13c5eab39c71c35116c1eb9a44fab6\": container with ID starting with 6c54f8c37225b428e25e0b2eb3a551336c13c5eab39c71c35116c1eb9a44fab6 not found: ID does not exist" Mar 13 11:27:03 crc kubenswrapper[4632]: I0313 11:27:03.047140 4632 scope.go:117] "RemoveContainer" containerID="5c3a87ecc65399621b88755466dcd9fbdd959c39ae545371f0246e2e62cc5227" Mar 13 11:27:03 crc kubenswrapper[4632]: E0313 11:27:03.047402 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c3a87ecc65399621b88755466dcd9fbdd959c39ae545371f0246e2e62cc5227\": container with ID starting with 5c3a87ecc65399621b88755466dcd9fbdd959c39ae545371f0246e2e62cc5227 not found: ID does not exist" containerID="5c3a87ecc65399621b88755466dcd9fbdd959c39ae545371f0246e2e62cc5227" Mar 13 11:27:03 crc kubenswrapper[4632]: I0313 11:27:03.047484 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c3a87ecc65399621b88755466dcd9fbdd959c39ae545371f0246e2e62cc5227"} err="failed to get container status \"5c3a87ecc65399621b88755466dcd9fbdd959c39ae545371f0246e2e62cc5227\": rpc error: code = NotFound desc = could not find container \"5c3a87ecc65399621b88755466dcd9fbdd959c39ae545371f0246e2e62cc5227\": container with ID starting with 5c3a87ecc65399621b88755466dcd9fbdd959c39ae545371f0246e2e62cc5227 not found: ID does not exist" Mar 13 11:27:04 crc kubenswrapper[4632]: I0313 11:27:04.062347 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="45f934fb-e679-4600-98e5-a67888251b13" path="/var/lib/kubelet/pods/45f934fb-e679-4600-98e5-a67888251b13/volumes" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.509965 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zlblp"] Mar 13 11:27:28 crc kubenswrapper[4632]: E0313 11:27:28.510761 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45f934fb-e679-4600-98e5-a67888251b13" containerName="extract-content" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.510773 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="45f934fb-e679-4600-98e5-a67888251b13" containerName="extract-content" Mar 13 11:27:28 crc kubenswrapper[4632]: E0313 11:27:28.510785 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45f934fb-e679-4600-98e5-a67888251b13" containerName="extract-utilities" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.510792 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="45f934fb-e679-4600-98e5-a67888251b13" containerName="extract-utilities" Mar 13 11:27:28 crc kubenswrapper[4632]: E0313 11:27:28.510804 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="45f934fb-e679-4600-98e5-a67888251b13" containerName="registry-server" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.510810 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="45f934fb-e679-4600-98e5-a67888251b13" containerName="registry-server" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.511013 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="45f934fb-e679-4600-98e5-a67888251b13" containerName="registry-server" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.512277 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.541363 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zlblp"] Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.693655 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-catalog-content\") pod \"community-operators-zlblp\" (UID: \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\") " pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.693979 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-utilities\") pod \"community-operators-zlblp\" (UID: \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\") " pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.694089 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds2wt\" (UniqueName: \"kubernetes.io/projected/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-kube-api-access-ds2wt\") pod \"community-operators-zlblp\" (UID: \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\") " pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.795666 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-catalog-content\") pod \"community-operators-zlblp\" (UID: \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\") " pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.795837 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-utilities\") pod \"community-operators-zlblp\" (UID: \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\") " pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.795876 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds2wt\" (UniqueName: \"kubernetes.io/projected/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-kube-api-access-ds2wt\") pod \"community-operators-zlblp\" (UID: \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\") " pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.796246 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-catalog-content\") pod \"community-operators-zlblp\" (UID: \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\") " pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.796663 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-utilities\") pod \"community-operators-zlblp\" (UID: \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\") " pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.831318 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds2wt\" (UniqueName: \"kubernetes.io/projected/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-kube-api-access-ds2wt\") pod \"community-operators-zlblp\" (UID: \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\") " pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:28 crc kubenswrapper[4632]: I0313 11:27:28.900528 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:29 crc kubenswrapper[4632]: I0313 11:27:29.389274 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zlblp"] Mar 13 11:27:30 crc kubenswrapper[4632]: I0313 11:27:30.255608 4632 generic.go:334] "Generic (PLEG): container finished" podID="7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" containerID="588578b93ec0d7a4ab9503a78edfe460fc82f8b613c54d0795b8a8243a44b801" exitCode=0 Mar 13 11:27:30 crc kubenswrapper[4632]: I0313 11:27:30.255667 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlblp" event={"ID":"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7","Type":"ContainerDied","Data":"588578b93ec0d7a4ab9503a78edfe460fc82f8b613c54d0795b8a8243a44b801"} Mar 13 11:27:30 crc kubenswrapper[4632]: I0313 11:27:30.255731 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlblp" event={"ID":"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7","Type":"ContainerStarted","Data":"f9576dbf1fd0b8a99204ae893036fa6519be832d4a433a5ae4a842e9e21ad975"} Mar 13 11:27:31 crc kubenswrapper[4632]: I0313 11:27:31.270611 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlblp" event={"ID":"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7","Type":"ContainerStarted","Data":"ddd93ebe954f83eabb9f3a51aec00226efd4801cb41ff33b60684d2fb0414f63"} Mar 13 11:27:33 crc kubenswrapper[4632]: I0313 11:27:33.293746 4632 generic.go:334] "Generic (PLEG): container finished" podID="7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" containerID="ddd93ebe954f83eabb9f3a51aec00226efd4801cb41ff33b60684d2fb0414f63" exitCode=0 Mar 13 11:27:33 crc kubenswrapper[4632]: I0313 11:27:33.293833 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlblp" event={"ID":"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7","Type":"ContainerDied","Data":"ddd93ebe954f83eabb9f3a51aec00226efd4801cb41ff33b60684d2fb0414f63"} Mar 13 11:27:35 crc kubenswrapper[4632]: I0313 11:27:35.317260 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlblp" event={"ID":"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7","Type":"ContainerStarted","Data":"231d214dea21217dfb7e7d55a896c931c65d43b0af2f081bf87a1d9116d05023"} Mar 13 11:27:35 crc kubenswrapper[4632]: I0313 11:27:35.354435 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zlblp" podStartSLOduration=3.741217878 podStartE2EDuration="7.354410168s" podCreationTimestamp="2026-03-13 11:27:28 +0000 UTC" firstStartedPulling="2026-03-13 11:27:30.259033727 +0000 UTC m=+5024.281563860" lastFinishedPulling="2026-03-13 11:27:33.872226017 +0000 UTC m=+5027.894756150" observedRunningTime="2026-03-13 11:27:35.343819046 +0000 UTC m=+5029.366349189" watchObservedRunningTime="2026-03-13 11:27:35.354410168 +0000 UTC m=+5029.376940311" Mar 13 11:27:38 crc kubenswrapper[4632]: I0313 11:27:38.900891 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:38 crc kubenswrapper[4632]: I0313 11:27:38.901325 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:39 crc kubenswrapper[4632]: I0313 11:27:39.953533 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-zlblp" podUID="7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" containerName="registry-server" probeResult="failure" output=< Mar 13 11:27:39 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:27:39 crc kubenswrapper[4632]: > Mar 13 11:27:40 crc kubenswrapper[4632]: I0313 11:27:40.461001 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:27:40 crc kubenswrapper[4632]: I0313 11:27:40.461134 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:27:48 crc kubenswrapper[4632]: I0313 11:27:48.955232 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:49 crc kubenswrapper[4632]: I0313 11:27:49.010105 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:49 crc kubenswrapper[4632]: I0313 11:27:49.260973 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zlblp"] Mar 13 11:27:50 crc kubenswrapper[4632]: I0313 11:27:50.480209 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zlblp" podUID="7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" containerName="registry-server" containerID="cri-o://231d214dea21217dfb7e7d55a896c931c65d43b0af2f081bf87a1d9116d05023" gracePeriod=2 Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.010213 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.025097 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ds2wt\" (UniqueName: \"kubernetes.io/projected/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-kube-api-access-ds2wt\") pod \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\" (UID: \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\") " Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.025406 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-utilities\") pod \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\" (UID: \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\") " Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.025847 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-catalog-content\") pod \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\" (UID: \"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7\") " Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.031201 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-kube-api-access-ds2wt" (OuterVolumeSpecName: "kube-api-access-ds2wt") pod "7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" (UID: "7f615080-fa27-4ea2-ab8b-bd6409bdcaf7"). InnerVolumeSpecName "kube-api-access-ds2wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.031586 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-utilities" (OuterVolumeSpecName: "utilities") pod "7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" (UID: "7f615080-fa27-4ea2-ab8b-bd6409bdcaf7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.106801 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" (UID: "7f615080-fa27-4ea2-ab8b-bd6409bdcaf7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.128665 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.128727 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ds2wt\" (UniqueName: \"kubernetes.io/projected/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-kube-api-access-ds2wt\") on node \"crc\" DevicePath \"\"" Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.128739 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.492584 4632 generic.go:334] "Generic (PLEG): container finished" podID="7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" containerID="231d214dea21217dfb7e7d55a896c931c65d43b0af2f081bf87a1d9116d05023" exitCode=0 Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.492669 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlblp" event={"ID":"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7","Type":"ContainerDied","Data":"231d214dea21217dfb7e7d55a896c931c65d43b0af2f081bf87a1d9116d05023"} Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.492761 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zlblp" event={"ID":"7f615080-fa27-4ea2-ab8b-bd6409bdcaf7","Type":"ContainerDied","Data":"f9576dbf1fd0b8a99204ae893036fa6519be832d4a433a5ae4a842e9e21ad975"} Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.492807 4632 scope.go:117] "RemoveContainer" containerID="231d214dea21217dfb7e7d55a896c931c65d43b0af2f081bf87a1d9116d05023" Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.493180 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zlblp" Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.518994 4632 scope.go:117] "RemoveContainer" containerID="ddd93ebe954f83eabb9f3a51aec00226efd4801cb41ff33b60684d2fb0414f63" Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.539384 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zlblp"] Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.551823 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zlblp"] Mar 13 11:27:51 crc kubenswrapper[4632]: I0313 11:27:51.979862 4632 scope.go:117] "RemoveContainer" containerID="588578b93ec0d7a4ab9503a78edfe460fc82f8b613c54d0795b8a8243a44b801" Mar 13 11:27:52 crc kubenswrapper[4632]: I0313 11:27:52.064027 4632 scope.go:117] "RemoveContainer" containerID="231d214dea21217dfb7e7d55a896c931c65d43b0af2f081bf87a1d9116d05023" Mar 13 11:27:52 crc kubenswrapper[4632]: E0313 11:27:52.064554 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"231d214dea21217dfb7e7d55a896c931c65d43b0af2f081bf87a1d9116d05023\": container with ID starting with 231d214dea21217dfb7e7d55a896c931c65d43b0af2f081bf87a1d9116d05023 not found: ID does not exist" containerID="231d214dea21217dfb7e7d55a896c931c65d43b0af2f081bf87a1d9116d05023" Mar 13 11:27:52 crc kubenswrapper[4632]: I0313 11:27:52.064599 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"231d214dea21217dfb7e7d55a896c931c65d43b0af2f081bf87a1d9116d05023"} err="failed to get container status \"231d214dea21217dfb7e7d55a896c931c65d43b0af2f081bf87a1d9116d05023\": rpc error: code = NotFound desc = could not find container \"231d214dea21217dfb7e7d55a896c931c65d43b0af2f081bf87a1d9116d05023\": container with ID starting with 231d214dea21217dfb7e7d55a896c931c65d43b0af2f081bf87a1d9116d05023 not found: ID does not exist" Mar 13 11:27:52 crc kubenswrapper[4632]: I0313 11:27:52.064622 4632 scope.go:117] "RemoveContainer" containerID="ddd93ebe954f83eabb9f3a51aec00226efd4801cb41ff33b60684d2fb0414f63" Mar 13 11:27:52 crc kubenswrapper[4632]: I0313 11:27:52.065008 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" path="/var/lib/kubelet/pods/7f615080-fa27-4ea2-ab8b-bd6409bdcaf7/volumes" Mar 13 11:27:52 crc kubenswrapper[4632]: E0313 11:27:52.065075 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ddd93ebe954f83eabb9f3a51aec00226efd4801cb41ff33b60684d2fb0414f63\": container with ID starting with ddd93ebe954f83eabb9f3a51aec00226efd4801cb41ff33b60684d2fb0414f63 not found: ID does not exist" containerID="ddd93ebe954f83eabb9f3a51aec00226efd4801cb41ff33b60684d2fb0414f63" Mar 13 11:27:52 crc kubenswrapper[4632]: I0313 11:27:52.065111 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ddd93ebe954f83eabb9f3a51aec00226efd4801cb41ff33b60684d2fb0414f63"} err="failed to get container status \"ddd93ebe954f83eabb9f3a51aec00226efd4801cb41ff33b60684d2fb0414f63\": rpc error: code = NotFound desc = could not find container \"ddd93ebe954f83eabb9f3a51aec00226efd4801cb41ff33b60684d2fb0414f63\": container with ID starting with ddd93ebe954f83eabb9f3a51aec00226efd4801cb41ff33b60684d2fb0414f63 not found: ID does not exist" Mar 13 11:27:52 crc kubenswrapper[4632]: I0313 11:27:52.065136 4632 scope.go:117] "RemoveContainer" containerID="588578b93ec0d7a4ab9503a78edfe460fc82f8b613c54d0795b8a8243a44b801" Mar 13 11:27:52 crc kubenswrapper[4632]: E0313 11:27:52.065613 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"588578b93ec0d7a4ab9503a78edfe460fc82f8b613c54d0795b8a8243a44b801\": container with ID starting with 588578b93ec0d7a4ab9503a78edfe460fc82f8b613c54d0795b8a8243a44b801 not found: ID does not exist" containerID="588578b93ec0d7a4ab9503a78edfe460fc82f8b613c54d0795b8a8243a44b801" Mar 13 11:27:52 crc kubenswrapper[4632]: I0313 11:27:52.065649 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"588578b93ec0d7a4ab9503a78edfe460fc82f8b613c54d0795b8a8243a44b801"} err="failed to get container status \"588578b93ec0d7a4ab9503a78edfe460fc82f8b613c54d0795b8a8243a44b801\": rpc error: code = NotFound desc = could not find container \"588578b93ec0d7a4ab9503a78edfe460fc82f8b613c54d0795b8a8243a44b801\": container with ID starting with 588578b93ec0d7a4ab9503a78edfe460fc82f8b613c54d0795b8a8243a44b801 not found: ID does not exist" Mar 13 11:28:00 crc kubenswrapper[4632]: I0313 11:28:00.208167 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556688-btp9b"] Mar 13 11:28:00 crc kubenswrapper[4632]: E0313 11:28:00.209216 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" containerName="registry-server" Mar 13 11:28:00 crc kubenswrapper[4632]: I0313 11:28:00.209238 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" containerName="registry-server" Mar 13 11:28:00 crc kubenswrapper[4632]: E0313 11:28:00.209280 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" containerName="extract-utilities" Mar 13 11:28:00 crc kubenswrapper[4632]: I0313 11:28:00.209290 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" containerName="extract-utilities" Mar 13 11:28:00 crc kubenswrapper[4632]: E0313 11:28:00.209307 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" containerName="extract-content" Mar 13 11:28:00 crc kubenswrapper[4632]: I0313 11:28:00.209317 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" containerName="extract-content" Mar 13 11:28:00 crc kubenswrapper[4632]: I0313 11:28:00.209596 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f615080-fa27-4ea2-ab8b-bd6409bdcaf7" containerName="registry-server" Mar 13 11:28:00 crc kubenswrapper[4632]: I0313 11:28:00.210394 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556688-btp9b" Mar 13 11:28:00 crc kubenswrapper[4632]: I0313 11:28:00.218427 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:28:00 crc kubenswrapper[4632]: I0313 11:28:00.218668 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:28:00 crc kubenswrapper[4632]: I0313 11:28:00.218818 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:28:00 crc kubenswrapper[4632]: I0313 11:28:00.236649 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556688-btp9b"] Mar 13 11:28:00 crc kubenswrapper[4632]: I0313 11:28:00.311969 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwnbq\" (UniqueName: \"kubernetes.io/projected/99bcc88d-9858-4d4f-97e5-68e185f06401-kube-api-access-xwnbq\") pod \"auto-csr-approver-29556688-btp9b\" (UID: \"99bcc88d-9858-4d4f-97e5-68e185f06401\") " pod="openshift-infra/auto-csr-approver-29556688-btp9b" Mar 13 11:28:00 crc kubenswrapper[4632]: I0313 11:28:00.414385 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwnbq\" (UniqueName: \"kubernetes.io/projected/99bcc88d-9858-4d4f-97e5-68e185f06401-kube-api-access-xwnbq\") pod \"auto-csr-approver-29556688-btp9b\" (UID: \"99bcc88d-9858-4d4f-97e5-68e185f06401\") " pod="openshift-infra/auto-csr-approver-29556688-btp9b" Mar 13 11:28:00 crc kubenswrapper[4632]: I0313 11:28:00.450806 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwnbq\" (UniqueName: \"kubernetes.io/projected/99bcc88d-9858-4d4f-97e5-68e185f06401-kube-api-access-xwnbq\") pod \"auto-csr-approver-29556688-btp9b\" (UID: \"99bcc88d-9858-4d4f-97e5-68e185f06401\") " pod="openshift-infra/auto-csr-approver-29556688-btp9b" Mar 13 11:28:00 crc kubenswrapper[4632]: I0313 11:28:00.538788 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556688-btp9b" Mar 13 11:28:01 crc kubenswrapper[4632]: I0313 11:28:01.030304 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556688-btp9b"] Mar 13 11:28:01 crc kubenswrapper[4632]: I0313 11:28:01.607184 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556688-btp9b" event={"ID":"99bcc88d-9858-4d4f-97e5-68e185f06401","Type":"ContainerStarted","Data":"8530e4258bcde86671610ca0987044da96d15fd754611f558323ff659137e25a"} Mar 13 11:28:03 crc kubenswrapper[4632]: I0313 11:28:03.637410 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556688-btp9b" event={"ID":"99bcc88d-9858-4d4f-97e5-68e185f06401","Type":"ContainerStarted","Data":"419020d6577499249f09db214d08f7500440b2163407ca9b721f361af1ab72f7"} Mar 13 11:28:03 crc kubenswrapper[4632]: I0313 11:28:03.660222 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556688-btp9b" podStartSLOduration=2.405580217 podStartE2EDuration="3.660198468s" podCreationTimestamp="2026-03-13 11:28:00 +0000 UTC" firstStartedPulling="2026-03-13 11:28:01.036776138 +0000 UTC m=+5055.059306271" lastFinishedPulling="2026-03-13 11:28:02.291394369 +0000 UTC m=+5056.313924522" observedRunningTime="2026-03-13 11:28:03.657872702 +0000 UTC m=+5057.680402845" watchObservedRunningTime="2026-03-13 11:28:03.660198468 +0000 UTC m=+5057.682728611" Mar 13 11:28:04 crc kubenswrapper[4632]: I0313 11:28:04.650090 4632 generic.go:334] "Generic (PLEG): container finished" podID="99bcc88d-9858-4d4f-97e5-68e185f06401" containerID="419020d6577499249f09db214d08f7500440b2163407ca9b721f361af1ab72f7" exitCode=0 Mar 13 11:28:04 crc kubenswrapper[4632]: I0313 11:28:04.650215 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556688-btp9b" event={"ID":"99bcc88d-9858-4d4f-97e5-68e185f06401","Type":"ContainerDied","Data":"419020d6577499249f09db214d08f7500440b2163407ca9b721f361af1ab72f7"} Mar 13 11:28:06 crc kubenswrapper[4632]: I0313 11:28:06.032103 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556688-btp9b" Mar 13 11:28:06 crc kubenswrapper[4632]: I0313 11:28:06.132301 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xwnbq\" (UniqueName: \"kubernetes.io/projected/99bcc88d-9858-4d4f-97e5-68e185f06401-kube-api-access-xwnbq\") pod \"99bcc88d-9858-4d4f-97e5-68e185f06401\" (UID: \"99bcc88d-9858-4d4f-97e5-68e185f06401\") " Mar 13 11:28:06 crc kubenswrapper[4632]: I0313 11:28:06.155458 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99bcc88d-9858-4d4f-97e5-68e185f06401-kube-api-access-xwnbq" (OuterVolumeSpecName: "kube-api-access-xwnbq") pod "99bcc88d-9858-4d4f-97e5-68e185f06401" (UID: "99bcc88d-9858-4d4f-97e5-68e185f06401"). InnerVolumeSpecName "kube-api-access-xwnbq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:28:06 crc kubenswrapper[4632]: I0313 11:28:06.234038 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xwnbq\" (UniqueName: \"kubernetes.io/projected/99bcc88d-9858-4d4f-97e5-68e185f06401-kube-api-access-xwnbq\") on node \"crc\" DevicePath \"\"" Mar 13 11:28:06 crc kubenswrapper[4632]: I0313 11:28:06.671670 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556688-btp9b" event={"ID":"99bcc88d-9858-4d4f-97e5-68e185f06401","Type":"ContainerDied","Data":"8530e4258bcde86671610ca0987044da96d15fd754611f558323ff659137e25a"} Mar 13 11:28:06 crc kubenswrapper[4632]: I0313 11:28:06.671722 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8530e4258bcde86671610ca0987044da96d15fd754611f558323ff659137e25a" Mar 13 11:28:06 crc kubenswrapper[4632]: I0313 11:28:06.671782 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556688-btp9b" Mar 13 11:28:06 crc kubenswrapper[4632]: E0313 11:28:06.796819 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99bcc88d_9858_4d4f_97e5_68e185f06401.slice/crio-8530e4258bcde86671610ca0987044da96d15fd754611f558323ff659137e25a\": RecentStats: unable to find data in memory cache]" Mar 13 11:28:07 crc kubenswrapper[4632]: I0313 11:28:07.130084 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556682-pm7lm"] Mar 13 11:28:07 crc kubenswrapper[4632]: I0313 11:28:07.138401 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556682-pm7lm"] Mar 13 11:28:08 crc kubenswrapper[4632]: I0313 11:28:08.065997 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="191fb79a-448d-4181-8346-f9dec8721d81" path="/var/lib/kubelet/pods/191fb79a-448d-4181-8346-f9dec8721d81/volumes" Mar 13 11:28:10 crc kubenswrapper[4632]: I0313 11:28:10.460722 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:28:10 crc kubenswrapper[4632]: I0313 11:28:10.461059 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:28:19 crc kubenswrapper[4632]: I0313 11:28:19.843254 4632 scope.go:117] "RemoveContainer" containerID="6b6b471905ed6fd6c16476a8c40a8d65b889486b4d10fadb0c4b7b6cf7a150be" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.458922 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l9hln"] Mar 13 11:28:40 crc kubenswrapper[4632]: E0313 11:28:40.459688 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99bcc88d-9858-4d4f-97e5-68e185f06401" containerName="oc" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.459700 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="99bcc88d-9858-4d4f-97e5-68e185f06401" containerName="oc" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.459904 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="99bcc88d-9858-4d4f-97e5-68e185f06401" containerName="oc" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.461983 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.462037 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.463793 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.464547 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.464611 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" gracePeriod=600 Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.464721 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.482960 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l9hln"] Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.508702 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddpsp\" (UniqueName: \"kubernetes.io/projected/e855a265-8de4-49c0-b910-ff29ae34b9c9-kube-api-access-ddpsp\") pod \"redhat-operators-l9hln\" (UID: \"e855a265-8de4-49c0-b910-ff29ae34b9c9\") " pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.508837 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e855a265-8de4-49c0-b910-ff29ae34b9c9-utilities\") pod \"redhat-operators-l9hln\" (UID: \"e855a265-8de4-49c0-b910-ff29ae34b9c9\") " pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.508883 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e855a265-8de4-49c0-b910-ff29ae34b9c9-catalog-content\") pod \"redhat-operators-l9hln\" (UID: \"e855a265-8de4-49c0-b910-ff29ae34b9c9\") " pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:28:40 crc kubenswrapper[4632]: E0313 11:28:40.590905 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.611216 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e855a265-8de4-49c0-b910-ff29ae34b9c9-utilities\") pod \"redhat-operators-l9hln\" (UID: \"e855a265-8de4-49c0-b910-ff29ae34b9c9\") " pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.611282 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e855a265-8de4-49c0-b910-ff29ae34b9c9-catalog-content\") pod \"redhat-operators-l9hln\" (UID: \"e855a265-8de4-49c0-b910-ff29ae34b9c9\") " pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.611419 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddpsp\" (UniqueName: \"kubernetes.io/projected/e855a265-8de4-49c0-b910-ff29ae34b9c9-kube-api-access-ddpsp\") pod \"redhat-operators-l9hln\" (UID: \"e855a265-8de4-49c0-b910-ff29ae34b9c9\") " pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.611827 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e855a265-8de4-49c0-b910-ff29ae34b9c9-utilities\") pod \"redhat-operators-l9hln\" (UID: \"e855a265-8de4-49c0-b910-ff29ae34b9c9\") " pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.611912 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e855a265-8de4-49c0-b910-ff29ae34b9c9-catalog-content\") pod \"redhat-operators-l9hln\" (UID: \"e855a265-8de4-49c0-b910-ff29ae34b9c9\") " pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.635576 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddpsp\" (UniqueName: \"kubernetes.io/projected/e855a265-8de4-49c0-b910-ff29ae34b9c9-kube-api-access-ddpsp\") pod \"redhat-operators-l9hln\" (UID: \"e855a265-8de4-49c0-b910-ff29ae34b9c9\") " pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:28:40 crc kubenswrapper[4632]: I0313 11:28:40.788929 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:28:41 crc kubenswrapper[4632]: I0313 11:28:41.019875 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" exitCode=0 Mar 13 11:28:41 crc kubenswrapper[4632]: I0313 11:28:41.020269 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3"} Mar 13 11:28:41 crc kubenswrapper[4632]: I0313 11:28:41.020304 4632 scope.go:117] "RemoveContainer" containerID="49682358d72adf3dbebb4a70c4dbc847548d4046ae5ef96f55f2ae4dfd58b9f9" Mar 13 11:28:41 crc kubenswrapper[4632]: I0313 11:28:41.020732 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:28:41 crc kubenswrapper[4632]: E0313 11:28:41.020975 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:28:41 crc kubenswrapper[4632]: I0313 11:28:41.353448 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l9hln"] Mar 13 11:28:41 crc kubenswrapper[4632]: W0313 11:28:41.765656 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode855a265_8de4_49c0_b910_ff29ae34b9c9.slice/crio-b8a9e04b6a98018edb581e5ef7c81695b29ef6091ce835092c2938e31a0a2e11 WatchSource:0}: Error finding container b8a9e04b6a98018edb581e5ef7c81695b29ef6091ce835092c2938e31a0a2e11: Status 404 returned error can't find the container with id b8a9e04b6a98018edb581e5ef7c81695b29ef6091ce835092c2938e31a0a2e11 Mar 13 11:28:42 crc kubenswrapper[4632]: I0313 11:28:42.032553 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9hln" event={"ID":"e855a265-8de4-49c0-b910-ff29ae34b9c9","Type":"ContainerStarted","Data":"867f95f5d2082d718d8174c12306f380c336af715d4c9e98f5b40dbfb57626a7"} Mar 13 11:28:42 crc kubenswrapper[4632]: I0313 11:28:42.032605 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9hln" event={"ID":"e855a265-8de4-49c0-b910-ff29ae34b9c9","Type":"ContainerStarted","Data":"b8a9e04b6a98018edb581e5ef7c81695b29ef6091ce835092c2938e31a0a2e11"} Mar 13 11:28:43 crc kubenswrapper[4632]: I0313 11:28:43.048748 4632 generic.go:334] "Generic (PLEG): container finished" podID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerID="867f95f5d2082d718d8174c12306f380c336af715d4c9e98f5b40dbfb57626a7" exitCode=0 Mar 13 11:28:43 crc kubenswrapper[4632]: I0313 11:28:43.049023 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9hln" event={"ID":"e855a265-8de4-49c0-b910-ff29ae34b9c9","Type":"ContainerDied","Data":"867f95f5d2082d718d8174c12306f380c336af715d4c9e98f5b40dbfb57626a7"} Mar 13 11:28:44 crc kubenswrapper[4632]: I0313 11:28:44.068512 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9hln" event={"ID":"e855a265-8de4-49c0-b910-ff29ae34b9c9","Type":"ContainerStarted","Data":"0d7e36ef30c33fd9204b409ec56e01a15539fd48854a29ce545aac9958b99985"} Mar 13 11:28:51 crc kubenswrapper[4632]: I0313 11:28:51.148415 4632 generic.go:334] "Generic (PLEG): container finished" podID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerID="0d7e36ef30c33fd9204b409ec56e01a15539fd48854a29ce545aac9958b99985" exitCode=0 Mar 13 11:28:51 crc kubenswrapper[4632]: I0313 11:28:51.148489 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9hln" event={"ID":"e855a265-8de4-49c0-b910-ff29ae34b9c9","Type":"ContainerDied","Data":"0d7e36ef30c33fd9204b409ec56e01a15539fd48854a29ce545aac9958b99985"} Mar 13 11:28:53 crc kubenswrapper[4632]: I0313 11:28:53.045245 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:28:53 crc kubenswrapper[4632]: E0313 11:28:53.045776 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:28:53 crc kubenswrapper[4632]: I0313 11:28:53.174139 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9hln" event={"ID":"e855a265-8de4-49c0-b910-ff29ae34b9c9","Type":"ContainerStarted","Data":"1507c5b681588f963d5f2f4b7375480788c56bfd5b05e62a793bca31530d4e2f"} Mar 13 11:28:53 crc kubenswrapper[4632]: I0313 11:28:53.200578 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l9hln" podStartSLOduration=4.627011124 podStartE2EDuration="13.200547102s" podCreationTimestamp="2026-03-13 11:28:40 +0000 UTC" firstStartedPulling="2026-03-13 11:28:43.05072313 +0000 UTC m=+5097.073253263" lastFinishedPulling="2026-03-13 11:28:51.624259098 +0000 UTC m=+5105.646789241" observedRunningTime="2026-03-13 11:28:53.196991834 +0000 UTC m=+5107.219521967" watchObservedRunningTime="2026-03-13 11:28:53.200547102 +0000 UTC m=+5107.223077235" Mar 13 11:29:00 crc kubenswrapper[4632]: I0313 11:29:00.789651 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:29:00 crc kubenswrapper[4632]: I0313 11:29:00.790219 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:29:01 crc kubenswrapper[4632]: I0313 11:29:01.835450 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l9hln" podUID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerName="registry-server" probeResult="failure" output=< Mar 13 11:29:01 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:29:01 crc kubenswrapper[4632]: > Mar 13 11:29:06 crc kubenswrapper[4632]: I0313 11:29:06.044274 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:29:06 crc kubenswrapper[4632]: E0313 11:29:06.045160 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:29:11 crc kubenswrapper[4632]: I0313 11:29:11.851029 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l9hln" podUID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerName="registry-server" probeResult="failure" output=< Mar 13 11:29:11 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:29:11 crc kubenswrapper[4632]: > Mar 13 11:29:21 crc kubenswrapper[4632]: I0313 11:29:21.044537 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:29:21 crc kubenswrapper[4632]: E0313 11:29:21.045341 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:29:21 crc kubenswrapper[4632]: I0313 11:29:21.852051 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l9hln" podUID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerName="registry-server" probeResult="failure" output=< Mar 13 11:29:21 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:29:21 crc kubenswrapper[4632]: > Mar 13 11:29:31 crc kubenswrapper[4632]: I0313 11:29:31.870436 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l9hln" podUID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerName="registry-server" probeResult="failure" output=< Mar 13 11:29:31 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:29:31 crc kubenswrapper[4632]: > Mar 13 11:29:33 crc kubenswrapper[4632]: I0313 11:29:33.044266 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:29:33 crc kubenswrapper[4632]: E0313 11:29:33.044602 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:29:40 crc kubenswrapper[4632]: I0313 11:29:40.837039 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:29:40 crc kubenswrapper[4632]: I0313 11:29:40.906755 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:29:41 crc kubenswrapper[4632]: I0313 11:29:41.682826 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l9hln"] Mar 13 11:29:42 crc kubenswrapper[4632]: I0313 11:29:42.657234 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l9hln" podUID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerName="registry-server" containerID="cri-o://1507c5b681588f963d5f2f4b7375480788c56bfd5b05e62a793bca31530d4e2f" gracePeriod=2 Mar 13 11:29:43 crc kubenswrapper[4632]: I0313 11:29:43.674029 4632 generic.go:334] "Generic (PLEG): container finished" podID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerID="1507c5b681588f963d5f2f4b7375480788c56bfd5b05e62a793bca31530d4e2f" exitCode=0 Mar 13 11:29:43 crc kubenswrapper[4632]: I0313 11:29:43.674253 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9hln" event={"ID":"e855a265-8de4-49c0-b910-ff29ae34b9c9","Type":"ContainerDied","Data":"1507c5b681588f963d5f2f4b7375480788c56bfd5b05e62a793bca31530d4e2f"} Mar 13 11:29:43 crc kubenswrapper[4632]: I0313 11:29:43.985475 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.130026 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e855a265-8de4-49c0-b910-ff29ae34b9c9-utilities\") pod \"e855a265-8de4-49c0-b910-ff29ae34b9c9\" (UID: \"e855a265-8de4-49c0-b910-ff29ae34b9c9\") " Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.130136 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddpsp\" (UniqueName: \"kubernetes.io/projected/e855a265-8de4-49c0-b910-ff29ae34b9c9-kube-api-access-ddpsp\") pod \"e855a265-8de4-49c0-b910-ff29ae34b9c9\" (UID: \"e855a265-8de4-49c0-b910-ff29ae34b9c9\") " Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.130340 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e855a265-8de4-49c0-b910-ff29ae34b9c9-catalog-content\") pod \"e855a265-8de4-49c0-b910-ff29ae34b9c9\" (UID: \"e855a265-8de4-49c0-b910-ff29ae34b9c9\") " Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.133044 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e855a265-8de4-49c0-b910-ff29ae34b9c9-utilities" (OuterVolumeSpecName: "utilities") pod "e855a265-8de4-49c0-b910-ff29ae34b9c9" (UID: "e855a265-8de4-49c0-b910-ff29ae34b9c9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.221523 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e855a265-8de4-49c0-b910-ff29ae34b9c9-kube-api-access-ddpsp" (OuterVolumeSpecName: "kube-api-access-ddpsp") pod "e855a265-8de4-49c0-b910-ff29ae34b9c9" (UID: "e855a265-8de4-49c0-b910-ff29ae34b9c9"). InnerVolumeSpecName "kube-api-access-ddpsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.237872 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e855a265-8de4-49c0-b910-ff29ae34b9c9-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.238112 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddpsp\" (UniqueName: \"kubernetes.io/projected/e855a265-8de4-49c0-b910-ff29ae34b9c9-kube-api-access-ddpsp\") on node \"crc\" DevicePath \"\"" Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.318080 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e855a265-8de4-49c0-b910-ff29ae34b9c9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e855a265-8de4-49c0-b910-ff29ae34b9c9" (UID: "e855a265-8de4-49c0-b910-ff29ae34b9c9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.340117 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e855a265-8de4-49c0-b910-ff29ae34b9c9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.687407 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l9hln" event={"ID":"e855a265-8de4-49c0-b910-ff29ae34b9c9","Type":"ContainerDied","Data":"b8a9e04b6a98018edb581e5ef7c81695b29ef6091ce835092c2938e31a0a2e11"} Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.687451 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l9hln" Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.687474 4632 scope.go:117] "RemoveContainer" containerID="1507c5b681588f963d5f2f4b7375480788c56bfd5b05e62a793bca31530d4e2f" Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.720093 4632 scope.go:117] "RemoveContainer" containerID="0d7e36ef30c33fd9204b409ec56e01a15539fd48854a29ce545aac9958b99985" Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.733168 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l9hln"] Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.754054 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l9hln"] Mar 13 11:29:44 crc kubenswrapper[4632]: I0313 11:29:44.760070 4632 scope.go:117] "RemoveContainer" containerID="867f95f5d2082d718d8174c12306f380c336af715d4c9e98f5b40dbfb57626a7" Mar 13 11:29:46 crc kubenswrapper[4632]: I0313 11:29:46.113493 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e855a265-8de4-49c0-b910-ff29ae34b9c9" path="/var/lib/kubelet/pods/e855a265-8de4-49c0-b910-ff29ae34b9c9/volumes" Mar 13 11:29:47 crc kubenswrapper[4632]: I0313 11:29:47.044274 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:29:47 crc kubenswrapper[4632]: E0313 11:29:47.045055 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:29:58 crc kubenswrapper[4632]: I0313 11:29:58.051514 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:29:58 crc kubenswrapper[4632]: E0313 11:29:58.052851 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.191295 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556690-d6xbb"] Mar 13 11:30:00 crc kubenswrapper[4632]: E0313 11:30:00.193782 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerName="extract-utilities" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.193814 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerName="extract-utilities" Mar 13 11:30:00 crc kubenswrapper[4632]: E0313 11:30:00.193837 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerName="registry-server" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.193844 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerName="registry-server" Mar 13 11:30:00 crc kubenswrapper[4632]: E0313 11:30:00.193873 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerName="extract-content" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.193880 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerName="extract-content" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.194294 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="e855a265-8de4-49c0-b910-ff29ae34b9c9" containerName="registry-server" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.202881 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh"] Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.203240 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556690-d6xbb" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.205190 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.211510 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.211675 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.211799 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.216976 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556690-d6xbb"] Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.225339 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.245735 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh"] Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.249060 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.304040 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dc72a85-cdb5-4b11-9e0a-158d269edf96-secret-volume\") pod \"collect-profiles-29556690-p82jh\" (UID: \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.304131 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jclp2\" (UniqueName: \"kubernetes.io/projected/9dc72a85-cdb5-4b11-9e0a-158d269edf96-kube-api-access-jclp2\") pod \"collect-profiles-29556690-p82jh\" (UID: \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.304165 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqksf\" (UniqueName: \"kubernetes.io/projected/d7194752-8651-4e1b-8973-1f821bf23755-kube-api-access-cqksf\") pod \"auto-csr-approver-29556690-d6xbb\" (UID: \"d7194752-8651-4e1b-8973-1f821bf23755\") " pod="openshift-infra/auto-csr-approver-29556690-d6xbb" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.304207 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dc72a85-cdb5-4b11-9e0a-158d269edf96-config-volume\") pod \"collect-profiles-29556690-p82jh\" (UID: \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.406150 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqksf\" (UniqueName: \"kubernetes.io/projected/d7194752-8651-4e1b-8973-1f821bf23755-kube-api-access-cqksf\") pod \"auto-csr-approver-29556690-d6xbb\" (UID: \"d7194752-8651-4e1b-8973-1f821bf23755\") " pod="openshift-infra/auto-csr-approver-29556690-d6xbb" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.406286 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dc72a85-cdb5-4b11-9e0a-158d269edf96-config-volume\") pod \"collect-profiles-29556690-p82jh\" (UID: \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.406410 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dc72a85-cdb5-4b11-9e0a-158d269edf96-secret-volume\") pod \"collect-profiles-29556690-p82jh\" (UID: \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.406497 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jclp2\" (UniqueName: \"kubernetes.io/projected/9dc72a85-cdb5-4b11-9e0a-158d269edf96-kube-api-access-jclp2\") pod \"collect-profiles-29556690-p82jh\" (UID: \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.408005 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dc72a85-cdb5-4b11-9e0a-158d269edf96-config-volume\") pod \"collect-profiles-29556690-p82jh\" (UID: \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.430211 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqksf\" (UniqueName: \"kubernetes.io/projected/d7194752-8651-4e1b-8973-1f821bf23755-kube-api-access-cqksf\") pod \"auto-csr-approver-29556690-d6xbb\" (UID: \"d7194752-8651-4e1b-8973-1f821bf23755\") " pod="openshift-infra/auto-csr-approver-29556690-d6xbb" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.430736 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dc72a85-cdb5-4b11-9e0a-158d269edf96-secret-volume\") pod \"collect-profiles-29556690-p82jh\" (UID: \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.442133 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jclp2\" (UniqueName: \"kubernetes.io/projected/9dc72a85-cdb5-4b11-9e0a-158d269edf96-kube-api-access-jclp2\") pod \"collect-profiles-29556690-p82jh\" (UID: \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.546152 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556690-d6xbb" Mar 13 11:30:00 crc kubenswrapper[4632]: I0313 11:30:00.566102 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" Mar 13 11:30:01 crc kubenswrapper[4632]: I0313 11:30:01.100712 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556690-d6xbb"] Mar 13 11:30:01 crc kubenswrapper[4632]: W0313 11:30:01.107236 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7194752_8651_4e1b_8973_1f821bf23755.slice/crio-a3de5a9af6c5ad84f4ba96edda904272f20626a70fd4e02596d6dc30b9c18f6b WatchSource:0}: Error finding container a3de5a9af6c5ad84f4ba96edda904272f20626a70fd4e02596d6dc30b9c18f6b: Status 404 returned error can't find the container with id a3de5a9af6c5ad84f4ba96edda904272f20626a70fd4e02596d6dc30b9c18f6b Mar 13 11:30:01 crc kubenswrapper[4632]: I0313 11:30:01.214755 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh"] Mar 13 11:30:01 crc kubenswrapper[4632]: W0313 11:30:01.222631 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9dc72a85_cdb5_4b11_9e0a_158d269edf96.slice/crio-e633c3a60e4e59ec14ddf042913c9d08e67dd9113f7b4b60d8100e72e69b8565 WatchSource:0}: Error finding container e633c3a60e4e59ec14ddf042913c9d08e67dd9113f7b4b60d8100e72e69b8565: Status 404 returned error can't find the container with id e633c3a60e4e59ec14ddf042913c9d08e67dd9113f7b4b60d8100e72e69b8565 Mar 13 11:30:01 crc kubenswrapper[4632]: I0313 11:30:01.873141 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556690-d6xbb" event={"ID":"d7194752-8651-4e1b-8973-1f821bf23755","Type":"ContainerStarted","Data":"a3de5a9af6c5ad84f4ba96edda904272f20626a70fd4e02596d6dc30b9c18f6b"} Mar 13 11:30:01 crc kubenswrapper[4632]: I0313 11:30:01.874845 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" event={"ID":"9dc72a85-cdb5-4b11-9e0a-158d269edf96","Type":"ContainerStarted","Data":"526c0b7d143109242f29250c0cffd4a40f383eaf78da9d0786f09bf0aa0eccb3"} Mar 13 11:30:01 crc kubenswrapper[4632]: I0313 11:30:01.874933 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" event={"ID":"9dc72a85-cdb5-4b11-9e0a-158d269edf96","Type":"ContainerStarted","Data":"e633c3a60e4e59ec14ddf042913c9d08e67dd9113f7b4b60d8100e72e69b8565"} Mar 13 11:30:01 crc kubenswrapper[4632]: I0313 11:30:01.900354 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" podStartSLOduration=1.90033506 podStartE2EDuration="1.90033506s" podCreationTimestamp="2026-03-13 11:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 11:30:01.892367503 +0000 UTC m=+5175.914897646" watchObservedRunningTime="2026-03-13 11:30:01.90033506 +0000 UTC m=+5175.922865193" Mar 13 11:30:02 crc kubenswrapper[4632]: I0313 11:30:02.886461 4632 generic.go:334] "Generic (PLEG): container finished" podID="9dc72a85-cdb5-4b11-9e0a-158d269edf96" containerID="526c0b7d143109242f29250c0cffd4a40f383eaf78da9d0786f09bf0aa0eccb3" exitCode=0 Mar 13 11:30:02 crc kubenswrapper[4632]: I0313 11:30:02.886550 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" event={"ID":"9dc72a85-cdb5-4b11-9e0a-158d269edf96","Type":"ContainerDied","Data":"526c0b7d143109242f29250c0cffd4a40f383eaf78da9d0786f09bf0aa0eccb3"} Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.291965 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.391055 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dc72a85-cdb5-4b11-9e0a-158d269edf96-secret-volume\") pod \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\" (UID: \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\") " Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.391160 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dc72a85-cdb5-4b11-9e0a-158d269edf96-config-volume\") pod \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\" (UID: \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\") " Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.391448 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jclp2\" (UniqueName: \"kubernetes.io/projected/9dc72a85-cdb5-4b11-9e0a-158d269edf96-kube-api-access-jclp2\") pod \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\" (UID: \"9dc72a85-cdb5-4b11-9e0a-158d269edf96\") " Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.391788 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9dc72a85-cdb5-4b11-9e0a-158d269edf96-config-volume" (OuterVolumeSpecName: "config-volume") pod "9dc72a85-cdb5-4b11-9e0a-158d269edf96" (UID: "9dc72a85-cdb5-4b11-9e0a-158d269edf96"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.392518 4632 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/9dc72a85-cdb5-4b11-9e0a-158d269edf96-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.397610 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dc72a85-cdb5-4b11-9e0a-158d269edf96-kube-api-access-jclp2" (OuterVolumeSpecName: "kube-api-access-jclp2") pod "9dc72a85-cdb5-4b11-9e0a-158d269edf96" (UID: "9dc72a85-cdb5-4b11-9e0a-158d269edf96"). InnerVolumeSpecName "kube-api-access-jclp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.398137 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dc72a85-cdb5-4b11-9e0a-158d269edf96-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "9dc72a85-cdb5-4b11-9e0a-158d269edf96" (UID: "9dc72a85-cdb5-4b11-9e0a-158d269edf96"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.494404 4632 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/9dc72a85-cdb5-4b11-9e0a-158d269edf96-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.494435 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jclp2\" (UniqueName: \"kubernetes.io/projected/9dc72a85-cdb5-4b11-9e0a-158d269edf96-kube-api-access-jclp2\") on node \"crc\" DevicePath \"\"" Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.910047 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556690-d6xbb" event={"ID":"d7194752-8651-4e1b-8973-1f821bf23755","Type":"ContainerStarted","Data":"a14245a1819d4e34ca7541b00cd28e096f96a6ad8f1b997d5a64006b8dd2f7c5"} Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.917145 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" event={"ID":"9dc72a85-cdb5-4b11-9e0a-158d269edf96","Type":"ContainerDied","Data":"e633c3a60e4e59ec14ddf042913c9d08e67dd9113f7b4b60d8100e72e69b8565"} Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.917182 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e633c3a60e4e59ec14ddf042913c9d08e67dd9113f7b4b60d8100e72e69b8565" Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.917234 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh" Mar 13 11:30:04 crc kubenswrapper[4632]: I0313 11:30:04.937602 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556690-d6xbb" podStartSLOduration=2.906815343 podStartE2EDuration="4.937578749s" podCreationTimestamp="2026-03-13 11:30:00 +0000 UTC" firstStartedPulling="2026-03-13 11:30:01.111020509 +0000 UTC m=+5175.133550652" lastFinishedPulling="2026-03-13 11:30:03.141783925 +0000 UTC m=+5177.164314058" observedRunningTime="2026-03-13 11:30:04.92872052 +0000 UTC m=+5178.951250653" watchObservedRunningTime="2026-03-13 11:30:04.937578749 +0000 UTC m=+5178.960108882" Mar 13 11:30:05 crc kubenswrapper[4632]: I0313 11:30:05.370352 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8"] Mar 13 11:30:05 crc kubenswrapper[4632]: I0313 11:30:05.386610 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556645-4btb8"] Mar 13 11:30:05 crc kubenswrapper[4632]: I0313 11:30:05.929786 4632 generic.go:334] "Generic (PLEG): container finished" podID="d7194752-8651-4e1b-8973-1f821bf23755" containerID="a14245a1819d4e34ca7541b00cd28e096f96a6ad8f1b997d5a64006b8dd2f7c5" exitCode=0 Mar 13 11:30:05 crc kubenswrapper[4632]: I0313 11:30:05.929893 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556690-d6xbb" event={"ID":"d7194752-8651-4e1b-8973-1f821bf23755","Type":"ContainerDied","Data":"a14245a1819d4e34ca7541b00cd28e096f96a6ad8f1b997d5a64006b8dd2f7c5"} Mar 13 11:30:06 crc kubenswrapper[4632]: I0313 11:30:06.079897 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cf52265-21b3-40f0-a2f5-d379c03cc045" path="/var/lib/kubelet/pods/8cf52265-21b3-40f0-a2f5-d379c03cc045/volumes" Mar 13 11:30:07 crc kubenswrapper[4632]: I0313 11:30:07.372607 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556690-d6xbb" Mar 13 11:30:07 crc kubenswrapper[4632]: I0313 11:30:07.454485 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cqksf\" (UniqueName: \"kubernetes.io/projected/d7194752-8651-4e1b-8973-1f821bf23755-kube-api-access-cqksf\") pod \"d7194752-8651-4e1b-8973-1f821bf23755\" (UID: \"d7194752-8651-4e1b-8973-1f821bf23755\") " Mar 13 11:30:07 crc kubenswrapper[4632]: I0313 11:30:07.467209 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7194752-8651-4e1b-8973-1f821bf23755-kube-api-access-cqksf" (OuterVolumeSpecName: "kube-api-access-cqksf") pod "d7194752-8651-4e1b-8973-1f821bf23755" (UID: "d7194752-8651-4e1b-8973-1f821bf23755"). InnerVolumeSpecName "kube-api-access-cqksf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:30:07 crc kubenswrapper[4632]: I0313 11:30:07.557260 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cqksf\" (UniqueName: \"kubernetes.io/projected/d7194752-8651-4e1b-8973-1f821bf23755-kube-api-access-cqksf\") on node \"crc\" DevicePath \"\"" Mar 13 11:30:07 crc kubenswrapper[4632]: I0313 11:30:07.950093 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556690-d6xbb" event={"ID":"d7194752-8651-4e1b-8973-1f821bf23755","Type":"ContainerDied","Data":"a3de5a9af6c5ad84f4ba96edda904272f20626a70fd4e02596d6dc30b9c18f6b"} Mar 13 11:30:07 crc kubenswrapper[4632]: I0313 11:30:07.950191 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556690-d6xbb" Mar 13 11:30:07 crc kubenswrapper[4632]: I0313 11:30:07.950928 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3de5a9af6c5ad84f4ba96edda904272f20626a70fd4e02596d6dc30b9c18f6b" Mar 13 11:30:08 crc kubenswrapper[4632]: I0313 11:30:08.005326 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556684-bd9st"] Mar 13 11:30:08 crc kubenswrapper[4632]: I0313 11:30:08.022139 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556684-bd9st"] Mar 13 11:30:08 crc kubenswrapper[4632]: I0313 11:30:08.058068 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb45f1ce-58e0-4f55-afd6-2e14db5f24ca" path="/var/lib/kubelet/pods/fb45f1ce-58e0-4f55-afd6-2e14db5f24ca/volumes" Mar 13 11:30:13 crc kubenswrapper[4632]: I0313 11:30:13.045746 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:30:13 crc kubenswrapper[4632]: E0313 11:30:13.047215 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:30:20 crc kubenswrapper[4632]: I0313 11:30:20.431444 4632 scope.go:117] "RemoveContainer" containerID="8e20db958c001216e89a657171c617c2e4d78b297bcd654a9af9c2d8d32242ac" Mar 13 11:30:20 crc kubenswrapper[4632]: I0313 11:30:20.552116 4632 scope.go:117] "RemoveContainer" containerID="2af97c9efffc6f3dc7413dcff6c97889a640ef442506af8ad264876a675427dc" Mar 13 11:30:27 crc kubenswrapper[4632]: I0313 11:30:27.044908 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:30:27 crc kubenswrapper[4632]: E0313 11:30:27.047064 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:30:39 crc kubenswrapper[4632]: I0313 11:30:39.043877 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:30:39 crc kubenswrapper[4632]: E0313 11:30:39.044642 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:30:53 crc kubenswrapper[4632]: I0313 11:30:53.044145 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:30:53 crc kubenswrapper[4632]: E0313 11:30:53.045245 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:31:05 crc kubenswrapper[4632]: I0313 11:31:05.044679 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:31:05 crc kubenswrapper[4632]: E0313 11:31:05.045531 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:31:17 crc kubenswrapper[4632]: I0313 11:31:17.045212 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:31:17 crc kubenswrapper[4632]: E0313 11:31:17.046488 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:31:31 crc kubenswrapper[4632]: I0313 11:31:31.044684 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:31:31 crc kubenswrapper[4632]: E0313 11:31:31.045436 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:31:43 crc kubenswrapper[4632]: I0313 11:31:43.044391 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:31:43 crc kubenswrapper[4632]: E0313 11:31:43.046709 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:31:57 crc kubenswrapper[4632]: I0313 11:31:57.044085 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:31:57 crc kubenswrapper[4632]: E0313 11:31:57.045794 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.155404 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556692-4kd7v"] Mar 13 11:32:00 crc kubenswrapper[4632]: E0313 11:32:00.156147 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dc72a85-cdb5-4b11-9e0a-158d269edf96" containerName="collect-profiles" Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.156165 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dc72a85-cdb5-4b11-9e0a-158d269edf96" containerName="collect-profiles" Mar 13 11:32:00 crc kubenswrapper[4632]: E0313 11:32:00.156206 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7194752-8651-4e1b-8973-1f821bf23755" containerName="oc" Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.156214 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7194752-8651-4e1b-8973-1f821bf23755" containerName="oc" Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.156427 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dc72a85-cdb5-4b11-9e0a-158d269edf96" containerName="collect-profiles" Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.156457 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7194752-8651-4e1b-8973-1f821bf23755" containerName="oc" Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.157212 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556692-4kd7v" Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.162519 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.163081 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.163409 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.165748 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556692-4kd7v"] Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.319982 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2lwc\" (UniqueName: \"kubernetes.io/projected/77ae44b0-e101-4d21-87e5-9e213e024e9e-kube-api-access-v2lwc\") pod \"auto-csr-approver-29556692-4kd7v\" (UID: \"77ae44b0-e101-4d21-87e5-9e213e024e9e\") " pod="openshift-infra/auto-csr-approver-29556692-4kd7v" Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.421868 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2lwc\" (UniqueName: \"kubernetes.io/projected/77ae44b0-e101-4d21-87e5-9e213e024e9e-kube-api-access-v2lwc\") pod \"auto-csr-approver-29556692-4kd7v\" (UID: \"77ae44b0-e101-4d21-87e5-9e213e024e9e\") " pod="openshift-infra/auto-csr-approver-29556692-4kd7v" Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.452867 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2lwc\" (UniqueName: \"kubernetes.io/projected/77ae44b0-e101-4d21-87e5-9e213e024e9e-kube-api-access-v2lwc\") pod \"auto-csr-approver-29556692-4kd7v\" (UID: \"77ae44b0-e101-4d21-87e5-9e213e024e9e\") " pod="openshift-infra/auto-csr-approver-29556692-4kd7v" Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.481856 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556692-4kd7v" Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.958795 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556692-4kd7v"] Mar 13 11:32:00 crc kubenswrapper[4632]: I0313 11:32:00.967355 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 11:32:01 crc kubenswrapper[4632]: I0313 11:32:01.102605 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556692-4kd7v" event={"ID":"77ae44b0-e101-4d21-87e5-9e213e024e9e","Type":"ContainerStarted","Data":"daa9f46f0da478ec17ba30a85b783ac0ac768a525ff861a88de96fe1e24a934f"} Mar 13 11:32:03 crc kubenswrapper[4632]: I0313 11:32:03.121039 4632 generic.go:334] "Generic (PLEG): container finished" podID="77ae44b0-e101-4d21-87e5-9e213e024e9e" containerID="b8f22a62b885c530c5401c31b24e17ea8bcf63d6debf02e44b4f01a4ab4c1102" exitCode=0 Mar 13 11:32:03 crc kubenswrapper[4632]: I0313 11:32:03.121479 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556692-4kd7v" event={"ID":"77ae44b0-e101-4d21-87e5-9e213e024e9e","Type":"ContainerDied","Data":"b8f22a62b885c530c5401c31b24e17ea8bcf63d6debf02e44b4f01a4ab4c1102"} Mar 13 11:32:04 crc kubenswrapper[4632]: I0313 11:32:04.575209 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556692-4kd7v" Mar 13 11:32:04 crc kubenswrapper[4632]: I0313 11:32:04.721429 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2lwc\" (UniqueName: \"kubernetes.io/projected/77ae44b0-e101-4d21-87e5-9e213e024e9e-kube-api-access-v2lwc\") pod \"77ae44b0-e101-4d21-87e5-9e213e024e9e\" (UID: \"77ae44b0-e101-4d21-87e5-9e213e024e9e\") " Mar 13 11:32:04 crc kubenswrapper[4632]: I0313 11:32:04.730190 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/77ae44b0-e101-4d21-87e5-9e213e024e9e-kube-api-access-v2lwc" (OuterVolumeSpecName: "kube-api-access-v2lwc") pod "77ae44b0-e101-4d21-87e5-9e213e024e9e" (UID: "77ae44b0-e101-4d21-87e5-9e213e024e9e"). InnerVolumeSpecName "kube-api-access-v2lwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:32:04 crc kubenswrapper[4632]: I0313 11:32:04.824390 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2lwc\" (UniqueName: \"kubernetes.io/projected/77ae44b0-e101-4d21-87e5-9e213e024e9e-kube-api-access-v2lwc\") on node \"crc\" DevicePath \"\"" Mar 13 11:32:05 crc kubenswrapper[4632]: I0313 11:32:05.143833 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556692-4kd7v" event={"ID":"77ae44b0-e101-4d21-87e5-9e213e024e9e","Type":"ContainerDied","Data":"daa9f46f0da478ec17ba30a85b783ac0ac768a525ff861a88de96fe1e24a934f"} Mar 13 11:32:05 crc kubenswrapper[4632]: I0313 11:32:05.144195 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="daa9f46f0da478ec17ba30a85b783ac0ac768a525ff861a88de96fe1e24a934f" Mar 13 11:32:05 crc kubenswrapper[4632]: I0313 11:32:05.143904 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556692-4kd7v" Mar 13 11:32:05 crc kubenswrapper[4632]: I0313 11:32:05.691056 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556686-v9c2p"] Mar 13 11:32:05 crc kubenswrapper[4632]: I0313 11:32:05.704263 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556686-v9c2p"] Mar 13 11:32:06 crc kubenswrapper[4632]: I0313 11:32:06.064188 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="78358d6e-af24-4da5-8c77-9453e6228cda" path="/var/lib/kubelet/pods/78358d6e-af24-4da5-8c77-9453e6228cda/volumes" Mar 13 11:32:08 crc kubenswrapper[4632]: I0313 11:32:08.046786 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:32:08 crc kubenswrapper[4632]: E0313 11:32:08.047405 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:32:20 crc kubenswrapper[4632]: I0313 11:32:20.731294 4632 scope.go:117] "RemoveContainer" containerID="0306bee576d03326b01d2c08c76cf2909394a8ec9e729a13ca12d86ebb721532" Mar 13 11:32:23 crc kubenswrapper[4632]: I0313 11:32:23.044834 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:32:23 crc kubenswrapper[4632]: E0313 11:32:23.045885 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:32:38 crc kubenswrapper[4632]: I0313 11:32:38.050781 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:32:38 crc kubenswrapper[4632]: E0313 11:32:38.051530 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:32:52 crc kubenswrapper[4632]: I0313 11:32:52.044368 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:32:52 crc kubenswrapper[4632]: E0313 11:32:52.045206 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:33:04 crc kubenswrapper[4632]: I0313 11:33:04.045589 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:33:04 crc kubenswrapper[4632]: E0313 11:33:04.046698 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:33:15 crc kubenswrapper[4632]: I0313 11:33:15.044027 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:33:15 crc kubenswrapper[4632]: E0313 11:33:15.044660 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:33:30 crc kubenswrapper[4632]: I0313 11:33:30.044108 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:33:30 crc kubenswrapper[4632]: E0313 11:33:30.045921 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:33:42 crc kubenswrapper[4632]: I0313 11:33:42.045683 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:33:43 crc kubenswrapper[4632]: I0313 11:33:43.102562 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"b28a3031014e23a161560bdf4de3a19a21d26729102cf99acd465c2bd90c33f9"} Mar 13 11:34:00 crc kubenswrapper[4632]: I0313 11:34:00.147281 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556694-jmf5t"] Mar 13 11:34:00 crc kubenswrapper[4632]: E0313 11:34:00.149490 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="77ae44b0-e101-4d21-87e5-9e213e024e9e" containerName="oc" Mar 13 11:34:00 crc kubenswrapper[4632]: I0313 11:34:00.149664 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="77ae44b0-e101-4d21-87e5-9e213e024e9e" containerName="oc" Mar 13 11:34:00 crc kubenswrapper[4632]: I0313 11:34:00.150251 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="77ae44b0-e101-4d21-87e5-9e213e024e9e" containerName="oc" Mar 13 11:34:00 crc kubenswrapper[4632]: I0313 11:34:00.151880 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556694-jmf5t" Mar 13 11:34:00 crc kubenswrapper[4632]: I0313 11:34:00.158555 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:34:00 crc kubenswrapper[4632]: I0313 11:34:00.159262 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:34:00 crc kubenswrapper[4632]: I0313 11:34:00.159277 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:34:00 crc kubenswrapper[4632]: I0313 11:34:00.168335 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556694-jmf5t"] Mar 13 11:34:00 crc kubenswrapper[4632]: I0313 11:34:00.217856 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5tl6\" (UniqueName: \"kubernetes.io/projected/266f7f6e-de91-4256-8605-0a71adef85fc-kube-api-access-l5tl6\") pod \"auto-csr-approver-29556694-jmf5t\" (UID: \"266f7f6e-de91-4256-8605-0a71adef85fc\") " pod="openshift-infra/auto-csr-approver-29556694-jmf5t" Mar 13 11:34:00 crc kubenswrapper[4632]: I0313 11:34:00.318855 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5tl6\" (UniqueName: \"kubernetes.io/projected/266f7f6e-de91-4256-8605-0a71adef85fc-kube-api-access-l5tl6\") pod \"auto-csr-approver-29556694-jmf5t\" (UID: \"266f7f6e-de91-4256-8605-0a71adef85fc\") " pod="openshift-infra/auto-csr-approver-29556694-jmf5t" Mar 13 11:34:00 crc kubenswrapper[4632]: I0313 11:34:00.340718 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5tl6\" (UniqueName: \"kubernetes.io/projected/266f7f6e-de91-4256-8605-0a71adef85fc-kube-api-access-l5tl6\") pod \"auto-csr-approver-29556694-jmf5t\" (UID: \"266f7f6e-de91-4256-8605-0a71adef85fc\") " pod="openshift-infra/auto-csr-approver-29556694-jmf5t" Mar 13 11:34:00 crc kubenswrapper[4632]: I0313 11:34:00.477138 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556694-jmf5t" Mar 13 11:34:01 crc kubenswrapper[4632]: I0313 11:34:01.616488 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556694-jmf5t"] Mar 13 11:34:02 crc kubenswrapper[4632]: I0313 11:34:02.282109 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556694-jmf5t" event={"ID":"266f7f6e-de91-4256-8605-0a71adef85fc","Type":"ContainerStarted","Data":"6159272a14d87ccbf5e486985e550d0df0ab2f55be3e4a304e8ba69362c2f3cf"} Mar 13 11:34:03 crc kubenswrapper[4632]: I0313 11:34:03.292113 4632 generic.go:334] "Generic (PLEG): container finished" podID="266f7f6e-de91-4256-8605-0a71adef85fc" containerID="dd29187096f712bf2f18fa46086683fcb900aea6c3d89212b78286e73075a17b" exitCode=0 Mar 13 11:34:03 crc kubenswrapper[4632]: I0313 11:34:03.292226 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556694-jmf5t" event={"ID":"266f7f6e-de91-4256-8605-0a71adef85fc","Type":"ContainerDied","Data":"dd29187096f712bf2f18fa46086683fcb900aea6c3d89212b78286e73075a17b"} Mar 13 11:34:04 crc kubenswrapper[4632]: I0313 11:34:04.717281 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556694-jmf5t" Mar 13 11:34:04 crc kubenswrapper[4632]: I0313 11:34:04.824244 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5tl6\" (UniqueName: \"kubernetes.io/projected/266f7f6e-de91-4256-8605-0a71adef85fc-kube-api-access-l5tl6\") pod \"266f7f6e-de91-4256-8605-0a71adef85fc\" (UID: \"266f7f6e-de91-4256-8605-0a71adef85fc\") " Mar 13 11:34:04 crc kubenswrapper[4632]: I0313 11:34:04.833288 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/266f7f6e-de91-4256-8605-0a71adef85fc-kube-api-access-l5tl6" (OuterVolumeSpecName: "kube-api-access-l5tl6") pod "266f7f6e-de91-4256-8605-0a71adef85fc" (UID: "266f7f6e-de91-4256-8605-0a71adef85fc"). InnerVolumeSpecName "kube-api-access-l5tl6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:34:04 crc kubenswrapper[4632]: I0313 11:34:04.926621 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5tl6\" (UniqueName: \"kubernetes.io/projected/266f7f6e-de91-4256-8605-0a71adef85fc-kube-api-access-l5tl6\") on node \"crc\" DevicePath \"\"" Mar 13 11:34:05 crc kubenswrapper[4632]: I0313 11:34:05.311021 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556694-jmf5t" event={"ID":"266f7f6e-de91-4256-8605-0a71adef85fc","Type":"ContainerDied","Data":"6159272a14d87ccbf5e486985e550d0df0ab2f55be3e4a304e8ba69362c2f3cf"} Mar 13 11:34:05 crc kubenswrapper[4632]: I0313 11:34:05.311086 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6159272a14d87ccbf5e486985e550d0df0ab2f55be3e4a304e8ba69362c2f3cf" Mar 13 11:34:05 crc kubenswrapper[4632]: I0313 11:34:05.311159 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556694-jmf5t" Mar 13 11:34:05 crc kubenswrapper[4632]: I0313 11:34:05.796812 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556688-btp9b"] Mar 13 11:34:05 crc kubenswrapper[4632]: I0313 11:34:05.807478 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556688-btp9b"] Mar 13 11:34:06 crc kubenswrapper[4632]: I0313 11:34:06.057358 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99bcc88d-9858-4d4f-97e5-68e185f06401" path="/var/lib/kubelet/pods/99bcc88d-9858-4d4f-97e5-68e185f06401/volumes" Mar 13 11:34:20 crc kubenswrapper[4632]: I0313 11:34:20.875772 4632 scope.go:117] "RemoveContainer" containerID="419020d6577499249f09db214d08f7500440b2163407ca9b721f361af1ab72f7" Mar 13 11:36:00 crc kubenswrapper[4632]: I0313 11:36:00.151652 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556696-6dgq6"] Mar 13 11:36:00 crc kubenswrapper[4632]: E0313 11:36:00.152713 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="266f7f6e-de91-4256-8605-0a71adef85fc" containerName="oc" Mar 13 11:36:00 crc kubenswrapper[4632]: I0313 11:36:00.152757 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="266f7f6e-de91-4256-8605-0a71adef85fc" containerName="oc" Mar 13 11:36:00 crc kubenswrapper[4632]: I0313 11:36:00.153042 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="266f7f6e-de91-4256-8605-0a71adef85fc" containerName="oc" Mar 13 11:36:00 crc kubenswrapper[4632]: I0313 11:36:00.153775 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556696-6dgq6" Mar 13 11:36:00 crc kubenswrapper[4632]: I0313 11:36:00.155764 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:36:00 crc kubenswrapper[4632]: I0313 11:36:00.156085 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:36:00 crc kubenswrapper[4632]: I0313 11:36:00.156738 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:36:00 crc kubenswrapper[4632]: I0313 11:36:00.173606 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556696-6dgq6"] Mar 13 11:36:00 crc kubenswrapper[4632]: I0313 11:36:00.294176 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rtg5\" (UniqueName: \"kubernetes.io/projected/f8600f7f-45fb-4aa6-b13b-9d6be5354009-kube-api-access-2rtg5\") pod \"auto-csr-approver-29556696-6dgq6\" (UID: \"f8600f7f-45fb-4aa6-b13b-9d6be5354009\") " pod="openshift-infra/auto-csr-approver-29556696-6dgq6" Mar 13 11:36:00 crc kubenswrapper[4632]: I0313 11:36:00.397042 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rtg5\" (UniqueName: \"kubernetes.io/projected/f8600f7f-45fb-4aa6-b13b-9d6be5354009-kube-api-access-2rtg5\") pod \"auto-csr-approver-29556696-6dgq6\" (UID: \"f8600f7f-45fb-4aa6-b13b-9d6be5354009\") " pod="openshift-infra/auto-csr-approver-29556696-6dgq6" Mar 13 11:36:00 crc kubenswrapper[4632]: I0313 11:36:00.425438 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rtg5\" (UniqueName: \"kubernetes.io/projected/f8600f7f-45fb-4aa6-b13b-9d6be5354009-kube-api-access-2rtg5\") pod \"auto-csr-approver-29556696-6dgq6\" (UID: \"f8600f7f-45fb-4aa6-b13b-9d6be5354009\") " pod="openshift-infra/auto-csr-approver-29556696-6dgq6" Mar 13 11:36:00 crc kubenswrapper[4632]: I0313 11:36:00.479648 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556696-6dgq6" Mar 13 11:36:01 crc kubenswrapper[4632]: I0313 11:36:01.156420 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556696-6dgq6"] Mar 13 11:36:01 crc kubenswrapper[4632]: W0313 11:36:01.359496 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8600f7f_45fb_4aa6_b13b_9d6be5354009.slice/crio-80667f994774ca7286e4b847a63dc26b70323a254a39964981595c2cdd8a61d7 WatchSource:0}: Error finding container 80667f994774ca7286e4b847a63dc26b70323a254a39964981595c2cdd8a61d7: Status 404 returned error can't find the container with id 80667f994774ca7286e4b847a63dc26b70323a254a39964981595c2cdd8a61d7 Mar 13 11:36:01 crc kubenswrapper[4632]: I0313 11:36:01.448245 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556696-6dgq6" event={"ID":"f8600f7f-45fb-4aa6-b13b-9d6be5354009","Type":"ContainerStarted","Data":"80667f994774ca7286e4b847a63dc26b70323a254a39964981595c2cdd8a61d7"} Mar 13 11:36:02 crc kubenswrapper[4632]: I0313 11:36:02.459719 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556696-6dgq6" event={"ID":"f8600f7f-45fb-4aa6-b13b-9d6be5354009","Type":"ContainerStarted","Data":"ffbf598df91f4bb7277b432bb2bc1355e735cdb640ec4482a312abc6e198f0af"} Mar 13 11:36:05 crc kubenswrapper[4632]: I0313 11:36:05.489006 4632 generic.go:334] "Generic (PLEG): container finished" podID="f8600f7f-45fb-4aa6-b13b-9d6be5354009" containerID="ffbf598df91f4bb7277b432bb2bc1355e735cdb640ec4482a312abc6e198f0af" exitCode=0 Mar 13 11:36:05 crc kubenswrapper[4632]: I0313 11:36:05.489086 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556696-6dgq6" event={"ID":"f8600f7f-45fb-4aa6-b13b-9d6be5354009","Type":"ContainerDied","Data":"ffbf598df91f4bb7277b432bb2bc1355e735cdb640ec4482a312abc6e198f0af"} Mar 13 11:36:06 crc kubenswrapper[4632]: I0313 11:36:06.963577 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556696-6dgq6" Mar 13 11:36:07 crc kubenswrapper[4632]: I0313 11:36:07.037547 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2rtg5\" (UniqueName: \"kubernetes.io/projected/f8600f7f-45fb-4aa6-b13b-9d6be5354009-kube-api-access-2rtg5\") pod \"f8600f7f-45fb-4aa6-b13b-9d6be5354009\" (UID: \"f8600f7f-45fb-4aa6-b13b-9d6be5354009\") " Mar 13 11:36:07 crc kubenswrapper[4632]: I0313 11:36:07.045356 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8600f7f-45fb-4aa6-b13b-9d6be5354009-kube-api-access-2rtg5" (OuterVolumeSpecName: "kube-api-access-2rtg5") pod "f8600f7f-45fb-4aa6-b13b-9d6be5354009" (UID: "f8600f7f-45fb-4aa6-b13b-9d6be5354009"). InnerVolumeSpecName "kube-api-access-2rtg5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:36:07 crc kubenswrapper[4632]: I0313 11:36:07.142395 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2rtg5\" (UniqueName: \"kubernetes.io/projected/f8600f7f-45fb-4aa6-b13b-9d6be5354009-kube-api-access-2rtg5\") on node \"crc\" DevicePath \"\"" Mar 13 11:36:07 crc kubenswrapper[4632]: I0313 11:36:07.511117 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556696-6dgq6" event={"ID":"f8600f7f-45fb-4aa6-b13b-9d6be5354009","Type":"ContainerDied","Data":"80667f994774ca7286e4b847a63dc26b70323a254a39964981595c2cdd8a61d7"} Mar 13 11:36:07 crc kubenswrapper[4632]: I0313 11:36:07.511436 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="80667f994774ca7286e4b847a63dc26b70323a254a39964981595c2cdd8a61d7" Mar 13 11:36:07 crc kubenswrapper[4632]: I0313 11:36:07.511180 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556696-6dgq6" Mar 13 11:36:07 crc kubenswrapper[4632]: I0313 11:36:07.578907 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556690-d6xbb"] Mar 13 11:36:07 crc kubenswrapper[4632]: I0313 11:36:07.587309 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556690-d6xbb"] Mar 13 11:36:08 crc kubenswrapper[4632]: I0313 11:36:08.064655 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7194752-8651-4e1b-8973-1f821bf23755" path="/var/lib/kubelet/pods/d7194752-8651-4e1b-8973-1f821bf23755/volumes" Mar 13 11:36:10 crc kubenswrapper[4632]: I0313 11:36:10.461449 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:36:10 crc kubenswrapper[4632]: I0313 11:36:10.476680 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:36:18 crc kubenswrapper[4632]: I0313 11:36:18.978455 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bv9w5"] Mar 13 11:36:18 crc kubenswrapper[4632]: E0313 11:36:18.980518 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8600f7f-45fb-4aa6-b13b-9d6be5354009" containerName="oc" Mar 13 11:36:18 crc kubenswrapper[4632]: I0313 11:36:18.980628 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8600f7f-45fb-4aa6-b13b-9d6be5354009" containerName="oc" Mar 13 11:36:18 crc kubenswrapper[4632]: I0313 11:36:18.980886 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8600f7f-45fb-4aa6-b13b-9d6be5354009" containerName="oc" Mar 13 11:36:18 crc kubenswrapper[4632]: I0313 11:36:18.982464 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:18 crc kubenswrapper[4632]: I0313 11:36:18.988574 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb192b24-7638-4c11-9936-ad94a6842ce9-catalog-content\") pod \"redhat-marketplace-bv9w5\" (UID: \"bb192b24-7638-4c11-9936-ad94a6842ce9\") " pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:18 crc kubenswrapper[4632]: I0313 11:36:18.988820 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb192b24-7638-4c11-9936-ad94a6842ce9-utilities\") pod \"redhat-marketplace-bv9w5\" (UID: \"bb192b24-7638-4c11-9936-ad94a6842ce9\") " pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:18 crc kubenswrapper[4632]: I0313 11:36:18.988931 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97zsq\" (UniqueName: \"kubernetes.io/projected/bb192b24-7638-4c11-9936-ad94a6842ce9-kube-api-access-97zsq\") pod \"redhat-marketplace-bv9w5\" (UID: \"bb192b24-7638-4c11-9936-ad94a6842ce9\") " pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:19 crc kubenswrapper[4632]: I0313 11:36:19.002281 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bv9w5"] Mar 13 11:36:19 crc kubenswrapper[4632]: I0313 11:36:19.091730 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb192b24-7638-4c11-9936-ad94a6842ce9-catalog-content\") pod \"redhat-marketplace-bv9w5\" (UID: \"bb192b24-7638-4c11-9936-ad94a6842ce9\") " pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:19 crc kubenswrapper[4632]: I0313 11:36:19.091784 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb192b24-7638-4c11-9936-ad94a6842ce9-utilities\") pod \"redhat-marketplace-bv9w5\" (UID: \"bb192b24-7638-4c11-9936-ad94a6842ce9\") " pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:19 crc kubenswrapper[4632]: I0313 11:36:19.091851 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97zsq\" (UniqueName: \"kubernetes.io/projected/bb192b24-7638-4c11-9936-ad94a6842ce9-kube-api-access-97zsq\") pod \"redhat-marketplace-bv9w5\" (UID: \"bb192b24-7638-4c11-9936-ad94a6842ce9\") " pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:19 crc kubenswrapper[4632]: I0313 11:36:19.092595 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb192b24-7638-4c11-9936-ad94a6842ce9-utilities\") pod \"redhat-marketplace-bv9w5\" (UID: \"bb192b24-7638-4c11-9936-ad94a6842ce9\") " pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:19 crc kubenswrapper[4632]: I0313 11:36:19.093041 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb192b24-7638-4c11-9936-ad94a6842ce9-catalog-content\") pod \"redhat-marketplace-bv9w5\" (UID: \"bb192b24-7638-4c11-9936-ad94a6842ce9\") " pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:19 crc kubenswrapper[4632]: I0313 11:36:19.112787 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97zsq\" (UniqueName: \"kubernetes.io/projected/bb192b24-7638-4c11-9936-ad94a6842ce9-kube-api-access-97zsq\") pod \"redhat-marketplace-bv9w5\" (UID: \"bb192b24-7638-4c11-9936-ad94a6842ce9\") " pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:19 crc kubenswrapper[4632]: I0313 11:36:19.306273 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:20 crc kubenswrapper[4632]: I0313 11:36:20.169391 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bv9w5"] Mar 13 11:36:20 crc kubenswrapper[4632]: I0313 11:36:20.677965 4632 generic.go:334] "Generic (PLEG): container finished" podID="bb192b24-7638-4c11-9936-ad94a6842ce9" containerID="6d026b8061dc2c3d9960ed047459f56a28590e7270cb04b536cdac0b85745294" exitCode=0 Mar 13 11:36:20 crc kubenswrapper[4632]: I0313 11:36:20.678068 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bv9w5" event={"ID":"bb192b24-7638-4c11-9936-ad94a6842ce9","Type":"ContainerDied","Data":"6d026b8061dc2c3d9960ed047459f56a28590e7270cb04b536cdac0b85745294"} Mar 13 11:36:20 crc kubenswrapper[4632]: I0313 11:36:20.678290 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bv9w5" event={"ID":"bb192b24-7638-4c11-9936-ad94a6842ce9","Type":"ContainerStarted","Data":"ea657aa4cd516e96e0356217039e8d93bddea9701a5de2aa302d85b9b36bebd7"} Mar 13 11:36:21 crc kubenswrapper[4632]: I0313 11:36:21.082096 4632 scope.go:117] "RemoveContainer" containerID="a14245a1819d4e34ca7541b00cd28e096f96a6ad8f1b997d5a64006b8dd2f7c5" Mar 13 11:36:21 crc kubenswrapper[4632]: I0313 11:36:21.695584 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bv9w5" event={"ID":"bb192b24-7638-4c11-9936-ad94a6842ce9","Type":"ContainerStarted","Data":"f1ce3fc4d0f620c27005352e1949578082140c75a9cff811a6bb65a60520f0d9"} Mar 13 11:36:23 crc kubenswrapper[4632]: I0313 11:36:23.717207 4632 generic.go:334] "Generic (PLEG): container finished" podID="bb192b24-7638-4c11-9936-ad94a6842ce9" containerID="f1ce3fc4d0f620c27005352e1949578082140c75a9cff811a6bb65a60520f0d9" exitCode=0 Mar 13 11:36:23 crc kubenswrapper[4632]: I0313 11:36:23.717310 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bv9w5" event={"ID":"bb192b24-7638-4c11-9936-ad94a6842ce9","Type":"ContainerDied","Data":"f1ce3fc4d0f620c27005352e1949578082140c75a9cff811a6bb65a60520f0d9"} Mar 13 11:36:24 crc kubenswrapper[4632]: I0313 11:36:24.731226 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bv9w5" event={"ID":"bb192b24-7638-4c11-9936-ad94a6842ce9","Type":"ContainerStarted","Data":"471b3e9129676bd35b3dfeb779906ed80e7614a37e876dbde9d3751b6bc56c61"} Mar 13 11:36:24 crc kubenswrapper[4632]: I0313 11:36:24.759509 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bv9w5" podStartSLOduration=3.223662111 podStartE2EDuration="6.759484632s" podCreationTimestamp="2026-03-13 11:36:18 +0000 UTC" firstStartedPulling="2026-03-13 11:36:20.680004387 +0000 UTC m=+5554.702534520" lastFinishedPulling="2026-03-13 11:36:24.215826908 +0000 UTC m=+5558.238357041" observedRunningTime="2026-03-13 11:36:24.757484113 +0000 UTC m=+5558.780014246" watchObservedRunningTime="2026-03-13 11:36:24.759484632 +0000 UTC m=+5558.782014775" Mar 13 11:36:29 crc kubenswrapper[4632]: I0313 11:36:29.307443 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:29 crc kubenswrapper[4632]: I0313 11:36:29.310702 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:30 crc kubenswrapper[4632]: I0313 11:36:30.356528 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-bv9w5" podUID="bb192b24-7638-4c11-9936-ad94a6842ce9" containerName="registry-server" probeResult="failure" output=< Mar 13 11:36:30 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:36:30 crc kubenswrapper[4632]: > Mar 13 11:36:39 crc kubenswrapper[4632]: I0313 11:36:39.367266 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:39 crc kubenswrapper[4632]: I0313 11:36:39.432219 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:39 crc kubenswrapper[4632]: I0313 11:36:39.625718 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bv9w5"] Mar 13 11:36:40 crc kubenswrapper[4632]: I0313 11:36:40.460681 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:36:40 crc kubenswrapper[4632]: I0313 11:36:40.460759 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:36:40 crc kubenswrapper[4632]: I0313 11:36:40.893859 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bv9w5" podUID="bb192b24-7638-4c11-9936-ad94a6842ce9" containerName="registry-server" containerID="cri-o://471b3e9129676bd35b3dfeb779906ed80e7614a37e876dbde9d3751b6bc56c61" gracePeriod=2 Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.501164 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.677836 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97zsq\" (UniqueName: \"kubernetes.io/projected/bb192b24-7638-4c11-9936-ad94a6842ce9-kube-api-access-97zsq\") pod \"bb192b24-7638-4c11-9936-ad94a6842ce9\" (UID: \"bb192b24-7638-4c11-9936-ad94a6842ce9\") " Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.678002 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb192b24-7638-4c11-9936-ad94a6842ce9-utilities\") pod \"bb192b24-7638-4c11-9936-ad94a6842ce9\" (UID: \"bb192b24-7638-4c11-9936-ad94a6842ce9\") " Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.678202 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb192b24-7638-4c11-9936-ad94a6842ce9-catalog-content\") pod \"bb192b24-7638-4c11-9936-ad94a6842ce9\" (UID: \"bb192b24-7638-4c11-9936-ad94a6842ce9\") " Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.678726 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb192b24-7638-4c11-9936-ad94a6842ce9-utilities" (OuterVolumeSpecName: "utilities") pod "bb192b24-7638-4c11-9936-ad94a6842ce9" (UID: "bb192b24-7638-4c11-9936-ad94a6842ce9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.684289 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb192b24-7638-4c11-9936-ad94a6842ce9-kube-api-access-97zsq" (OuterVolumeSpecName: "kube-api-access-97zsq") pod "bb192b24-7638-4c11-9936-ad94a6842ce9" (UID: "bb192b24-7638-4c11-9936-ad94a6842ce9"). InnerVolumeSpecName "kube-api-access-97zsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.710655 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb192b24-7638-4c11-9936-ad94a6842ce9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bb192b24-7638-4c11-9936-ad94a6842ce9" (UID: "bb192b24-7638-4c11-9936-ad94a6842ce9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.780671 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bb192b24-7638-4c11-9936-ad94a6842ce9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.780726 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97zsq\" (UniqueName: \"kubernetes.io/projected/bb192b24-7638-4c11-9936-ad94a6842ce9-kube-api-access-97zsq\") on node \"crc\" DevicePath \"\"" Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.780742 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bb192b24-7638-4c11-9936-ad94a6842ce9-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.904768 4632 generic.go:334] "Generic (PLEG): container finished" podID="bb192b24-7638-4c11-9936-ad94a6842ce9" containerID="471b3e9129676bd35b3dfeb779906ed80e7614a37e876dbde9d3751b6bc56c61" exitCode=0 Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.904815 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bv9w5" event={"ID":"bb192b24-7638-4c11-9936-ad94a6842ce9","Type":"ContainerDied","Data":"471b3e9129676bd35b3dfeb779906ed80e7614a37e876dbde9d3751b6bc56c61"} Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.904852 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bv9w5" event={"ID":"bb192b24-7638-4c11-9936-ad94a6842ce9","Type":"ContainerDied","Data":"ea657aa4cd516e96e0356217039e8d93bddea9701a5de2aa302d85b9b36bebd7"} Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.904846 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bv9w5" Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.904868 4632 scope.go:117] "RemoveContainer" containerID="471b3e9129676bd35b3dfeb779906ed80e7614a37e876dbde9d3751b6bc56c61" Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.936605 4632 scope.go:117] "RemoveContainer" containerID="f1ce3fc4d0f620c27005352e1949578082140c75a9cff811a6bb65a60520f0d9" Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.950386 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bv9w5"] Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.958418 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bv9w5"] Mar 13 11:36:41 crc kubenswrapper[4632]: I0313 11:36:41.959765 4632 scope.go:117] "RemoveContainer" containerID="6d026b8061dc2c3d9960ed047459f56a28590e7270cb04b536cdac0b85745294" Mar 13 11:36:42 crc kubenswrapper[4632]: I0313 11:36:42.003660 4632 scope.go:117] "RemoveContainer" containerID="471b3e9129676bd35b3dfeb779906ed80e7614a37e876dbde9d3751b6bc56c61" Mar 13 11:36:42 crc kubenswrapper[4632]: E0313 11:36:42.008480 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"471b3e9129676bd35b3dfeb779906ed80e7614a37e876dbde9d3751b6bc56c61\": container with ID starting with 471b3e9129676bd35b3dfeb779906ed80e7614a37e876dbde9d3751b6bc56c61 not found: ID does not exist" containerID="471b3e9129676bd35b3dfeb779906ed80e7614a37e876dbde9d3751b6bc56c61" Mar 13 11:36:42 crc kubenswrapper[4632]: I0313 11:36:42.008537 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"471b3e9129676bd35b3dfeb779906ed80e7614a37e876dbde9d3751b6bc56c61"} err="failed to get container status \"471b3e9129676bd35b3dfeb779906ed80e7614a37e876dbde9d3751b6bc56c61\": rpc error: code = NotFound desc = could not find container \"471b3e9129676bd35b3dfeb779906ed80e7614a37e876dbde9d3751b6bc56c61\": container with ID starting with 471b3e9129676bd35b3dfeb779906ed80e7614a37e876dbde9d3751b6bc56c61 not found: ID does not exist" Mar 13 11:36:42 crc kubenswrapper[4632]: I0313 11:36:42.008565 4632 scope.go:117] "RemoveContainer" containerID="f1ce3fc4d0f620c27005352e1949578082140c75a9cff811a6bb65a60520f0d9" Mar 13 11:36:42 crc kubenswrapper[4632]: E0313 11:36:42.009086 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1ce3fc4d0f620c27005352e1949578082140c75a9cff811a6bb65a60520f0d9\": container with ID starting with f1ce3fc4d0f620c27005352e1949578082140c75a9cff811a6bb65a60520f0d9 not found: ID does not exist" containerID="f1ce3fc4d0f620c27005352e1949578082140c75a9cff811a6bb65a60520f0d9" Mar 13 11:36:42 crc kubenswrapper[4632]: I0313 11:36:42.009110 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1ce3fc4d0f620c27005352e1949578082140c75a9cff811a6bb65a60520f0d9"} err="failed to get container status \"f1ce3fc4d0f620c27005352e1949578082140c75a9cff811a6bb65a60520f0d9\": rpc error: code = NotFound desc = could not find container \"f1ce3fc4d0f620c27005352e1949578082140c75a9cff811a6bb65a60520f0d9\": container with ID starting with f1ce3fc4d0f620c27005352e1949578082140c75a9cff811a6bb65a60520f0d9 not found: ID does not exist" Mar 13 11:36:42 crc kubenswrapper[4632]: I0313 11:36:42.009124 4632 scope.go:117] "RemoveContainer" containerID="6d026b8061dc2c3d9960ed047459f56a28590e7270cb04b536cdac0b85745294" Mar 13 11:36:42 crc kubenswrapper[4632]: E0313 11:36:42.009385 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d026b8061dc2c3d9960ed047459f56a28590e7270cb04b536cdac0b85745294\": container with ID starting with 6d026b8061dc2c3d9960ed047459f56a28590e7270cb04b536cdac0b85745294 not found: ID does not exist" containerID="6d026b8061dc2c3d9960ed047459f56a28590e7270cb04b536cdac0b85745294" Mar 13 11:36:42 crc kubenswrapper[4632]: I0313 11:36:42.009421 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d026b8061dc2c3d9960ed047459f56a28590e7270cb04b536cdac0b85745294"} err="failed to get container status \"6d026b8061dc2c3d9960ed047459f56a28590e7270cb04b536cdac0b85745294\": rpc error: code = NotFound desc = could not find container \"6d026b8061dc2c3d9960ed047459f56a28590e7270cb04b536cdac0b85745294\": container with ID starting with 6d026b8061dc2c3d9960ed047459f56a28590e7270cb04b536cdac0b85745294 not found: ID does not exist" Mar 13 11:36:42 crc kubenswrapper[4632]: I0313 11:36:42.057312 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb192b24-7638-4c11-9936-ad94a6842ce9" path="/var/lib/kubelet/pods/bb192b24-7638-4c11-9936-ad94a6842ce9/volumes" Mar 13 11:37:10 crc kubenswrapper[4632]: I0313 11:37:10.462073 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:37:10 crc kubenswrapper[4632]: I0313 11:37:10.462706 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:37:10 crc kubenswrapper[4632]: I0313 11:37:10.462761 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 11:37:10 crc kubenswrapper[4632]: I0313 11:37:10.482612 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b28a3031014e23a161560bdf4de3a19a21d26729102cf99acd465c2bd90c33f9"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 11:37:10 crc kubenswrapper[4632]: I0313 11:37:10.482769 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://b28a3031014e23a161560bdf4de3a19a21d26729102cf99acd465c2bd90c33f9" gracePeriod=600 Mar 13 11:37:11 crc kubenswrapper[4632]: I0313 11:37:11.246349 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="b28a3031014e23a161560bdf4de3a19a21d26729102cf99acd465c2bd90c33f9" exitCode=0 Mar 13 11:37:11 crc kubenswrapper[4632]: I0313 11:37:11.246392 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"b28a3031014e23a161560bdf4de3a19a21d26729102cf99acd465c2bd90c33f9"} Mar 13 11:37:11 crc kubenswrapper[4632]: I0313 11:37:11.246663 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6"} Mar 13 11:37:11 crc kubenswrapper[4632]: I0313 11:37:11.246683 4632 scope.go:117] "RemoveContainer" containerID="d23237a6fa10676a84af0ea53cff2f624fc9045a8ba857339106215f64dbb8e3" Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.152556 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556698-ncrvx"] Mar 13 11:38:00 crc kubenswrapper[4632]: E0313 11:38:00.153567 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb192b24-7638-4c11-9936-ad94a6842ce9" containerName="extract-utilities" Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.153586 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb192b24-7638-4c11-9936-ad94a6842ce9" containerName="extract-utilities" Mar 13 11:38:00 crc kubenswrapper[4632]: E0313 11:38:00.153614 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb192b24-7638-4c11-9936-ad94a6842ce9" containerName="registry-server" Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.153623 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb192b24-7638-4c11-9936-ad94a6842ce9" containerName="registry-server" Mar 13 11:38:00 crc kubenswrapper[4632]: E0313 11:38:00.153642 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb192b24-7638-4c11-9936-ad94a6842ce9" containerName="extract-content" Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.153649 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb192b24-7638-4c11-9936-ad94a6842ce9" containerName="extract-content" Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.153852 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb192b24-7638-4c11-9936-ad94a6842ce9" containerName="registry-server" Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.155018 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556698-ncrvx" Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.161816 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.162664 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.167480 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556698-ncrvx"] Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.172814 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.233803 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lt2x\" (UniqueName: \"kubernetes.io/projected/08ee143e-f1cf-4c38-a811-d31496082a75-kube-api-access-4lt2x\") pod \"auto-csr-approver-29556698-ncrvx\" (UID: \"08ee143e-f1cf-4c38-a811-d31496082a75\") " pod="openshift-infra/auto-csr-approver-29556698-ncrvx" Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.335453 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lt2x\" (UniqueName: \"kubernetes.io/projected/08ee143e-f1cf-4c38-a811-d31496082a75-kube-api-access-4lt2x\") pod \"auto-csr-approver-29556698-ncrvx\" (UID: \"08ee143e-f1cf-4c38-a811-d31496082a75\") " pod="openshift-infra/auto-csr-approver-29556698-ncrvx" Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.357934 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lt2x\" (UniqueName: \"kubernetes.io/projected/08ee143e-f1cf-4c38-a811-d31496082a75-kube-api-access-4lt2x\") pod \"auto-csr-approver-29556698-ncrvx\" (UID: \"08ee143e-f1cf-4c38-a811-d31496082a75\") " pod="openshift-infra/auto-csr-approver-29556698-ncrvx" Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.481139 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556698-ncrvx" Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.975176 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556698-ncrvx"] Mar 13 11:38:00 crc kubenswrapper[4632]: I0313 11:38:00.994086 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 11:38:01 crc kubenswrapper[4632]: I0313 11:38:01.006679 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556698-ncrvx" event={"ID":"08ee143e-f1cf-4c38-a811-d31496082a75","Type":"ContainerStarted","Data":"204b62853491c1d97261545b38204bb2892aa81ad3ee8d8220f0b5d0fdcd889b"} Mar 13 11:38:03 crc kubenswrapper[4632]: I0313 11:38:03.027144 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556698-ncrvx" event={"ID":"08ee143e-f1cf-4c38-a811-d31496082a75","Type":"ContainerStarted","Data":"19b28d2a56d1971c59c024b2b42655c24314722844900a0860bc74bbd0e6dfd4"} Mar 13 11:38:03 crc kubenswrapper[4632]: I0313 11:38:03.054568 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556698-ncrvx" podStartSLOduration=1.916666324 podStartE2EDuration="3.054549273s" podCreationTimestamp="2026-03-13 11:38:00 +0000 UTC" firstStartedPulling="2026-03-13 11:38:00.993825497 +0000 UTC m=+5655.016355620" lastFinishedPulling="2026-03-13 11:38:02.131708436 +0000 UTC m=+5656.154238569" observedRunningTime="2026-03-13 11:38:03.054138614 +0000 UTC m=+5657.076668757" watchObservedRunningTime="2026-03-13 11:38:03.054549273 +0000 UTC m=+5657.077079406" Mar 13 11:38:05 crc kubenswrapper[4632]: I0313 11:38:05.057893 4632 generic.go:334] "Generic (PLEG): container finished" podID="08ee143e-f1cf-4c38-a811-d31496082a75" containerID="19b28d2a56d1971c59c024b2b42655c24314722844900a0860bc74bbd0e6dfd4" exitCode=0 Mar 13 11:38:05 crc kubenswrapper[4632]: I0313 11:38:05.058006 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556698-ncrvx" event={"ID":"08ee143e-f1cf-4c38-a811-d31496082a75","Type":"ContainerDied","Data":"19b28d2a56d1971c59c024b2b42655c24314722844900a0860bc74bbd0e6dfd4"} Mar 13 11:38:06 crc kubenswrapper[4632]: I0313 11:38:06.487054 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556698-ncrvx" Mar 13 11:38:06 crc kubenswrapper[4632]: I0313 11:38:06.562824 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4lt2x\" (UniqueName: \"kubernetes.io/projected/08ee143e-f1cf-4c38-a811-d31496082a75-kube-api-access-4lt2x\") pod \"08ee143e-f1cf-4c38-a811-d31496082a75\" (UID: \"08ee143e-f1cf-4c38-a811-d31496082a75\") " Mar 13 11:38:06 crc kubenswrapper[4632]: I0313 11:38:06.583271 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08ee143e-f1cf-4c38-a811-d31496082a75-kube-api-access-4lt2x" (OuterVolumeSpecName: "kube-api-access-4lt2x") pod "08ee143e-f1cf-4c38-a811-d31496082a75" (UID: "08ee143e-f1cf-4c38-a811-d31496082a75"). InnerVolumeSpecName "kube-api-access-4lt2x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:38:06 crc kubenswrapper[4632]: I0313 11:38:06.666172 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4lt2x\" (UniqueName: \"kubernetes.io/projected/08ee143e-f1cf-4c38-a811-d31496082a75-kube-api-access-4lt2x\") on node \"crc\" DevicePath \"\"" Mar 13 11:38:07 crc kubenswrapper[4632]: I0313 11:38:07.089637 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556698-ncrvx" event={"ID":"08ee143e-f1cf-4c38-a811-d31496082a75","Type":"ContainerDied","Data":"204b62853491c1d97261545b38204bb2892aa81ad3ee8d8220f0b5d0fdcd889b"} Mar 13 11:38:07 crc kubenswrapper[4632]: I0313 11:38:07.089902 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556698-ncrvx" Mar 13 11:38:07 crc kubenswrapper[4632]: I0313 11:38:07.089909 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="204b62853491c1d97261545b38204bb2892aa81ad3ee8d8220f0b5d0fdcd889b" Mar 13 11:38:07 crc kubenswrapper[4632]: I0313 11:38:07.166426 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556692-4kd7v"] Mar 13 11:38:07 crc kubenswrapper[4632]: I0313 11:38:07.177705 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556692-4kd7v"] Mar 13 11:38:08 crc kubenswrapper[4632]: I0313 11:38:08.058194 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="77ae44b0-e101-4d21-87e5-9e213e024e9e" path="/var/lib/kubelet/pods/77ae44b0-e101-4d21-87e5-9e213e024e9e/volumes" Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.031426 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-c87c2"] Mar 13 11:38:17 crc kubenswrapper[4632]: E0313 11:38:17.040669 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08ee143e-f1cf-4c38-a811-d31496082a75" containerName="oc" Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.040711 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="08ee143e-f1cf-4c38-a811-d31496082a75" containerName="oc" Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.040885 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="08ee143e-f1cf-4c38-a811-d31496082a75" containerName="oc" Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.042841 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.045557 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c87c2"] Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.188466 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bbc76a6-d812-41c7-a63b-09f6fdb37405-catalog-content\") pod \"community-operators-c87c2\" (UID: \"7bbc76a6-d812-41c7-a63b-09f6fdb37405\") " pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.190483 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntrdr\" (UniqueName: \"kubernetes.io/projected/7bbc76a6-d812-41c7-a63b-09f6fdb37405-kube-api-access-ntrdr\") pod \"community-operators-c87c2\" (UID: \"7bbc76a6-d812-41c7-a63b-09f6fdb37405\") " pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.190792 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bbc76a6-d812-41c7-a63b-09f6fdb37405-utilities\") pod \"community-operators-c87c2\" (UID: \"7bbc76a6-d812-41c7-a63b-09f6fdb37405\") " pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.292873 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntrdr\" (UniqueName: \"kubernetes.io/projected/7bbc76a6-d812-41c7-a63b-09f6fdb37405-kube-api-access-ntrdr\") pod \"community-operators-c87c2\" (UID: \"7bbc76a6-d812-41c7-a63b-09f6fdb37405\") " pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.293382 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bbc76a6-d812-41c7-a63b-09f6fdb37405-utilities\") pod \"community-operators-c87c2\" (UID: \"7bbc76a6-d812-41c7-a63b-09f6fdb37405\") " pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.293733 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bbc76a6-d812-41c7-a63b-09f6fdb37405-catalog-content\") pod \"community-operators-c87c2\" (UID: \"7bbc76a6-d812-41c7-a63b-09f6fdb37405\") " pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.294520 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7bbc76a6-d812-41c7-a63b-09f6fdb37405-utilities\") pod \"community-operators-c87c2\" (UID: \"7bbc76a6-d812-41c7-a63b-09f6fdb37405\") " pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.294683 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7bbc76a6-d812-41c7-a63b-09f6fdb37405-catalog-content\") pod \"community-operators-c87c2\" (UID: \"7bbc76a6-d812-41c7-a63b-09f6fdb37405\") " pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.312142 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntrdr\" (UniqueName: \"kubernetes.io/projected/7bbc76a6-d812-41c7-a63b-09f6fdb37405-kube-api-access-ntrdr\") pod \"community-operators-c87c2\" (UID: \"7bbc76a6-d812-41c7-a63b-09f6fdb37405\") " pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:17 crc kubenswrapper[4632]: I0313 11:38:17.393460 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:18 crc kubenswrapper[4632]: I0313 11:38:18.134505 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c87c2"] Mar 13 11:38:18 crc kubenswrapper[4632]: I0313 11:38:18.200573 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c87c2" event={"ID":"7bbc76a6-d812-41c7-a63b-09f6fdb37405","Type":"ContainerStarted","Data":"8b195e74d7a07f4a43f98dd358110c2996bc32ef4665c1a538b5616819d59f0b"} Mar 13 11:38:19 crc kubenswrapper[4632]: I0313 11:38:19.224758 4632 generic.go:334] "Generic (PLEG): container finished" podID="7bbc76a6-d812-41c7-a63b-09f6fdb37405" containerID="29b9253bb98b4a2906b8867d032633c0a6084d2ee5611106305860f394c9da23" exitCode=0 Mar 13 11:38:19 crc kubenswrapper[4632]: I0313 11:38:19.224801 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c87c2" event={"ID":"7bbc76a6-d812-41c7-a63b-09f6fdb37405","Type":"ContainerDied","Data":"29b9253bb98b4a2906b8867d032633c0a6084d2ee5611106305860f394c9da23"} Mar 13 11:38:21 crc kubenswrapper[4632]: I0313 11:38:21.235912 4632 scope.go:117] "RemoveContainer" containerID="b8f22a62b885c530c5401c31b24e17ea8bcf63d6debf02e44b4f01a4ab4c1102" Mar 13 11:38:27 crc kubenswrapper[4632]: I0313 11:38:27.299649 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c87c2" event={"ID":"7bbc76a6-d812-41c7-a63b-09f6fdb37405","Type":"ContainerStarted","Data":"90efdae9ee51f26b819dfe72dc997f0e8e40c177d9f2a50823b81b5fdbf64e1e"} Mar 13 11:38:28 crc kubenswrapper[4632]: I0313 11:38:28.311488 4632 generic.go:334] "Generic (PLEG): container finished" podID="7bbc76a6-d812-41c7-a63b-09f6fdb37405" containerID="90efdae9ee51f26b819dfe72dc997f0e8e40c177d9f2a50823b81b5fdbf64e1e" exitCode=0 Mar 13 11:38:28 crc kubenswrapper[4632]: I0313 11:38:28.311544 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c87c2" event={"ID":"7bbc76a6-d812-41c7-a63b-09f6fdb37405","Type":"ContainerDied","Data":"90efdae9ee51f26b819dfe72dc997f0e8e40c177d9f2a50823b81b5fdbf64e1e"} Mar 13 11:38:29 crc kubenswrapper[4632]: I0313 11:38:29.322395 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-c87c2" event={"ID":"7bbc76a6-d812-41c7-a63b-09f6fdb37405","Type":"ContainerStarted","Data":"3e4f69b8253bf86d8b9955ae26ef60a4e60f02816ad73b2bf8a731285c5e7153"} Mar 13 11:38:29 crc kubenswrapper[4632]: I0313 11:38:29.342932 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-c87c2" podStartSLOduration=2.777515627 podStartE2EDuration="12.342913268s" podCreationTimestamp="2026-03-13 11:38:17 +0000 UTC" firstStartedPulling="2026-03-13 11:38:19.226827598 +0000 UTC m=+5673.249357741" lastFinishedPulling="2026-03-13 11:38:28.792225249 +0000 UTC m=+5682.814755382" observedRunningTime="2026-03-13 11:38:29.342278172 +0000 UTC m=+5683.364808315" watchObservedRunningTime="2026-03-13 11:38:29.342913268 +0000 UTC m=+5683.365443401" Mar 13 11:38:37 crc kubenswrapper[4632]: I0313 11:38:37.393928 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:37 crc kubenswrapper[4632]: I0313 11:38:37.395577 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:37 crc kubenswrapper[4632]: I0313 11:38:37.448023 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:38 crc kubenswrapper[4632]: I0313 11:38:38.455166 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-c87c2" Mar 13 11:38:38 crc kubenswrapper[4632]: I0313 11:38:38.655837 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-c87c2"] Mar 13 11:38:38 crc kubenswrapper[4632]: I0313 11:38:38.711800 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-774lb"] Mar 13 11:38:38 crc kubenswrapper[4632]: I0313 11:38:38.716774 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-774lb" podUID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerName="registry-server" containerID="cri-o://1b995d3ea46318dbc1da1ae83e15d1a1943f08993ba4772ae9cb4b946ae10e86" gracePeriod=2 Mar 13 11:38:39 crc kubenswrapper[4632]: I0313 11:38:39.412904 4632 generic.go:334] "Generic (PLEG): container finished" podID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerID="1b995d3ea46318dbc1da1ae83e15d1a1943f08993ba4772ae9cb4b946ae10e86" exitCode=0 Mar 13 11:38:39 crc kubenswrapper[4632]: I0313 11:38:39.412978 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-774lb" event={"ID":"560629a7-9dec-4eb7-8c73-a8f097293daa","Type":"ContainerDied","Data":"1b995d3ea46318dbc1da1ae83e15d1a1943f08993ba4772ae9cb4b946ae10e86"} Mar 13 11:38:39 crc kubenswrapper[4632]: I0313 11:38:39.922503 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-774lb" Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.091111 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rc4f5\" (UniqueName: \"kubernetes.io/projected/560629a7-9dec-4eb7-8c73-a8f097293daa-kube-api-access-rc4f5\") pod \"560629a7-9dec-4eb7-8c73-a8f097293daa\" (UID: \"560629a7-9dec-4eb7-8c73-a8f097293daa\") " Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.091473 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/560629a7-9dec-4eb7-8c73-a8f097293daa-catalog-content\") pod \"560629a7-9dec-4eb7-8c73-a8f097293daa\" (UID: \"560629a7-9dec-4eb7-8c73-a8f097293daa\") " Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.091845 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/560629a7-9dec-4eb7-8c73-a8f097293daa-utilities\") pod \"560629a7-9dec-4eb7-8c73-a8f097293daa\" (UID: \"560629a7-9dec-4eb7-8c73-a8f097293daa\") " Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.093655 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/560629a7-9dec-4eb7-8c73-a8f097293daa-utilities" (OuterVolumeSpecName: "utilities") pod "560629a7-9dec-4eb7-8c73-a8f097293daa" (UID: "560629a7-9dec-4eb7-8c73-a8f097293daa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.119718 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/560629a7-9dec-4eb7-8c73-a8f097293daa-kube-api-access-rc4f5" (OuterVolumeSpecName: "kube-api-access-rc4f5") pod "560629a7-9dec-4eb7-8c73-a8f097293daa" (UID: "560629a7-9dec-4eb7-8c73-a8f097293daa"). InnerVolumeSpecName "kube-api-access-rc4f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.193797 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/560629a7-9dec-4eb7-8c73-a8f097293daa-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.193832 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rc4f5\" (UniqueName: \"kubernetes.io/projected/560629a7-9dec-4eb7-8c73-a8f097293daa-kube-api-access-rc4f5\") on node \"crc\" DevicePath \"\"" Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.304275 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/560629a7-9dec-4eb7-8c73-a8f097293daa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "560629a7-9dec-4eb7-8c73-a8f097293daa" (UID: "560629a7-9dec-4eb7-8c73-a8f097293daa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.397500 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/560629a7-9dec-4eb7-8c73-a8f097293daa-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.425341 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-774lb" event={"ID":"560629a7-9dec-4eb7-8c73-a8f097293daa","Type":"ContainerDied","Data":"4ae9494de264dfa5dcfb2c9e6166d64886aa8f640f54445b6eadb498ad356c8c"} Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.425409 4632 scope.go:117] "RemoveContainer" containerID="1b995d3ea46318dbc1da1ae83e15d1a1943f08993ba4772ae9cb4b946ae10e86" Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.426146 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-774lb" Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.464101 4632 scope.go:117] "RemoveContainer" containerID="d5e77ef64ff23f92ed48258b81d7d0310ada291a691626009608a75068a59888" Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.508206 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-774lb"] Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.514114 4632 scope.go:117] "RemoveContainer" containerID="f9fbff406d14d8da11f86810a1b1b035215dd5c6179ac20e5ddd29194bd3f5d6" Mar 13 11:38:40 crc kubenswrapper[4632]: I0313 11:38:40.530849 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-774lb"] Mar 13 11:38:42 crc kubenswrapper[4632]: I0313 11:38:42.062588 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="560629a7-9dec-4eb7-8c73-a8f097293daa" path="/var/lib/kubelet/pods/560629a7-9dec-4eb7-8c73-a8f097293daa/volumes" Mar 13 11:39:10 crc kubenswrapper[4632]: I0313 11:39:10.461164 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:39:10 crc kubenswrapper[4632]: I0313 11:39:10.461686 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:39:40 crc kubenswrapper[4632]: I0313 11:39:40.460700 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:39:40 crc kubenswrapper[4632]: I0313 11:39:40.461120 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.166741 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556700-rjhgg"] Mar 13 11:40:00 crc kubenswrapper[4632]: E0313 11:40:00.168041 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerName="extract-content" Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.168066 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerName="extract-content" Mar 13 11:40:00 crc kubenswrapper[4632]: E0313 11:40:00.168100 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerName="extract-utilities" Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.168112 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerName="extract-utilities" Mar 13 11:40:00 crc kubenswrapper[4632]: E0313 11:40:00.168165 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerName="registry-server" Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.168178 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerName="registry-server" Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.168582 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="560629a7-9dec-4eb7-8c73-a8f097293daa" containerName="registry-server" Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.169570 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556700-rjhgg" Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.172144 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.177813 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.179706 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.209921 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556700-rjhgg"] Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.232572 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrnzz\" (UniqueName: \"kubernetes.io/projected/d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5-kube-api-access-xrnzz\") pod \"auto-csr-approver-29556700-rjhgg\" (UID: \"d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5\") " pod="openshift-infra/auto-csr-approver-29556700-rjhgg" Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.335539 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrnzz\" (UniqueName: \"kubernetes.io/projected/d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5-kube-api-access-xrnzz\") pod \"auto-csr-approver-29556700-rjhgg\" (UID: \"d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5\") " pod="openshift-infra/auto-csr-approver-29556700-rjhgg" Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.356400 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrnzz\" (UniqueName: \"kubernetes.io/projected/d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5-kube-api-access-xrnzz\") pod \"auto-csr-approver-29556700-rjhgg\" (UID: \"d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5\") " pod="openshift-infra/auto-csr-approver-29556700-rjhgg" Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.488931 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556700-rjhgg" Mar 13 11:40:00 crc kubenswrapper[4632]: I0313 11:40:00.978923 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556700-rjhgg"] Mar 13 11:40:01 crc kubenswrapper[4632]: I0313 11:40:01.213352 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556700-rjhgg" event={"ID":"d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5","Type":"ContainerStarted","Data":"ffa8ab9e687203b969b410a56351f55d13c66e081d5375016372e44985afa4a7"} Mar 13 11:40:03 crc kubenswrapper[4632]: I0313 11:40:03.231885 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556700-rjhgg" event={"ID":"d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5","Type":"ContainerStarted","Data":"e2dc32a91f84dbc41f05188f09b5ec2790c5d429e256411e1defae69e3e43deb"} Mar 13 11:40:03 crc kubenswrapper[4632]: I0313 11:40:03.252989 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556700-rjhgg" podStartSLOduration=1.854841785 podStartE2EDuration="3.252931148s" podCreationTimestamp="2026-03-13 11:40:00 +0000 UTC" firstStartedPulling="2026-03-13 11:40:00.985329014 +0000 UTC m=+5775.007859147" lastFinishedPulling="2026-03-13 11:40:02.383418377 +0000 UTC m=+5776.405948510" observedRunningTime="2026-03-13 11:40:03.244716225 +0000 UTC m=+5777.267246358" watchObservedRunningTime="2026-03-13 11:40:03.252931148 +0000 UTC m=+5777.275461281" Mar 13 11:40:05 crc kubenswrapper[4632]: I0313 11:40:05.251445 4632 generic.go:334] "Generic (PLEG): container finished" podID="d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5" containerID="e2dc32a91f84dbc41f05188f09b5ec2790c5d429e256411e1defae69e3e43deb" exitCode=0 Mar 13 11:40:05 crc kubenswrapper[4632]: I0313 11:40:05.251668 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556700-rjhgg" event={"ID":"d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5","Type":"ContainerDied","Data":"e2dc32a91f84dbc41f05188f09b5ec2790c5d429e256411e1defae69e3e43deb"} Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.281401 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-flb98"] Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.286382 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.316469 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-flb98"] Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.360772 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz4bb\" (UniqueName: \"kubernetes.io/projected/dda62ea0-a8e2-46bb-a080-1b771e45feec-kube-api-access-wz4bb\") pod \"redhat-operators-flb98\" (UID: \"dda62ea0-a8e2-46bb-a080-1b771e45feec\") " pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.360860 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dda62ea0-a8e2-46bb-a080-1b771e45feec-utilities\") pod \"redhat-operators-flb98\" (UID: \"dda62ea0-a8e2-46bb-a080-1b771e45feec\") " pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.360951 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dda62ea0-a8e2-46bb-a080-1b771e45feec-catalog-content\") pod \"redhat-operators-flb98\" (UID: \"dda62ea0-a8e2-46bb-a080-1b771e45feec\") " pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.463762 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz4bb\" (UniqueName: \"kubernetes.io/projected/dda62ea0-a8e2-46bb-a080-1b771e45feec-kube-api-access-wz4bb\") pod \"redhat-operators-flb98\" (UID: \"dda62ea0-a8e2-46bb-a080-1b771e45feec\") " pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.463971 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dda62ea0-a8e2-46bb-a080-1b771e45feec-utilities\") pod \"redhat-operators-flb98\" (UID: \"dda62ea0-a8e2-46bb-a080-1b771e45feec\") " pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.464787 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dda62ea0-a8e2-46bb-a080-1b771e45feec-utilities\") pod \"redhat-operators-flb98\" (UID: \"dda62ea0-a8e2-46bb-a080-1b771e45feec\") " pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.464836 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dda62ea0-a8e2-46bb-a080-1b771e45feec-catalog-content\") pod \"redhat-operators-flb98\" (UID: \"dda62ea0-a8e2-46bb-a080-1b771e45feec\") " pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.464851 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dda62ea0-a8e2-46bb-a080-1b771e45feec-catalog-content\") pod \"redhat-operators-flb98\" (UID: \"dda62ea0-a8e2-46bb-a080-1b771e45feec\") " pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.482754 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz4bb\" (UniqueName: \"kubernetes.io/projected/dda62ea0-a8e2-46bb-a080-1b771e45feec-kube-api-access-wz4bb\") pod \"redhat-operators-flb98\" (UID: \"dda62ea0-a8e2-46bb-a080-1b771e45feec\") " pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.618363 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.857362 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556700-rjhgg" Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.980105 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrnzz\" (UniqueName: \"kubernetes.io/projected/d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5-kube-api-access-xrnzz\") pod \"d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5\" (UID: \"d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5\") " Mar 13 11:40:06 crc kubenswrapper[4632]: I0313 11:40:06.986610 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5-kube-api-access-xrnzz" (OuterVolumeSpecName: "kube-api-access-xrnzz") pod "d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5" (UID: "d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5"). InnerVolumeSpecName "kube-api-access-xrnzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:40:07 crc kubenswrapper[4632]: I0313 11:40:07.083398 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrnzz\" (UniqueName: \"kubernetes.io/projected/d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5-kube-api-access-xrnzz\") on node \"crc\" DevicePath \"\"" Mar 13 11:40:07 crc kubenswrapper[4632]: I0313 11:40:07.270726 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556700-rjhgg" event={"ID":"d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5","Type":"ContainerDied","Data":"ffa8ab9e687203b969b410a56351f55d13c66e081d5375016372e44985afa4a7"} Mar 13 11:40:07 crc kubenswrapper[4632]: I0313 11:40:07.270784 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffa8ab9e687203b969b410a56351f55d13c66e081d5375016372e44985afa4a7" Mar 13 11:40:07 crc kubenswrapper[4632]: I0313 11:40:07.270782 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556700-rjhgg" Mar 13 11:40:07 crc kubenswrapper[4632]: I0313 11:40:07.361324 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-flb98"] Mar 13 11:40:07 crc kubenswrapper[4632]: W0313 11:40:07.373762 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddda62ea0_a8e2_46bb_a080_1b771e45feec.slice/crio-bc8f50d20d6be8ef47a9be50fca7331e6e4176f53dc58080b0e546eb62f4a959 WatchSource:0}: Error finding container bc8f50d20d6be8ef47a9be50fca7331e6e4176f53dc58080b0e546eb62f4a959: Status 404 returned error can't find the container with id bc8f50d20d6be8ef47a9be50fca7331e6e4176f53dc58080b0e546eb62f4a959 Mar 13 11:40:07 crc kubenswrapper[4632]: I0313 11:40:07.374532 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556694-jmf5t"] Mar 13 11:40:07 crc kubenswrapper[4632]: I0313 11:40:07.420693 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556694-jmf5t"] Mar 13 11:40:08 crc kubenswrapper[4632]: I0313 11:40:08.060575 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="266f7f6e-de91-4256-8605-0a71adef85fc" path="/var/lib/kubelet/pods/266f7f6e-de91-4256-8605-0a71adef85fc/volumes" Mar 13 11:40:08 crc kubenswrapper[4632]: I0313 11:40:08.281680 4632 generic.go:334] "Generic (PLEG): container finished" podID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerID="f8176fa837f25c1697cac7d62b754ee2cfecbadb578ca4e441339c0f0c3b11c5" exitCode=0 Mar 13 11:40:08 crc kubenswrapper[4632]: I0313 11:40:08.281731 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-flb98" event={"ID":"dda62ea0-a8e2-46bb-a080-1b771e45feec","Type":"ContainerDied","Data":"f8176fa837f25c1697cac7d62b754ee2cfecbadb578ca4e441339c0f0c3b11c5"} Mar 13 11:40:08 crc kubenswrapper[4632]: I0313 11:40:08.281780 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-flb98" event={"ID":"dda62ea0-a8e2-46bb-a080-1b771e45feec","Type":"ContainerStarted","Data":"bc8f50d20d6be8ef47a9be50fca7331e6e4176f53dc58080b0e546eb62f4a959"} Mar 13 11:40:09 crc kubenswrapper[4632]: I0313 11:40:09.292150 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-flb98" event={"ID":"dda62ea0-a8e2-46bb-a080-1b771e45feec","Type":"ContainerStarted","Data":"879dd7d998eed4fd26b7e2d5e411da489d9eceae2f429769ea2bcc4ae3aaf24d"} Mar 13 11:40:10 crc kubenswrapper[4632]: I0313 11:40:10.462696 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:40:10 crc kubenswrapper[4632]: I0313 11:40:10.463175 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:40:10 crc kubenswrapper[4632]: I0313 11:40:10.463229 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 11:40:10 crc kubenswrapper[4632]: I0313 11:40:10.464071 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 11:40:10 crc kubenswrapper[4632]: I0313 11:40:10.464153 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" gracePeriod=600 Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.082511 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kc89d"] Mar 13 11:40:11 crc kubenswrapper[4632]: E0313 11:40:11.083326 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5" containerName="oc" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.083343 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5" containerName="oc" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.083533 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5" containerName="oc" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.084843 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.104038 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kc89d"] Mar 13 11:40:11 crc kubenswrapper[4632]: E0313 11:40:11.162115 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.169392 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ad0d32e-2227-4d12-bf62-19eb24597391-utilities\") pod \"certified-operators-kc89d\" (UID: \"2ad0d32e-2227-4d12-bf62-19eb24597391\") " pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.169562 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbnzm\" (UniqueName: \"kubernetes.io/projected/2ad0d32e-2227-4d12-bf62-19eb24597391-kube-api-access-cbnzm\") pod \"certified-operators-kc89d\" (UID: \"2ad0d32e-2227-4d12-bf62-19eb24597391\") " pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.169735 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ad0d32e-2227-4d12-bf62-19eb24597391-catalog-content\") pod \"certified-operators-kc89d\" (UID: \"2ad0d32e-2227-4d12-bf62-19eb24597391\") " pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.271727 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ad0d32e-2227-4d12-bf62-19eb24597391-utilities\") pod \"certified-operators-kc89d\" (UID: \"2ad0d32e-2227-4d12-bf62-19eb24597391\") " pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.271824 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbnzm\" (UniqueName: \"kubernetes.io/projected/2ad0d32e-2227-4d12-bf62-19eb24597391-kube-api-access-cbnzm\") pod \"certified-operators-kc89d\" (UID: \"2ad0d32e-2227-4d12-bf62-19eb24597391\") " pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.271923 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ad0d32e-2227-4d12-bf62-19eb24597391-catalog-content\") pod \"certified-operators-kc89d\" (UID: \"2ad0d32e-2227-4d12-bf62-19eb24597391\") " pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.272863 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ad0d32e-2227-4d12-bf62-19eb24597391-catalog-content\") pod \"certified-operators-kc89d\" (UID: \"2ad0d32e-2227-4d12-bf62-19eb24597391\") " pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.272909 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ad0d32e-2227-4d12-bf62-19eb24597391-utilities\") pod \"certified-operators-kc89d\" (UID: \"2ad0d32e-2227-4d12-bf62-19eb24597391\") " pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.294848 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbnzm\" (UniqueName: \"kubernetes.io/projected/2ad0d32e-2227-4d12-bf62-19eb24597391-kube-api-access-cbnzm\") pod \"certified-operators-kc89d\" (UID: \"2ad0d32e-2227-4d12-bf62-19eb24597391\") " pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.310332 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" exitCode=0 Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.310382 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6"} Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.310424 4632 scope.go:117] "RemoveContainer" containerID="b28a3031014e23a161560bdf4de3a19a21d26729102cf99acd465c2bd90c33f9" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.311107 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:40:11 crc kubenswrapper[4632]: E0313 11:40:11.311622 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.406451 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:11 crc kubenswrapper[4632]: I0313 11:40:11.899397 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kc89d"] Mar 13 11:40:12 crc kubenswrapper[4632]: I0313 11:40:12.323838 4632 generic.go:334] "Generic (PLEG): container finished" podID="2ad0d32e-2227-4d12-bf62-19eb24597391" containerID="4f578f7aba25fc35c2994f8884e7599a303084b2db092f8a745f47b140231631" exitCode=0 Mar 13 11:40:12 crc kubenswrapper[4632]: I0313 11:40:12.323905 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kc89d" event={"ID":"2ad0d32e-2227-4d12-bf62-19eb24597391","Type":"ContainerDied","Data":"4f578f7aba25fc35c2994f8884e7599a303084b2db092f8a745f47b140231631"} Mar 13 11:40:12 crc kubenswrapper[4632]: I0313 11:40:12.323984 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kc89d" event={"ID":"2ad0d32e-2227-4d12-bf62-19eb24597391","Type":"ContainerStarted","Data":"7f8725f4be963837d7f72aed20800307d623dcf5ab9a99a449d6c1c7ef19f63e"} Mar 13 11:40:14 crc kubenswrapper[4632]: I0313 11:40:14.345324 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kc89d" event={"ID":"2ad0d32e-2227-4d12-bf62-19eb24597391","Type":"ContainerStarted","Data":"38648afa2577e5a9fb5d9d7e00b5f9e414dd8af022764679faa186da16c72b26"} Mar 13 11:40:19 crc kubenswrapper[4632]: I0313 11:40:19.396349 4632 generic.go:334] "Generic (PLEG): container finished" podID="2ad0d32e-2227-4d12-bf62-19eb24597391" containerID="38648afa2577e5a9fb5d9d7e00b5f9e414dd8af022764679faa186da16c72b26" exitCode=0 Mar 13 11:40:19 crc kubenswrapper[4632]: I0313 11:40:19.396413 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kc89d" event={"ID":"2ad0d32e-2227-4d12-bf62-19eb24597391","Type":"ContainerDied","Data":"38648afa2577e5a9fb5d9d7e00b5f9e414dd8af022764679faa186da16c72b26"} Mar 13 11:40:21 crc kubenswrapper[4632]: I0313 11:40:21.437979 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kc89d" event={"ID":"2ad0d32e-2227-4d12-bf62-19eb24597391","Type":"ContainerStarted","Data":"9d9bde8c088f7036968dfb6799f01abee05cb617fe7dd0b877a31e0a9fec56bb"} Mar 13 11:40:21 crc kubenswrapper[4632]: I0313 11:40:21.442558 4632 generic.go:334] "Generic (PLEG): container finished" podID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerID="879dd7d998eed4fd26b7e2d5e411da489d9eceae2f429769ea2bcc4ae3aaf24d" exitCode=0 Mar 13 11:40:21 crc kubenswrapper[4632]: I0313 11:40:21.442615 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-flb98" event={"ID":"dda62ea0-a8e2-46bb-a080-1b771e45feec","Type":"ContainerDied","Data":"879dd7d998eed4fd26b7e2d5e411da489d9eceae2f429769ea2bcc4ae3aaf24d"} Mar 13 11:40:21 crc kubenswrapper[4632]: I0313 11:40:21.488261 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kc89d" podStartSLOduration=2.65704836 podStartE2EDuration="10.488219566s" podCreationTimestamp="2026-03-13 11:40:11 +0000 UTC" firstStartedPulling="2026-03-13 11:40:12.326371451 +0000 UTC m=+5786.348901594" lastFinishedPulling="2026-03-13 11:40:20.157542667 +0000 UTC m=+5794.180072800" observedRunningTime="2026-03-13 11:40:21.467766691 +0000 UTC m=+5795.490296844" watchObservedRunningTime="2026-03-13 11:40:21.488219566 +0000 UTC m=+5795.510749719" Mar 13 11:40:21 crc kubenswrapper[4632]: I0313 11:40:21.709604 4632 scope.go:117] "RemoveContainer" containerID="dd29187096f712bf2f18fa46086683fcb900aea6c3d89212b78286e73075a17b" Mar 13 11:40:23 crc kubenswrapper[4632]: I0313 11:40:23.044797 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:40:23 crc kubenswrapper[4632]: E0313 11:40:23.045469 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:40:23 crc kubenswrapper[4632]: I0313 11:40:23.462735 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-flb98" event={"ID":"dda62ea0-a8e2-46bb-a080-1b771e45feec","Type":"ContainerStarted","Data":"902661c18f068b146e4aecda28b4265925f2ef61195dd89d1bcb5ed19cf93dd6"} Mar 13 11:40:23 crc kubenswrapper[4632]: I0313 11:40:23.486131 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-flb98" podStartSLOduration=3.603650438 podStartE2EDuration="17.48611495s" podCreationTimestamp="2026-03-13 11:40:06 +0000 UTC" firstStartedPulling="2026-03-13 11:40:08.283971682 +0000 UTC m=+5782.306501815" lastFinishedPulling="2026-03-13 11:40:22.166436194 +0000 UTC m=+5796.188966327" observedRunningTime="2026-03-13 11:40:23.48567673 +0000 UTC m=+5797.508206883" watchObservedRunningTime="2026-03-13 11:40:23.48611495 +0000 UTC m=+5797.508645083" Mar 13 11:40:26 crc kubenswrapper[4632]: I0313 11:40:26.618604 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:40:26 crc kubenswrapper[4632]: I0313 11:40:26.619141 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:40:27 crc kubenswrapper[4632]: I0313 11:40:27.677866 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-flb98" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerName="registry-server" probeResult="failure" output=< Mar 13 11:40:27 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:40:27 crc kubenswrapper[4632]: > Mar 13 11:40:31 crc kubenswrapper[4632]: I0313 11:40:31.407580 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:31 crc kubenswrapper[4632]: I0313 11:40:31.408300 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:32 crc kubenswrapper[4632]: I0313 11:40:32.459577 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-kc89d" podUID="2ad0d32e-2227-4d12-bf62-19eb24597391" containerName="registry-server" probeResult="failure" output=< Mar 13 11:40:32 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:40:32 crc kubenswrapper[4632]: > Mar 13 11:40:36 crc kubenswrapper[4632]: I0313 11:40:36.044453 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:40:36 crc kubenswrapper[4632]: E0313 11:40:36.045373 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:40:37 crc kubenswrapper[4632]: I0313 11:40:37.665380 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-flb98" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerName="registry-server" probeResult="failure" output=< Mar 13 11:40:37 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:40:37 crc kubenswrapper[4632]: > Mar 13 11:40:42 crc kubenswrapper[4632]: I0313 11:40:42.466876 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-kc89d" podUID="2ad0d32e-2227-4d12-bf62-19eb24597391" containerName="registry-server" probeResult="failure" output=< Mar 13 11:40:42 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:40:42 crc kubenswrapper[4632]: > Mar 13 11:40:47 crc kubenswrapper[4632]: I0313 11:40:47.666537 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-flb98" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerName="registry-server" probeResult="failure" output=< Mar 13 11:40:47 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:40:47 crc kubenswrapper[4632]: > Mar 13 11:40:49 crc kubenswrapper[4632]: I0313 11:40:49.044796 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:40:49 crc kubenswrapper[4632]: E0313 11:40:49.045622 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:40:51 crc kubenswrapper[4632]: I0313 11:40:51.524963 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:51 crc kubenswrapper[4632]: I0313 11:40:51.608864 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:51 crc kubenswrapper[4632]: I0313 11:40:51.775625 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kc89d"] Mar 13 11:40:52 crc kubenswrapper[4632]: I0313 11:40:52.765227 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-kc89d" podUID="2ad0d32e-2227-4d12-bf62-19eb24597391" containerName="registry-server" containerID="cri-o://9d9bde8c088f7036968dfb6799f01abee05cb617fe7dd0b877a31e0a9fec56bb" gracePeriod=2 Mar 13 11:40:53 crc kubenswrapper[4632]: I0313 11:40:53.777830 4632 generic.go:334] "Generic (PLEG): container finished" podID="2ad0d32e-2227-4d12-bf62-19eb24597391" containerID="9d9bde8c088f7036968dfb6799f01abee05cb617fe7dd0b877a31e0a9fec56bb" exitCode=0 Mar 13 11:40:53 crc kubenswrapper[4632]: I0313 11:40:53.778142 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kc89d" event={"ID":"2ad0d32e-2227-4d12-bf62-19eb24597391","Type":"ContainerDied","Data":"9d9bde8c088f7036968dfb6799f01abee05cb617fe7dd0b877a31e0a9fec56bb"} Mar 13 11:40:53 crc kubenswrapper[4632]: I0313 11:40:53.778174 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kc89d" event={"ID":"2ad0d32e-2227-4d12-bf62-19eb24597391","Type":"ContainerDied","Data":"7f8725f4be963837d7f72aed20800307d623dcf5ab9a99a449d6c1c7ef19f63e"} Mar 13 11:40:53 crc kubenswrapper[4632]: I0313 11:40:53.778188 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f8725f4be963837d7f72aed20800307d623dcf5ab9a99a449d6c1c7ef19f63e" Mar 13 11:40:53 crc kubenswrapper[4632]: I0313 11:40:53.824360 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:53 crc kubenswrapper[4632]: I0313 11:40:53.859347 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ad0d32e-2227-4d12-bf62-19eb24597391-catalog-content\") pod \"2ad0d32e-2227-4d12-bf62-19eb24597391\" (UID: \"2ad0d32e-2227-4d12-bf62-19eb24597391\") " Mar 13 11:40:53 crc kubenswrapper[4632]: I0313 11:40:53.859665 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ad0d32e-2227-4d12-bf62-19eb24597391-utilities\") pod \"2ad0d32e-2227-4d12-bf62-19eb24597391\" (UID: \"2ad0d32e-2227-4d12-bf62-19eb24597391\") " Mar 13 11:40:53 crc kubenswrapper[4632]: I0313 11:40:53.859756 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbnzm\" (UniqueName: \"kubernetes.io/projected/2ad0d32e-2227-4d12-bf62-19eb24597391-kube-api-access-cbnzm\") pod \"2ad0d32e-2227-4d12-bf62-19eb24597391\" (UID: \"2ad0d32e-2227-4d12-bf62-19eb24597391\") " Mar 13 11:40:53 crc kubenswrapper[4632]: I0313 11:40:53.861490 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ad0d32e-2227-4d12-bf62-19eb24597391-utilities" (OuterVolumeSpecName: "utilities") pod "2ad0d32e-2227-4d12-bf62-19eb24597391" (UID: "2ad0d32e-2227-4d12-bf62-19eb24597391"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:40:53 crc kubenswrapper[4632]: I0313 11:40:53.885064 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ad0d32e-2227-4d12-bf62-19eb24597391-kube-api-access-cbnzm" (OuterVolumeSpecName: "kube-api-access-cbnzm") pod "2ad0d32e-2227-4d12-bf62-19eb24597391" (UID: "2ad0d32e-2227-4d12-bf62-19eb24597391"). InnerVolumeSpecName "kube-api-access-cbnzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:40:53 crc kubenswrapper[4632]: I0313 11:40:53.963017 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbnzm\" (UniqueName: \"kubernetes.io/projected/2ad0d32e-2227-4d12-bf62-19eb24597391-kube-api-access-cbnzm\") on node \"crc\" DevicePath \"\"" Mar 13 11:40:53 crc kubenswrapper[4632]: I0313 11:40:53.963482 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2ad0d32e-2227-4d12-bf62-19eb24597391-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:40:53 crc kubenswrapper[4632]: I0313 11:40:53.998619 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2ad0d32e-2227-4d12-bf62-19eb24597391-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2ad0d32e-2227-4d12-bf62-19eb24597391" (UID: "2ad0d32e-2227-4d12-bf62-19eb24597391"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:40:54 crc kubenswrapper[4632]: I0313 11:40:54.065626 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2ad0d32e-2227-4d12-bf62-19eb24597391-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:40:54 crc kubenswrapper[4632]: I0313 11:40:54.786851 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kc89d" Mar 13 11:40:54 crc kubenswrapper[4632]: I0313 11:40:54.818637 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-kc89d"] Mar 13 11:40:54 crc kubenswrapper[4632]: I0313 11:40:54.827375 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-kc89d"] Mar 13 11:40:56 crc kubenswrapper[4632]: I0313 11:40:56.069148 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ad0d32e-2227-4d12-bf62-19eb24597391" path="/var/lib/kubelet/pods/2ad0d32e-2227-4d12-bf62-19eb24597391/volumes" Mar 13 11:40:57 crc kubenswrapper[4632]: I0313 11:40:57.667211 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-flb98" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerName="registry-server" probeResult="failure" output=< Mar 13 11:40:57 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:40:57 crc kubenswrapper[4632]: > Mar 13 11:41:00 crc kubenswrapper[4632]: I0313 11:41:00.045040 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:41:00 crc kubenswrapper[4632]: E0313 11:41:00.045606 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:41:07 crc kubenswrapper[4632]: I0313 11:41:07.665762 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-flb98" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerName="registry-server" probeResult="failure" output=< Mar 13 11:41:07 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:41:07 crc kubenswrapper[4632]: > Mar 13 11:41:11 crc kubenswrapper[4632]: I0313 11:41:11.044148 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:41:11 crc kubenswrapper[4632]: E0313 11:41:11.044992 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:41:17 crc kubenswrapper[4632]: I0313 11:41:17.674552 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-flb98" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerName="registry-server" probeResult="failure" output=< Mar 13 11:41:17 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:41:17 crc kubenswrapper[4632]: > Mar 13 11:41:22 crc kubenswrapper[4632]: I0313 11:41:22.044232 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:41:22 crc kubenswrapper[4632]: E0313 11:41:22.044866 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:41:26 crc kubenswrapper[4632]: I0313 11:41:26.673343 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:41:26 crc kubenswrapper[4632]: I0313 11:41:26.725998 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:41:26 crc kubenswrapper[4632]: I0313 11:41:26.925353 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-flb98"] Mar 13 11:41:28 crc kubenswrapper[4632]: I0313 11:41:28.147237 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-flb98" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerName="registry-server" containerID="cri-o://902661c18f068b146e4aecda28b4265925f2ef61195dd89d1bcb5ed19cf93dd6" gracePeriod=2 Mar 13 11:41:28 crc kubenswrapper[4632]: I0313 11:41:28.934566 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.043796 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dda62ea0-a8e2-46bb-a080-1b771e45feec-catalog-content\") pod \"dda62ea0-a8e2-46bb-a080-1b771e45feec\" (UID: \"dda62ea0-a8e2-46bb-a080-1b771e45feec\") " Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.044154 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dda62ea0-a8e2-46bb-a080-1b771e45feec-utilities\") pod \"dda62ea0-a8e2-46bb-a080-1b771e45feec\" (UID: \"dda62ea0-a8e2-46bb-a080-1b771e45feec\") " Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.044372 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wz4bb\" (UniqueName: \"kubernetes.io/projected/dda62ea0-a8e2-46bb-a080-1b771e45feec-kube-api-access-wz4bb\") pod \"dda62ea0-a8e2-46bb-a080-1b771e45feec\" (UID: \"dda62ea0-a8e2-46bb-a080-1b771e45feec\") " Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.047087 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dda62ea0-a8e2-46bb-a080-1b771e45feec-utilities" (OuterVolumeSpecName: "utilities") pod "dda62ea0-a8e2-46bb-a080-1b771e45feec" (UID: "dda62ea0-a8e2-46bb-a080-1b771e45feec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.068049 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dda62ea0-a8e2-46bb-a080-1b771e45feec-kube-api-access-wz4bb" (OuterVolumeSpecName: "kube-api-access-wz4bb") pod "dda62ea0-a8e2-46bb-a080-1b771e45feec" (UID: "dda62ea0-a8e2-46bb-a080-1b771e45feec"). InnerVolumeSpecName "kube-api-access-wz4bb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.147523 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wz4bb\" (UniqueName: \"kubernetes.io/projected/dda62ea0-a8e2-46bb-a080-1b771e45feec-kube-api-access-wz4bb\") on node \"crc\" DevicePath \"\"" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.147732 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dda62ea0-a8e2-46bb-a080-1b771e45feec-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.161924 4632 generic.go:334] "Generic (PLEG): container finished" podID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerID="902661c18f068b146e4aecda28b4265925f2ef61195dd89d1bcb5ed19cf93dd6" exitCode=0 Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.162008 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-flb98" event={"ID":"dda62ea0-a8e2-46bb-a080-1b771e45feec","Type":"ContainerDied","Data":"902661c18f068b146e4aecda28b4265925f2ef61195dd89d1bcb5ed19cf93dd6"} Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.162048 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-flb98" event={"ID":"dda62ea0-a8e2-46bb-a080-1b771e45feec","Type":"ContainerDied","Data":"bc8f50d20d6be8ef47a9be50fca7331e6e4176f53dc58080b0e546eb62f4a959"} Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.162073 4632 scope.go:117] "RemoveContainer" containerID="902661c18f068b146e4aecda28b4265925f2ef61195dd89d1bcb5ed19cf93dd6" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.162292 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-flb98" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.205740 4632 scope.go:117] "RemoveContainer" containerID="879dd7d998eed4fd26b7e2d5e411da489d9eceae2f429769ea2bcc4ae3aaf24d" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.242362 4632 scope.go:117] "RemoveContainer" containerID="f8176fa837f25c1697cac7d62b754ee2cfecbadb578ca4e441339c0f0c3b11c5" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.252054 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dda62ea0-a8e2-46bb-a080-1b771e45feec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dda62ea0-a8e2-46bb-a080-1b771e45feec" (UID: "dda62ea0-a8e2-46bb-a080-1b771e45feec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.279292 4632 scope.go:117] "RemoveContainer" containerID="902661c18f068b146e4aecda28b4265925f2ef61195dd89d1bcb5ed19cf93dd6" Mar 13 11:41:29 crc kubenswrapper[4632]: E0313 11:41:29.286377 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"902661c18f068b146e4aecda28b4265925f2ef61195dd89d1bcb5ed19cf93dd6\": container with ID starting with 902661c18f068b146e4aecda28b4265925f2ef61195dd89d1bcb5ed19cf93dd6 not found: ID does not exist" containerID="902661c18f068b146e4aecda28b4265925f2ef61195dd89d1bcb5ed19cf93dd6" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.286458 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"902661c18f068b146e4aecda28b4265925f2ef61195dd89d1bcb5ed19cf93dd6"} err="failed to get container status \"902661c18f068b146e4aecda28b4265925f2ef61195dd89d1bcb5ed19cf93dd6\": rpc error: code = NotFound desc = could not find container \"902661c18f068b146e4aecda28b4265925f2ef61195dd89d1bcb5ed19cf93dd6\": container with ID starting with 902661c18f068b146e4aecda28b4265925f2ef61195dd89d1bcb5ed19cf93dd6 not found: ID does not exist" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.286493 4632 scope.go:117] "RemoveContainer" containerID="879dd7d998eed4fd26b7e2d5e411da489d9eceae2f429769ea2bcc4ae3aaf24d" Mar 13 11:41:29 crc kubenswrapper[4632]: E0313 11:41:29.286991 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"879dd7d998eed4fd26b7e2d5e411da489d9eceae2f429769ea2bcc4ae3aaf24d\": container with ID starting with 879dd7d998eed4fd26b7e2d5e411da489d9eceae2f429769ea2bcc4ae3aaf24d not found: ID does not exist" containerID="879dd7d998eed4fd26b7e2d5e411da489d9eceae2f429769ea2bcc4ae3aaf24d" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.287065 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"879dd7d998eed4fd26b7e2d5e411da489d9eceae2f429769ea2bcc4ae3aaf24d"} err="failed to get container status \"879dd7d998eed4fd26b7e2d5e411da489d9eceae2f429769ea2bcc4ae3aaf24d\": rpc error: code = NotFound desc = could not find container \"879dd7d998eed4fd26b7e2d5e411da489d9eceae2f429769ea2bcc4ae3aaf24d\": container with ID starting with 879dd7d998eed4fd26b7e2d5e411da489d9eceae2f429769ea2bcc4ae3aaf24d not found: ID does not exist" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.287118 4632 scope.go:117] "RemoveContainer" containerID="f8176fa837f25c1697cac7d62b754ee2cfecbadb578ca4e441339c0f0c3b11c5" Mar 13 11:41:29 crc kubenswrapper[4632]: E0313 11:41:29.287821 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8176fa837f25c1697cac7d62b754ee2cfecbadb578ca4e441339c0f0c3b11c5\": container with ID starting with f8176fa837f25c1697cac7d62b754ee2cfecbadb578ca4e441339c0f0c3b11c5 not found: ID does not exist" containerID="f8176fa837f25c1697cac7d62b754ee2cfecbadb578ca4e441339c0f0c3b11c5" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.287856 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8176fa837f25c1697cac7d62b754ee2cfecbadb578ca4e441339c0f0c3b11c5"} err="failed to get container status \"f8176fa837f25c1697cac7d62b754ee2cfecbadb578ca4e441339c0f0c3b11c5\": rpc error: code = NotFound desc = could not find container \"f8176fa837f25c1697cac7d62b754ee2cfecbadb578ca4e441339c0f0c3b11c5\": container with ID starting with f8176fa837f25c1697cac7d62b754ee2cfecbadb578ca4e441339c0f0c3b11c5 not found: ID does not exist" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.352388 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dda62ea0-a8e2-46bb-a080-1b771e45feec-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.516449 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-flb98"] Mar 13 11:41:29 crc kubenswrapper[4632]: I0313 11:41:29.524547 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-flb98"] Mar 13 11:41:30 crc kubenswrapper[4632]: I0313 11:41:30.056539 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" path="/var/lib/kubelet/pods/dda62ea0-a8e2-46bb-a080-1b771e45feec/volumes" Mar 13 11:41:33 crc kubenswrapper[4632]: I0313 11:41:33.044953 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:41:33 crc kubenswrapper[4632]: E0313 11:41:33.045898 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:41:45 crc kubenswrapper[4632]: I0313 11:41:45.045368 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:41:45 crc kubenswrapper[4632]: E0313 11:41:45.046262 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:41:56 crc kubenswrapper[4632]: I0313 11:41:56.049232 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:41:56 crc kubenswrapper[4632]: E0313 11:41:56.050922 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.159401 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556702-prtsf"] Mar 13 11:42:00 crc kubenswrapper[4632]: E0313 11:42:00.160555 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad0d32e-2227-4d12-bf62-19eb24597391" containerName="extract-utilities" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.160574 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad0d32e-2227-4d12-bf62-19eb24597391" containerName="extract-utilities" Mar 13 11:42:00 crc kubenswrapper[4632]: E0313 11:42:00.160595 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerName="extract-content" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.160604 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerName="extract-content" Mar 13 11:42:00 crc kubenswrapper[4632]: E0313 11:42:00.160628 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerName="registry-server" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.160636 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerName="registry-server" Mar 13 11:42:00 crc kubenswrapper[4632]: E0313 11:42:00.160659 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad0d32e-2227-4d12-bf62-19eb24597391" containerName="registry-server" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.160668 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad0d32e-2227-4d12-bf62-19eb24597391" containerName="registry-server" Mar 13 11:42:00 crc kubenswrapper[4632]: E0313 11:42:00.160690 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ad0d32e-2227-4d12-bf62-19eb24597391" containerName="extract-content" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.160698 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ad0d32e-2227-4d12-bf62-19eb24597391" containerName="extract-content" Mar 13 11:42:00 crc kubenswrapper[4632]: E0313 11:42:00.160729 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerName="extract-utilities" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.160737 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerName="extract-utilities" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.161030 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="dda62ea0-a8e2-46bb-a080-1b771e45feec" containerName="registry-server" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.161085 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ad0d32e-2227-4d12-bf62-19eb24597391" containerName="registry-server" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.162515 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556702-prtsf" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.165420 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.165684 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.169050 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.169548 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556702-prtsf"] Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.318004 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb6f7\" (UniqueName: \"kubernetes.io/projected/ec8968a0-0c4c-47e1-87d8-3703bea87a89-kube-api-access-mb6f7\") pod \"auto-csr-approver-29556702-prtsf\" (UID: \"ec8968a0-0c4c-47e1-87d8-3703bea87a89\") " pod="openshift-infra/auto-csr-approver-29556702-prtsf" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.420178 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb6f7\" (UniqueName: \"kubernetes.io/projected/ec8968a0-0c4c-47e1-87d8-3703bea87a89-kube-api-access-mb6f7\") pod \"auto-csr-approver-29556702-prtsf\" (UID: \"ec8968a0-0c4c-47e1-87d8-3703bea87a89\") " pod="openshift-infra/auto-csr-approver-29556702-prtsf" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.445793 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb6f7\" (UniqueName: \"kubernetes.io/projected/ec8968a0-0c4c-47e1-87d8-3703bea87a89-kube-api-access-mb6f7\") pod \"auto-csr-approver-29556702-prtsf\" (UID: \"ec8968a0-0c4c-47e1-87d8-3703bea87a89\") " pod="openshift-infra/auto-csr-approver-29556702-prtsf" Mar 13 11:42:00 crc kubenswrapper[4632]: I0313 11:42:00.492681 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556702-prtsf" Mar 13 11:42:01 crc kubenswrapper[4632]: I0313 11:42:01.072767 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556702-prtsf"] Mar 13 11:42:01 crc kubenswrapper[4632]: I0313 11:42:01.463017 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556702-prtsf" event={"ID":"ec8968a0-0c4c-47e1-87d8-3703bea87a89","Type":"ContainerStarted","Data":"3d873398315a97348760f80182a041f20e51b27ec456a28c5d8999e39a7149d1"} Mar 13 11:42:03 crc kubenswrapper[4632]: I0313 11:42:03.481798 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556702-prtsf" event={"ID":"ec8968a0-0c4c-47e1-87d8-3703bea87a89","Type":"ContainerStarted","Data":"6ff82271933ceb662a6f6b867ecc2729be9d4acd3b4299ec77fdefa80de44bf3"} Mar 13 11:42:04 crc kubenswrapper[4632]: I0313 11:42:04.552202 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556702-prtsf" podStartSLOduration=3.662121479 podStartE2EDuration="4.552182376s" podCreationTimestamp="2026-03-13 11:42:00 +0000 UTC" firstStartedPulling="2026-03-13 11:42:01.079267308 +0000 UTC m=+5895.101797441" lastFinishedPulling="2026-03-13 11:42:01.969328205 +0000 UTC m=+5895.991858338" observedRunningTime="2026-03-13 11:42:04.549628843 +0000 UTC m=+5898.572158976" watchObservedRunningTime="2026-03-13 11:42:04.552182376 +0000 UTC m=+5898.574712509" Mar 13 11:42:06 crc kubenswrapper[4632]: I0313 11:42:06.547111 4632 generic.go:334] "Generic (PLEG): container finished" podID="ec8968a0-0c4c-47e1-87d8-3703bea87a89" containerID="6ff82271933ceb662a6f6b867ecc2729be9d4acd3b4299ec77fdefa80de44bf3" exitCode=0 Mar 13 11:42:06 crc kubenswrapper[4632]: I0313 11:42:06.547252 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556702-prtsf" event={"ID":"ec8968a0-0c4c-47e1-87d8-3703bea87a89","Type":"ContainerDied","Data":"6ff82271933ceb662a6f6b867ecc2729be9d4acd3b4299ec77fdefa80de44bf3"} Mar 13 11:42:07 crc kubenswrapper[4632]: I0313 11:42:07.952499 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556702-prtsf" Mar 13 11:42:08 crc kubenswrapper[4632]: I0313 11:42:08.090329 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb6f7\" (UniqueName: \"kubernetes.io/projected/ec8968a0-0c4c-47e1-87d8-3703bea87a89-kube-api-access-mb6f7\") pod \"ec8968a0-0c4c-47e1-87d8-3703bea87a89\" (UID: \"ec8968a0-0c4c-47e1-87d8-3703bea87a89\") " Mar 13 11:42:08 crc kubenswrapper[4632]: I0313 11:42:08.102158 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec8968a0-0c4c-47e1-87d8-3703bea87a89-kube-api-access-mb6f7" (OuterVolumeSpecName: "kube-api-access-mb6f7") pod "ec8968a0-0c4c-47e1-87d8-3703bea87a89" (UID: "ec8968a0-0c4c-47e1-87d8-3703bea87a89"). InnerVolumeSpecName "kube-api-access-mb6f7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:42:08 crc kubenswrapper[4632]: I0313 11:42:08.195057 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb6f7\" (UniqueName: \"kubernetes.io/projected/ec8968a0-0c4c-47e1-87d8-3703bea87a89-kube-api-access-mb6f7\") on node \"crc\" DevicePath \"\"" Mar 13 11:42:08 crc kubenswrapper[4632]: I0313 11:42:08.565027 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556702-prtsf" event={"ID":"ec8968a0-0c4c-47e1-87d8-3703bea87a89","Type":"ContainerDied","Data":"3d873398315a97348760f80182a041f20e51b27ec456a28c5d8999e39a7149d1"} Mar 13 11:42:08 crc kubenswrapper[4632]: I0313 11:42:08.565076 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d873398315a97348760f80182a041f20e51b27ec456a28c5d8999e39a7149d1" Mar 13 11:42:08 crc kubenswrapper[4632]: I0313 11:42:08.565119 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556702-prtsf" Mar 13 11:42:08 crc kubenswrapper[4632]: I0313 11:42:08.647553 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556696-6dgq6"] Mar 13 11:42:08 crc kubenswrapper[4632]: I0313 11:42:08.657610 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556696-6dgq6"] Mar 13 11:42:10 crc kubenswrapper[4632]: I0313 11:42:10.044214 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:42:10 crc kubenswrapper[4632]: E0313 11:42:10.044853 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:42:10 crc kubenswrapper[4632]: I0313 11:42:10.058312 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8600f7f-45fb-4aa6-b13b-9d6be5354009" path="/var/lib/kubelet/pods/f8600f7f-45fb-4aa6-b13b-9d6be5354009/volumes" Mar 13 11:42:21 crc kubenswrapper[4632]: I0313 11:42:21.954877 4632 scope.go:117] "RemoveContainer" containerID="ffbf598df91f4bb7277b432bb2bc1355e735cdb640ec4482a312abc6e198f0af" Mar 13 11:42:24 crc kubenswrapper[4632]: I0313 11:42:24.044217 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:42:24 crc kubenswrapper[4632]: E0313 11:42:24.045216 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:42:38 crc kubenswrapper[4632]: I0313 11:42:38.050571 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:42:38 crc kubenswrapper[4632]: E0313 11:42:38.052609 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:42:53 crc kubenswrapper[4632]: I0313 11:42:53.044926 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:42:53 crc kubenswrapper[4632]: E0313 11:42:53.046106 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:43:06 crc kubenswrapper[4632]: I0313 11:43:06.048660 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:43:06 crc kubenswrapper[4632]: E0313 11:43:06.053865 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:43:17 crc kubenswrapper[4632]: I0313 11:43:17.044221 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:43:17 crc kubenswrapper[4632]: E0313 11:43:17.045030 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:43:28 crc kubenswrapper[4632]: I0313 11:43:28.054324 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:43:28 crc kubenswrapper[4632]: E0313 11:43:28.055497 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:43:43 crc kubenswrapper[4632]: I0313 11:43:43.045022 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:43:43 crc kubenswrapper[4632]: E0313 11:43:43.045915 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:43:54 crc kubenswrapper[4632]: I0313 11:43:54.046655 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:43:54 crc kubenswrapper[4632]: E0313 11:43:54.047191 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:44:00 crc kubenswrapper[4632]: I0313 11:44:00.156473 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556704-r82jw"] Mar 13 11:44:00 crc kubenswrapper[4632]: E0313 11:44:00.158417 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec8968a0-0c4c-47e1-87d8-3703bea87a89" containerName="oc" Mar 13 11:44:00 crc kubenswrapper[4632]: I0313 11:44:00.158501 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec8968a0-0c4c-47e1-87d8-3703bea87a89" containerName="oc" Mar 13 11:44:00 crc kubenswrapper[4632]: I0313 11:44:00.158743 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec8968a0-0c4c-47e1-87d8-3703bea87a89" containerName="oc" Mar 13 11:44:00 crc kubenswrapper[4632]: I0313 11:44:00.159451 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556704-r82jw" Mar 13 11:44:00 crc kubenswrapper[4632]: I0313 11:44:00.167861 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556704-r82jw"] Mar 13 11:44:00 crc kubenswrapper[4632]: I0313 11:44:00.168603 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:44:00 crc kubenswrapper[4632]: I0313 11:44:00.168601 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:44:00 crc kubenswrapper[4632]: I0313 11:44:00.177669 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:44:00 crc kubenswrapper[4632]: I0313 11:44:00.337300 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcks9\" (UniqueName: \"kubernetes.io/projected/6300cb33-fba3-4d08-948b-0c6584d2ef26-kube-api-access-dcks9\") pod \"auto-csr-approver-29556704-r82jw\" (UID: \"6300cb33-fba3-4d08-948b-0c6584d2ef26\") " pod="openshift-infra/auto-csr-approver-29556704-r82jw" Mar 13 11:44:00 crc kubenswrapper[4632]: I0313 11:44:00.441863 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcks9\" (UniqueName: \"kubernetes.io/projected/6300cb33-fba3-4d08-948b-0c6584d2ef26-kube-api-access-dcks9\") pod \"auto-csr-approver-29556704-r82jw\" (UID: \"6300cb33-fba3-4d08-948b-0c6584d2ef26\") " pod="openshift-infra/auto-csr-approver-29556704-r82jw" Mar 13 11:44:00 crc kubenswrapper[4632]: I0313 11:44:00.475217 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcks9\" (UniqueName: \"kubernetes.io/projected/6300cb33-fba3-4d08-948b-0c6584d2ef26-kube-api-access-dcks9\") pod \"auto-csr-approver-29556704-r82jw\" (UID: \"6300cb33-fba3-4d08-948b-0c6584d2ef26\") " pod="openshift-infra/auto-csr-approver-29556704-r82jw" Mar 13 11:44:00 crc kubenswrapper[4632]: I0313 11:44:00.480065 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556704-r82jw" Mar 13 11:44:00 crc kubenswrapper[4632]: I0313 11:44:00.968527 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 11:44:00 crc kubenswrapper[4632]: I0313 11:44:00.970605 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556704-r82jw"] Mar 13 11:44:01 crc kubenswrapper[4632]: I0313 11:44:01.808487 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556704-r82jw" event={"ID":"6300cb33-fba3-4d08-948b-0c6584d2ef26","Type":"ContainerStarted","Data":"3ec1b3268e2b207a1a34f7039203387c5cea5c789a84c70b274d99fe9ffa654c"} Mar 13 11:44:02 crc kubenswrapper[4632]: E0313 11:44:02.497060 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6300cb33_fba3_4d08_948b_0c6584d2ef26.slice/crio-be61603d8895006f002bb067cf8ee34fe273ab45dfec9b8aa73261bbdd0ea048.scope\": RecentStats: unable to find data in memory cache]" Mar 13 11:44:02 crc kubenswrapper[4632]: I0313 11:44:02.822428 4632 generic.go:334] "Generic (PLEG): container finished" podID="6300cb33-fba3-4d08-948b-0c6584d2ef26" containerID="be61603d8895006f002bb067cf8ee34fe273ab45dfec9b8aa73261bbdd0ea048" exitCode=0 Mar 13 11:44:02 crc kubenswrapper[4632]: I0313 11:44:02.822486 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556704-r82jw" event={"ID":"6300cb33-fba3-4d08-948b-0c6584d2ef26","Type":"ContainerDied","Data":"be61603d8895006f002bb067cf8ee34fe273ab45dfec9b8aa73261bbdd0ea048"} Mar 13 11:44:04 crc kubenswrapper[4632]: I0313 11:44:04.201490 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556704-r82jw" Mar 13 11:44:04 crc kubenswrapper[4632]: I0313 11:44:04.318361 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcks9\" (UniqueName: \"kubernetes.io/projected/6300cb33-fba3-4d08-948b-0c6584d2ef26-kube-api-access-dcks9\") pod \"6300cb33-fba3-4d08-948b-0c6584d2ef26\" (UID: \"6300cb33-fba3-4d08-948b-0c6584d2ef26\") " Mar 13 11:44:04 crc kubenswrapper[4632]: I0313 11:44:04.324162 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6300cb33-fba3-4d08-948b-0c6584d2ef26-kube-api-access-dcks9" (OuterVolumeSpecName: "kube-api-access-dcks9") pod "6300cb33-fba3-4d08-948b-0c6584d2ef26" (UID: "6300cb33-fba3-4d08-948b-0c6584d2ef26"). InnerVolumeSpecName "kube-api-access-dcks9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:44:04 crc kubenswrapper[4632]: I0313 11:44:04.421411 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcks9\" (UniqueName: \"kubernetes.io/projected/6300cb33-fba3-4d08-948b-0c6584d2ef26-kube-api-access-dcks9\") on node \"crc\" DevicePath \"\"" Mar 13 11:44:04 crc kubenswrapper[4632]: I0313 11:44:04.845652 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556704-r82jw" event={"ID":"6300cb33-fba3-4d08-948b-0c6584d2ef26","Type":"ContainerDied","Data":"3ec1b3268e2b207a1a34f7039203387c5cea5c789a84c70b274d99fe9ffa654c"} Mar 13 11:44:04 crc kubenswrapper[4632]: I0313 11:44:04.845911 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3ec1b3268e2b207a1a34f7039203387c5cea5c789a84c70b274d99fe9ffa654c" Mar 13 11:44:04 crc kubenswrapper[4632]: I0313 11:44:04.845764 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556704-r82jw" Mar 13 11:44:05 crc kubenswrapper[4632]: I0313 11:44:05.314901 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556698-ncrvx"] Mar 13 11:44:05 crc kubenswrapper[4632]: I0313 11:44:05.330274 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556698-ncrvx"] Mar 13 11:44:06 crc kubenswrapper[4632]: I0313 11:44:06.056705 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08ee143e-f1cf-4c38-a811-d31496082a75" path="/var/lib/kubelet/pods/08ee143e-f1cf-4c38-a811-d31496082a75/volumes" Mar 13 11:44:09 crc kubenswrapper[4632]: I0313 11:44:09.043512 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:44:09 crc kubenswrapper[4632]: E0313 11:44:09.043815 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:44:22 crc kubenswrapper[4632]: I0313 11:44:22.199556 4632 scope.go:117] "RemoveContainer" containerID="19b28d2a56d1971c59c024b2b42655c24314722844900a0860bc74bbd0e6dfd4" Mar 13 11:44:23 crc kubenswrapper[4632]: I0313 11:44:23.045359 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:44:23 crc kubenswrapper[4632]: E0313 11:44:23.046027 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:44:34 crc kubenswrapper[4632]: I0313 11:44:34.044048 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:44:34 crc kubenswrapper[4632]: E0313 11:44:34.045001 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:44:49 crc kubenswrapper[4632]: I0313 11:44:49.045351 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:44:49 crc kubenswrapper[4632]: E0313 11:44:49.046440 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.044884 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:45:00 crc kubenswrapper[4632]: E0313 11:45:00.045686 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.201268 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv"] Mar 13 11:45:00 crc kubenswrapper[4632]: E0313 11:45:00.201965 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6300cb33-fba3-4d08-948b-0c6584d2ef26" containerName="oc" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.201979 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6300cb33-fba3-4d08-948b-0c6584d2ef26" containerName="oc" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.202361 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6300cb33-fba3-4d08-948b-0c6584d2ef26" containerName="oc" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.204601 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.207953 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.208136 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.208701 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv"] Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.364023 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-config-volume\") pod \"collect-profiles-29556705-cdjnv\" (UID: \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.364063 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-secret-volume\") pod \"collect-profiles-29556705-cdjnv\" (UID: \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.364118 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fbkn7\" (UniqueName: \"kubernetes.io/projected/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-kube-api-access-fbkn7\") pod \"collect-profiles-29556705-cdjnv\" (UID: \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.465989 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-config-volume\") pod \"collect-profiles-29556705-cdjnv\" (UID: \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.466031 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-secret-volume\") pod \"collect-profiles-29556705-cdjnv\" (UID: \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.466094 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fbkn7\" (UniqueName: \"kubernetes.io/projected/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-kube-api-access-fbkn7\") pod \"collect-profiles-29556705-cdjnv\" (UID: \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.466810 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-config-volume\") pod \"collect-profiles-29556705-cdjnv\" (UID: \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.479091 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-secret-volume\") pod \"collect-profiles-29556705-cdjnv\" (UID: \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.488833 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fbkn7\" (UniqueName: \"kubernetes.io/projected/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-kube-api-access-fbkn7\") pod \"collect-profiles-29556705-cdjnv\" (UID: \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" Mar 13 11:45:00 crc kubenswrapper[4632]: I0313 11:45:00.538611 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" Mar 13 11:45:01 crc kubenswrapper[4632]: I0313 11:45:01.138851 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv"] Mar 13 11:45:01 crc kubenswrapper[4632]: I0313 11:45:01.408069 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" event={"ID":"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7","Type":"ContainerStarted","Data":"a50d24de30277dacbb16bc71e07335e3c84d2cedb12dfb6c3d660775ff2f0c54"} Mar 13 11:45:01 crc kubenswrapper[4632]: I0313 11:45:01.408116 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" event={"ID":"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7","Type":"ContainerStarted","Data":"21cccd049d8f0c721f003daa0c370957ce291331c311203a423c768e560fb2f2"} Mar 13 11:45:02 crc kubenswrapper[4632]: I0313 11:45:02.421797 4632 generic.go:334] "Generic (PLEG): container finished" podID="964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7" containerID="a50d24de30277dacbb16bc71e07335e3c84d2cedb12dfb6c3d660775ff2f0c54" exitCode=0 Mar 13 11:45:02 crc kubenswrapper[4632]: I0313 11:45:02.421894 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" event={"ID":"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7","Type":"ContainerDied","Data":"a50d24de30277dacbb16bc71e07335e3c84d2cedb12dfb6c3d660775ff2f0c54"} Mar 13 11:45:03 crc kubenswrapper[4632]: I0313 11:45:03.806296 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" Mar 13 11:45:03 crc kubenswrapper[4632]: I0313 11:45:03.948587 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fbkn7\" (UniqueName: \"kubernetes.io/projected/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-kube-api-access-fbkn7\") pod \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\" (UID: \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\") " Mar 13 11:45:03 crc kubenswrapper[4632]: I0313 11:45:03.948641 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-config-volume\") pod \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\" (UID: \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\") " Mar 13 11:45:03 crc kubenswrapper[4632]: I0313 11:45:03.948759 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-secret-volume\") pod \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\" (UID: \"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7\") " Mar 13 11:45:03 crc kubenswrapper[4632]: I0313 11:45:03.949674 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-config-volume" (OuterVolumeSpecName: "config-volume") pod "964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7" (UID: "964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 11:45:03 crc kubenswrapper[4632]: I0313 11:45:03.950863 4632 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 11:45:03 crc kubenswrapper[4632]: I0313 11:45:03.969588 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7" (UID: "964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 11:45:03 crc kubenswrapper[4632]: I0313 11:45:03.969621 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-kube-api-access-fbkn7" (OuterVolumeSpecName: "kube-api-access-fbkn7") pod "964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7" (UID: "964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7"). InnerVolumeSpecName "kube-api-access-fbkn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:45:04 crc kubenswrapper[4632]: I0313 11:45:04.052435 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fbkn7\" (UniqueName: \"kubernetes.io/projected/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-kube-api-access-fbkn7\") on node \"crc\" DevicePath \"\"" Mar 13 11:45:04 crc kubenswrapper[4632]: I0313 11:45:04.052472 4632 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 11:45:04 crc kubenswrapper[4632]: I0313 11:45:04.449978 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" event={"ID":"964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7","Type":"ContainerDied","Data":"21cccd049d8f0c721f003daa0c370957ce291331c311203a423c768e560fb2f2"} Mar 13 11:45:04 crc kubenswrapper[4632]: I0313 11:45:04.450022 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21cccd049d8f0c721f003daa0c370957ce291331c311203a423c768e560fb2f2" Mar 13 11:45:04 crc kubenswrapper[4632]: I0313 11:45:04.450092 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv" Mar 13 11:45:04 crc kubenswrapper[4632]: I0313 11:45:04.521205 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8"] Mar 13 11:45:04 crc kubenswrapper[4632]: I0313 11:45:04.530092 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556660-7vph8"] Mar 13 11:45:06 crc kubenswrapper[4632]: I0313 11:45:06.057959 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f506e288-f3da-4d62-a6a2-bb598a62ed13" path="/var/lib/kubelet/pods/f506e288-f3da-4d62-a6a2-bb598a62ed13/volumes" Mar 13 11:45:13 crc kubenswrapper[4632]: I0313 11:45:13.044542 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:45:13 crc kubenswrapper[4632]: I0313 11:45:13.546835 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"a91f451a2842f8b1b73b10a99ff94ea342a17276601161b96bf6802b9f5327a9"} Mar 13 11:45:22 crc kubenswrapper[4632]: I0313 11:45:22.316873 4632 scope.go:117] "RemoveContainer" containerID="df99b126bcdc13810e89ae823dc76bf43eab9d932c52b6dd430fa449a698c642" Mar 13 11:46:00 crc kubenswrapper[4632]: I0313 11:46:00.154455 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556706-psc7t"] Mar 13 11:46:00 crc kubenswrapper[4632]: E0313 11:46:00.155431 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7" containerName="collect-profiles" Mar 13 11:46:00 crc kubenswrapper[4632]: I0313 11:46:00.155445 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7" containerName="collect-profiles" Mar 13 11:46:00 crc kubenswrapper[4632]: I0313 11:46:00.155656 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7" containerName="collect-profiles" Mar 13 11:46:00 crc kubenswrapper[4632]: I0313 11:46:00.156415 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556706-psc7t" Mar 13 11:46:00 crc kubenswrapper[4632]: I0313 11:46:00.159106 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:46:00 crc kubenswrapper[4632]: I0313 11:46:00.159616 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:46:00 crc kubenswrapper[4632]: I0313 11:46:00.161775 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:46:00 crc kubenswrapper[4632]: I0313 11:46:00.169414 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556706-psc7t"] Mar 13 11:46:00 crc kubenswrapper[4632]: I0313 11:46:00.279753 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wcp2\" (UniqueName: \"kubernetes.io/projected/8980f067-488f-497f-8ba7-5ee2d3069d62-kube-api-access-2wcp2\") pod \"auto-csr-approver-29556706-psc7t\" (UID: \"8980f067-488f-497f-8ba7-5ee2d3069d62\") " pod="openshift-infra/auto-csr-approver-29556706-psc7t" Mar 13 11:46:00 crc kubenswrapper[4632]: I0313 11:46:00.381509 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wcp2\" (UniqueName: \"kubernetes.io/projected/8980f067-488f-497f-8ba7-5ee2d3069d62-kube-api-access-2wcp2\") pod \"auto-csr-approver-29556706-psc7t\" (UID: \"8980f067-488f-497f-8ba7-5ee2d3069d62\") " pod="openshift-infra/auto-csr-approver-29556706-psc7t" Mar 13 11:46:00 crc kubenswrapper[4632]: I0313 11:46:00.412187 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wcp2\" (UniqueName: \"kubernetes.io/projected/8980f067-488f-497f-8ba7-5ee2d3069d62-kube-api-access-2wcp2\") pod \"auto-csr-approver-29556706-psc7t\" (UID: \"8980f067-488f-497f-8ba7-5ee2d3069d62\") " pod="openshift-infra/auto-csr-approver-29556706-psc7t" Mar 13 11:46:00 crc kubenswrapper[4632]: I0313 11:46:00.479214 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556706-psc7t" Mar 13 11:46:01 crc kubenswrapper[4632]: I0313 11:46:01.032056 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556706-psc7t"] Mar 13 11:46:01 crc kubenswrapper[4632]: I0313 11:46:01.235244 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556706-psc7t" event={"ID":"8980f067-488f-497f-8ba7-5ee2d3069d62","Type":"ContainerStarted","Data":"5c5f685d468e35ff54098d18f75b65f3dcbdea73c7e4e084c9f376a192c7c67f"} Mar 13 11:46:03 crc kubenswrapper[4632]: I0313 11:46:03.259205 4632 generic.go:334] "Generic (PLEG): container finished" podID="8980f067-488f-497f-8ba7-5ee2d3069d62" containerID="49f7bf435fba27e68a413e86a923b4ddacb7432c6b3ec46cefd0935c8e2aecc2" exitCode=0 Mar 13 11:46:03 crc kubenswrapper[4632]: I0313 11:46:03.259313 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556706-psc7t" event={"ID":"8980f067-488f-497f-8ba7-5ee2d3069d62","Type":"ContainerDied","Data":"49f7bf435fba27e68a413e86a923b4ddacb7432c6b3ec46cefd0935c8e2aecc2"} Mar 13 11:46:04 crc kubenswrapper[4632]: I0313 11:46:04.679919 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556706-psc7t" Mar 13 11:46:04 crc kubenswrapper[4632]: I0313 11:46:04.802772 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wcp2\" (UniqueName: \"kubernetes.io/projected/8980f067-488f-497f-8ba7-5ee2d3069d62-kube-api-access-2wcp2\") pod \"8980f067-488f-497f-8ba7-5ee2d3069d62\" (UID: \"8980f067-488f-497f-8ba7-5ee2d3069d62\") " Mar 13 11:46:04 crc kubenswrapper[4632]: I0313 11:46:04.809422 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8980f067-488f-497f-8ba7-5ee2d3069d62-kube-api-access-2wcp2" (OuterVolumeSpecName: "kube-api-access-2wcp2") pod "8980f067-488f-497f-8ba7-5ee2d3069d62" (UID: "8980f067-488f-497f-8ba7-5ee2d3069d62"). InnerVolumeSpecName "kube-api-access-2wcp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:46:04 crc kubenswrapper[4632]: I0313 11:46:04.905015 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wcp2\" (UniqueName: \"kubernetes.io/projected/8980f067-488f-497f-8ba7-5ee2d3069d62-kube-api-access-2wcp2\") on node \"crc\" DevicePath \"\"" Mar 13 11:46:05 crc kubenswrapper[4632]: I0313 11:46:05.280138 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556706-psc7t" event={"ID":"8980f067-488f-497f-8ba7-5ee2d3069d62","Type":"ContainerDied","Data":"5c5f685d468e35ff54098d18f75b65f3dcbdea73c7e4e084c9f376a192c7c67f"} Mar 13 11:46:05 crc kubenswrapper[4632]: I0313 11:46:05.280179 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c5f685d468e35ff54098d18f75b65f3dcbdea73c7e4e084c9f376a192c7c67f" Mar 13 11:46:05 crc kubenswrapper[4632]: I0313 11:46:05.280232 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556706-psc7t" Mar 13 11:46:05 crc kubenswrapper[4632]: I0313 11:46:05.764502 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556700-rjhgg"] Mar 13 11:46:05 crc kubenswrapper[4632]: I0313 11:46:05.776926 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556700-rjhgg"] Mar 13 11:46:06 crc kubenswrapper[4632]: I0313 11:46:06.058826 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5" path="/var/lib/kubelet/pods/d39670e0-5c6f-40f3-b7b1-46dc6fdc5fc5/volumes" Mar 13 11:46:22 crc kubenswrapper[4632]: I0313 11:46:22.380329 4632 scope.go:117] "RemoveContainer" containerID="9d9bde8c088f7036968dfb6799f01abee05cb617fe7dd0b877a31e0a9fec56bb" Mar 13 11:46:22 crc kubenswrapper[4632]: I0313 11:46:22.403396 4632 scope.go:117] "RemoveContainer" containerID="38648afa2577e5a9fb5d9d7e00b5f9e414dd8af022764679faa186da16c72b26" Mar 13 11:46:22 crc kubenswrapper[4632]: I0313 11:46:22.429544 4632 scope.go:117] "RemoveContainer" containerID="e2dc32a91f84dbc41f05188f09b5ec2790c5d429e256411e1defae69e3e43deb" Mar 13 11:46:22 crc kubenswrapper[4632]: I0313 11:46:22.504298 4632 scope.go:117] "RemoveContainer" containerID="4f578f7aba25fc35c2994f8884e7599a303084b2db092f8a745f47b140231631" Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.485381 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hrpzk"] Mar 13 11:47:13 crc kubenswrapper[4632]: E0313 11:47:13.486343 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8980f067-488f-497f-8ba7-5ee2d3069d62" containerName="oc" Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.486360 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8980f067-488f-497f-8ba7-5ee2d3069d62" containerName="oc" Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.487110 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8980f067-488f-497f-8ba7-5ee2d3069d62" containerName="oc" Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.488684 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.510479 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrpzk"] Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.616316 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqpsp\" (UniqueName: \"kubernetes.io/projected/b73e9e28-59a7-4e69-818c-03972ee9f6db-kube-api-access-bqpsp\") pod \"redhat-marketplace-hrpzk\" (UID: \"b73e9e28-59a7-4e69-818c-03972ee9f6db\") " pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.616386 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b73e9e28-59a7-4e69-818c-03972ee9f6db-catalog-content\") pod \"redhat-marketplace-hrpzk\" (UID: \"b73e9e28-59a7-4e69-818c-03972ee9f6db\") " pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.616464 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b73e9e28-59a7-4e69-818c-03972ee9f6db-utilities\") pod \"redhat-marketplace-hrpzk\" (UID: \"b73e9e28-59a7-4e69-818c-03972ee9f6db\") " pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.718109 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b73e9e28-59a7-4e69-818c-03972ee9f6db-utilities\") pod \"redhat-marketplace-hrpzk\" (UID: \"b73e9e28-59a7-4e69-818c-03972ee9f6db\") " pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.718736 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b73e9e28-59a7-4e69-818c-03972ee9f6db-utilities\") pod \"redhat-marketplace-hrpzk\" (UID: \"b73e9e28-59a7-4e69-818c-03972ee9f6db\") " pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.719087 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqpsp\" (UniqueName: \"kubernetes.io/projected/b73e9e28-59a7-4e69-818c-03972ee9f6db-kube-api-access-bqpsp\") pod \"redhat-marketplace-hrpzk\" (UID: \"b73e9e28-59a7-4e69-818c-03972ee9f6db\") " pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.719534 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b73e9e28-59a7-4e69-818c-03972ee9f6db-catalog-content\") pod \"redhat-marketplace-hrpzk\" (UID: \"b73e9e28-59a7-4e69-818c-03972ee9f6db\") " pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.719856 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b73e9e28-59a7-4e69-818c-03972ee9f6db-catalog-content\") pod \"redhat-marketplace-hrpzk\" (UID: \"b73e9e28-59a7-4e69-818c-03972ee9f6db\") " pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.737920 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqpsp\" (UniqueName: \"kubernetes.io/projected/b73e9e28-59a7-4e69-818c-03972ee9f6db-kube-api-access-bqpsp\") pod \"redhat-marketplace-hrpzk\" (UID: \"b73e9e28-59a7-4e69-818c-03972ee9f6db\") " pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:13 crc kubenswrapper[4632]: I0313 11:47:13.852748 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:14 crc kubenswrapper[4632]: I0313 11:47:14.349262 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrpzk"] Mar 13 11:47:15 crc kubenswrapper[4632]: I0313 11:47:15.000731 4632 generic.go:334] "Generic (PLEG): container finished" podID="b73e9e28-59a7-4e69-818c-03972ee9f6db" containerID="daed49207d919944fcd55fa320fc1c031d61060fce906a2b0f553c91d9ebbb5b" exitCode=0 Mar 13 11:47:15 crc kubenswrapper[4632]: I0313 11:47:15.001001 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrpzk" event={"ID":"b73e9e28-59a7-4e69-818c-03972ee9f6db","Type":"ContainerDied","Data":"daed49207d919944fcd55fa320fc1c031d61060fce906a2b0f553c91d9ebbb5b"} Mar 13 11:47:15 crc kubenswrapper[4632]: I0313 11:47:15.001061 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrpzk" event={"ID":"b73e9e28-59a7-4e69-818c-03972ee9f6db","Type":"ContainerStarted","Data":"68fa17ac498b5d3b5b99bb554433e7a897e9d668ce1084e5bebf2b51c96391f9"} Mar 13 11:47:17 crc kubenswrapper[4632]: I0313 11:47:17.032970 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrpzk" event={"ID":"b73e9e28-59a7-4e69-818c-03972ee9f6db","Type":"ContainerStarted","Data":"8f3b80c84e004d10de4ab88d81eca36dc2a9a31a2533f8d5d74d789e09b9f3c1"} Mar 13 11:47:18 crc kubenswrapper[4632]: I0313 11:47:18.045531 4632 generic.go:334] "Generic (PLEG): container finished" podID="b73e9e28-59a7-4e69-818c-03972ee9f6db" containerID="8f3b80c84e004d10de4ab88d81eca36dc2a9a31a2533f8d5d74d789e09b9f3c1" exitCode=0 Mar 13 11:47:18 crc kubenswrapper[4632]: I0313 11:47:18.057707 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrpzk" event={"ID":"b73e9e28-59a7-4e69-818c-03972ee9f6db","Type":"ContainerDied","Data":"8f3b80c84e004d10de4ab88d81eca36dc2a9a31a2533f8d5d74d789e09b9f3c1"} Mar 13 11:47:19 crc kubenswrapper[4632]: I0313 11:47:19.057401 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrpzk" event={"ID":"b73e9e28-59a7-4e69-818c-03972ee9f6db","Type":"ContainerStarted","Data":"aad74b1bbd85b819ef3895de0f161045236f8fae19bf8e68390588fd962447e5"} Mar 13 11:47:19 crc kubenswrapper[4632]: I0313 11:47:19.080964 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hrpzk" podStartSLOduration=2.362119064 podStartE2EDuration="6.080931074s" podCreationTimestamp="2026-03-13 11:47:13 +0000 UTC" firstStartedPulling="2026-03-13 11:47:15.007456103 +0000 UTC m=+6209.029986236" lastFinishedPulling="2026-03-13 11:47:18.726268113 +0000 UTC m=+6212.748798246" observedRunningTime="2026-03-13 11:47:19.074370742 +0000 UTC m=+6213.096900875" watchObservedRunningTime="2026-03-13 11:47:19.080931074 +0000 UTC m=+6213.103461207" Mar 13 11:47:23 crc kubenswrapper[4632]: I0313 11:47:23.853680 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:23 crc kubenswrapper[4632]: I0313 11:47:23.854263 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:24 crc kubenswrapper[4632]: I0313 11:47:24.896120 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-hrpzk" podUID="b73e9e28-59a7-4e69-818c-03972ee9f6db" containerName="registry-server" probeResult="failure" output=< Mar 13 11:47:24 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:47:24 crc kubenswrapper[4632]: > Mar 13 11:47:33 crc kubenswrapper[4632]: I0313 11:47:33.898543 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:33 crc kubenswrapper[4632]: I0313 11:47:33.954306 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:34 crc kubenswrapper[4632]: I0313 11:47:34.170143 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrpzk"] Mar 13 11:47:35 crc kubenswrapper[4632]: I0313 11:47:35.199407 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hrpzk" podUID="b73e9e28-59a7-4e69-818c-03972ee9f6db" containerName="registry-server" containerID="cri-o://aad74b1bbd85b819ef3895de0f161045236f8fae19bf8e68390588fd962447e5" gracePeriod=2 Mar 13 11:47:35 crc kubenswrapper[4632]: I0313 11:47:35.770095 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:35 crc kubenswrapper[4632]: I0313 11:47:35.897993 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqpsp\" (UniqueName: \"kubernetes.io/projected/b73e9e28-59a7-4e69-818c-03972ee9f6db-kube-api-access-bqpsp\") pod \"b73e9e28-59a7-4e69-818c-03972ee9f6db\" (UID: \"b73e9e28-59a7-4e69-818c-03972ee9f6db\") " Mar 13 11:47:35 crc kubenswrapper[4632]: I0313 11:47:35.898104 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b73e9e28-59a7-4e69-818c-03972ee9f6db-utilities\") pod \"b73e9e28-59a7-4e69-818c-03972ee9f6db\" (UID: \"b73e9e28-59a7-4e69-818c-03972ee9f6db\") " Mar 13 11:47:35 crc kubenswrapper[4632]: I0313 11:47:35.898349 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b73e9e28-59a7-4e69-818c-03972ee9f6db-catalog-content\") pod \"b73e9e28-59a7-4e69-818c-03972ee9f6db\" (UID: \"b73e9e28-59a7-4e69-818c-03972ee9f6db\") " Mar 13 11:47:35 crc kubenswrapper[4632]: I0313 11:47:35.899356 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b73e9e28-59a7-4e69-818c-03972ee9f6db-utilities" (OuterVolumeSpecName: "utilities") pod "b73e9e28-59a7-4e69-818c-03972ee9f6db" (UID: "b73e9e28-59a7-4e69-818c-03972ee9f6db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:47:35 crc kubenswrapper[4632]: I0313 11:47:35.911143 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b73e9e28-59a7-4e69-818c-03972ee9f6db-kube-api-access-bqpsp" (OuterVolumeSpecName: "kube-api-access-bqpsp") pod "b73e9e28-59a7-4e69-818c-03972ee9f6db" (UID: "b73e9e28-59a7-4e69-818c-03972ee9f6db"). InnerVolumeSpecName "kube-api-access-bqpsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:47:35 crc kubenswrapper[4632]: I0313 11:47:35.924959 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b73e9e28-59a7-4e69-818c-03972ee9f6db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b73e9e28-59a7-4e69-818c-03972ee9f6db" (UID: "b73e9e28-59a7-4e69-818c-03972ee9f6db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.000164 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b73e9e28-59a7-4e69-818c-03972ee9f6db-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.000217 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqpsp\" (UniqueName: \"kubernetes.io/projected/b73e9e28-59a7-4e69-818c-03972ee9f6db-kube-api-access-bqpsp\") on node \"crc\" DevicePath \"\"" Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.000231 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b73e9e28-59a7-4e69-818c-03972ee9f6db-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.211339 4632 generic.go:334] "Generic (PLEG): container finished" podID="b73e9e28-59a7-4e69-818c-03972ee9f6db" containerID="aad74b1bbd85b819ef3895de0f161045236f8fae19bf8e68390588fd962447e5" exitCode=0 Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.211381 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrpzk" event={"ID":"b73e9e28-59a7-4e69-818c-03972ee9f6db","Type":"ContainerDied","Data":"aad74b1bbd85b819ef3895de0f161045236f8fae19bf8e68390588fd962447e5"} Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.211648 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hrpzk" event={"ID":"b73e9e28-59a7-4e69-818c-03972ee9f6db","Type":"ContainerDied","Data":"68fa17ac498b5d3b5b99bb554433e7a897e9d668ce1084e5bebf2b51c96391f9"} Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.211668 4632 scope.go:117] "RemoveContainer" containerID="aad74b1bbd85b819ef3895de0f161045236f8fae19bf8e68390588fd962447e5" Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.211455 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hrpzk" Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.237708 4632 scope.go:117] "RemoveContainer" containerID="8f3b80c84e004d10de4ab88d81eca36dc2a9a31a2533f8d5d74d789e09b9f3c1" Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.261069 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrpzk"] Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.261615 4632 scope.go:117] "RemoveContainer" containerID="daed49207d919944fcd55fa320fc1c031d61060fce906a2b0f553c91d9ebbb5b" Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.268469 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hrpzk"] Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.338806 4632 scope.go:117] "RemoveContainer" containerID="aad74b1bbd85b819ef3895de0f161045236f8fae19bf8e68390588fd962447e5" Mar 13 11:47:36 crc kubenswrapper[4632]: E0313 11:47:36.341780 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aad74b1bbd85b819ef3895de0f161045236f8fae19bf8e68390588fd962447e5\": container with ID starting with aad74b1bbd85b819ef3895de0f161045236f8fae19bf8e68390588fd962447e5 not found: ID does not exist" containerID="aad74b1bbd85b819ef3895de0f161045236f8fae19bf8e68390588fd962447e5" Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.341832 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aad74b1bbd85b819ef3895de0f161045236f8fae19bf8e68390588fd962447e5"} err="failed to get container status \"aad74b1bbd85b819ef3895de0f161045236f8fae19bf8e68390588fd962447e5\": rpc error: code = NotFound desc = could not find container \"aad74b1bbd85b819ef3895de0f161045236f8fae19bf8e68390588fd962447e5\": container with ID starting with aad74b1bbd85b819ef3895de0f161045236f8fae19bf8e68390588fd962447e5 not found: ID does not exist" Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.341885 4632 scope.go:117] "RemoveContainer" containerID="8f3b80c84e004d10de4ab88d81eca36dc2a9a31a2533f8d5d74d789e09b9f3c1" Mar 13 11:47:36 crc kubenswrapper[4632]: E0313 11:47:36.342710 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f3b80c84e004d10de4ab88d81eca36dc2a9a31a2533f8d5d74d789e09b9f3c1\": container with ID starting with 8f3b80c84e004d10de4ab88d81eca36dc2a9a31a2533f8d5d74d789e09b9f3c1 not found: ID does not exist" containerID="8f3b80c84e004d10de4ab88d81eca36dc2a9a31a2533f8d5d74d789e09b9f3c1" Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.342904 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f3b80c84e004d10de4ab88d81eca36dc2a9a31a2533f8d5d74d789e09b9f3c1"} err="failed to get container status \"8f3b80c84e004d10de4ab88d81eca36dc2a9a31a2533f8d5d74d789e09b9f3c1\": rpc error: code = NotFound desc = could not find container \"8f3b80c84e004d10de4ab88d81eca36dc2a9a31a2533f8d5d74d789e09b9f3c1\": container with ID starting with 8f3b80c84e004d10de4ab88d81eca36dc2a9a31a2533f8d5d74d789e09b9f3c1 not found: ID does not exist" Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.342956 4632 scope.go:117] "RemoveContainer" containerID="daed49207d919944fcd55fa320fc1c031d61060fce906a2b0f553c91d9ebbb5b" Mar 13 11:47:36 crc kubenswrapper[4632]: E0313 11:47:36.343390 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daed49207d919944fcd55fa320fc1c031d61060fce906a2b0f553c91d9ebbb5b\": container with ID starting with daed49207d919944fcd55fa320fc1c031d61060fce906a2b0f553c91d9ebbb5b not found: ID does not exist" containerID="daed49207d919944fcd55fa320fc1c031d61060fce906a2b0f553c91d9ebbb5b" Mar 13 11:47:36 crc kubenswrapper[4632]: I0313 11:47:36.343408 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daed49207d919944fcd55fa320fc1c031d61060fce906a2b0f553c91d9ebbb5b"} err="failed to get container status \"daed49207d919944fcd55fa320fc1c031d61060fce906a2b0f553c91d9ebbb5b\": rpc error: code = NotFound desc = could not find container \"daed49207d919944fcd55fa320fc1c031d61060fce906a2b0f553c91d9ebbb5b\": container with ID starting with daed49207d919944fcd55fa320fc1c031d61060fce906a2b0f553c91d9ebbb5b not found: ID does not exist" Mar 13 11:47:38 crc kubenswrapper[4632]: I0313 11:47:38.060497 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b73e9e28-59a7-4e69-818c-03972ee9f6db" path="/var/lib/kubelet/pods/b73e9e28-59a7-4e69-818c-03972ee9f6db/volumes" Mar 13 11:47:40 crc kubenswrapper[4632]: I0313 11:47:40.461095 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:47:40 crc kubenswrapper[4632]: I0313 11:47:40.461553 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.191073 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556708-f6hsv"] Mar 13 11:48:00 crc kubenswrapper[4632]: E0313 11:48:00.192042 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b73e9e28-59a7-4e69-818c-03972ee9f6db" containerName="extract-content" Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.192057 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="b73e9e28-59a7-4e69-818c-03972ee9f6db" containerName="extract-content" Mar 13 11:48:00 crc kubenswrapper[4632]: E0313 11:48:00.192066 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b73e9e28-59a7-4e69-818c-03972ee9f6db" containerName="registry-server" Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.192072 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="b73e9e28-59a7-4e69-818c-03972ee9f6db" containerName="registry-server" Mar 13 11:48:00 crc kubenswrapper[4632]: E0313 11:48:00.192116 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b73e9e28-59a7-4e69-818c-03972ee9f6db" containerName="extract-utilities" Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.192125 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="b73e9e28-59a7-4e69-818c-03972ee9f6db" containerName="extract-utilities" Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.192301 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="b73e9e28-59a7-4e69-818c-03972ee9f6db" containerName="registry-server" Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.192926 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556708-f6hsv" Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.195872 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.196400 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.196679 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.210922 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556708-f6hsv"] Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.296205 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsjt5\" (UniqueName: \"kubernetes.io/projected/5464d278-31e9-45aa-9e87-78ef3e96115e-kube-api-access-gsjt5\") pod \"auto-csr-approver-29556708-f6hsv\" (UID: \"5464d278-31e9-45aa-9e87-78ef3e96115e\") " pod="openshift-infra/auto-csr-approver-29556708-f6hsv" Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.397894 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsjt5\" (UniqueName: \"kubernetes.io/projected/5464d278-31e9-45aa-9e87-78ef3e96115e-kube-api-access-gsjt5\") pod \"auto-csr-approver-29556708-f6hsv\" (UID: \"5464d278-31e9-45aa-9e87-78ef3e96115e\") " pod="openshift-infra/auto-csr-approver-29556708-f6hsv" Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.422077 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsjt5\" (UniqueName: \"kubernetes.io/projected/5464d278-31e9-45aa-9e87-78ef3e96115e-kube-api-access-gsjt5\") pod \"auto-csr-approver-29556708-f6hsv\" (UID: \"5464d278-31e9-45aa-9e87-78ef3e96115e\") " pod="openshift-infra/auto-csr-approver-29556708-f6hsv" Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.526530 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556708-f6hsv" Mar 13 11:48:00 crc kubenswrapper[4632]: I0313 11:48:00.869722 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556708-f6hsv"] Mar 13 11:48:01 crc kubenswrapper[4632]: I0313 11:48:01.462859 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556708-f6hsv" event={"ID":"5464d278-31e9-45aa-9e87-78ef3e96115e","Type":"ContainerStarted","Data":"3bbec0ef6d8358428ebfe87e34ae623dc5307c008bdbed4bc8220f8b08a51513"} Mar 13 11:48:03 crc kubenswrapper[4632]: I0313 11:48:03.489532 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556708-f6hsv" event={"ID":"5464d278-31e9-45aa-9e87-78ef3e96115e","Type":"ContainerStarted","Data":"6f4e30c3bf10310c255b10e3f6602511c866fd5af961f5f486fed69de586adb4"} Mar 13 11:48:03 crc kubenswrapper[4632]: I0313 11:48:03.515815 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556708-f6hsv" podStartSLOduration=2.099296206 podStartE2EDuration="3.515792034s" podCreationTimestamp="2026-03-13 11:48:00 +0000 UTC" firstStartedPulling="2026-03-13 11:48:00.878026856 +0000 UTC m=+6254.900556989" lastFinishedPulling="2026-03-13 11:48:02.294522684 +0000 UTC m=+6256.317052817" observedRunningTime="2026-03-13 11:48:03.506807703 +0000 UTC m=+6257.529337846" watchObservedRunningTime="2026-03-13 11:48:03.515792034 +0000 UTC m=+6257.538322167" Mar 13 11:48:04 crc kubenswrapper[4632]: I0313 11:48:04.503365 4632 generic.go:334] "Generic (PLEG): container finished" podID="5464d278-31e9-45aa-9e87-78ef3e96115e" containerID="6f4e30c3bf10310c255b10e3f6602511c866fd5af961f5f486fed69de586adb4" exitCode=0 Mar 13 11:48:04 crc kubenswrapper[4632]: I0313 11:48:04.503705 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556708-f6hsv" event={"ID":"5464d278-31e9-45aa-9e87-78ef3e96115e","Type":"ContainerDied","Data":"6f4e30c3bf10310c255b10e3f6602511c866fd5af961f5f486fed69de586adb4"} Mar 13 11:48:05 crc kubenswrapper[4632]: I0313 11:48:05.950993 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556708-f6hsv" Mar 13 11:48:06 crc kubenswrapper[4632]: I0313 11:48:06.104964 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsjt5\" (UniqueName: \"kubernetes.io/projected/5464d278-31e9-45aa-9e87-78ef3e96115e-kube-api-access-gsjt5\") pod \"5464d278-31e9-45aa-9e87-78ef3e96115e\" (UID: \"5464d278-31e9-45aa-9e87-78ef3e96115e\") " Mar 13 11:48:06 crc kubenswrapper[4632]: I0313 11:48:06.111203 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5464d278-31e9-45aa-9e87-78ef3e96115e-kube-api-access-gsjt5" (OuterVolumeSpecName: "kube-api-access-gsjt5") pod "5464d278-31e9-45aa-9e87-78ef3e96115e" (UID: "5464d278-31e9-45aa-9e87-78ef3e96115e"). InnerVolumeSpecName "kube-api-access-gsjt5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:48:06 crc kubenswrapper[4632]: I0313 11:48:06.208267 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsjt5\" (UniqueName: \"kubernetes.io/projected/5464d278-31e9-45aa-9e87-78ef3e96115e-kube-api-access-gsjt5\") on node \"crc\" DevicePath \"\"" Mar 13 11:48:06 crc kubenswrapper[4632]: I0313 11:48:06.521044 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556708-f6hsv" event={"ID":"5464d278-31e9-45aa-9e87-78ef3e96115e","Type":"ContainerDied","Data":"3bbec0ef6d8358428ebfe87e34ae623dc5307c008bdbed4bc8220f8b08a51513"} Mar 13 11:48:06 crc kubenswrapper[4632]: I0313 11:48:06.521108 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3bbec0ef6d8358428ebfe87e34ae623dc5307c008bdbed4bc8220f8b08a51513" Mar 13 11:48:06 crc kubenswrapper[4632]: I0313 11:48:06.521112 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556708-f6hsv" Mar 13 11:48:06 crc kubenswrapper[4632]: I0313 11:48:06.586298 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556702-prtsf"] Mar 13 11:48:06 crc kubenswrapper[4632]: I0313 11:48:06.593916 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556702-prtsf"] Mar 13 11:48:08 crc kubenswrapper[4632]: I0313 11:48:08.062322 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec8968a0-0c4c-47e1-87d8-3703bea87a89" path="/var/lib/kubelet/pods/ec8968a0-0c4c-47e1-87d8-3703bea87a89/volumes" Mar 13 11:48:10 crc kubenswrapper[4632]: I0313 11:48:10.461187 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:48:10 crc kubenswrapper[4632]: I0313 11:48:10.461554 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:48:19 crc kubenswrapper[4632]: I0313 11:48:19.771078 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-72gcl"] Mar 13 11:48:19 crc kubenswrapper[4632]: E0313 11:48:19.772325 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5464d278-31e9-45aa-9e87-78ef3e96115e" containerName="oc" Mar 13 11:48:19 crc kubenswrapper[4632]: I0313 11:48:19.772346 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="5464d278-31e9-45aa-9e87-78ef3e96115e" containerName="oc" Mar 13 11:48:19 crc kubenswrapper[4632]: I0313 11:48:19.772609 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="5464d278-31e9-45aa-9e87-78ef3e96115e" containerName="oc" Mar 13 11:48:19 crc kubenswrapper[4632]: I0313 11:48:19.778029 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:19 crc kubenswrapper[4632]: I0313 11:48:19.786115 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72gcl"] Mar 13 11:48:19 crc kubenswrapper[4632]: I0313 11:48:19.900397 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d87f4ca1-c949-445e-86d5-ec3f446e07d7-utilities\") pod \"community-operators-72gcl\" (UID: \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\") " pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:19 crc kubenswrapper[4632]: I0313 11:48:19.900488 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7sqz2\" (UniqueName: \"kubernetes.io/projected/d87f4ca1-c949-445e-86d5-ec3f446e07d7-kube-api-access-7sqz2\") pod \"community-operators-72gcl\" (UID: \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\") " pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:19 crc kubenswrapper[4632]: I0313 11:48:19.900571 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d87f4ca1-c949-445e-86d5-ec3f446e07d7-catalog-content\") pod \"community-operators-72gcl\" (UID: \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\") " pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:20 crc kubenswrapper[4632]: I0313 11:48:20.002698 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7sqz2\" (UniqueName: \"kubernetes.io/projected/d87f4ca1-c949-445e-86d5-ec3f446e07d7-kube-api-access-7sqz2\") pod \"community-operators-72gcl\" (UID: \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\") " pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:20 crc kubenswrapper[4632]: I0313 11:48:20.002752 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d87f4ca1-c949-445e-86d5-ec3f446e07d7-catalog-content\") pod \"community-operators-72gcl\" (UID: \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\") " pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:20 crc kubenswrapper[4632]: I0313 11:48:20.002930 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d87f4ca1-c949-445e-86d5-ec3f446e07d7-utilities\") pod \"community-operators-72gcl\" (UID: \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\") " pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:20 crc kubenswrapper[4632]: I0313 11:48:20.003755 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d87f4ca1-c949-445e-86d5-ec3f446e07d7-catalog-content\") pod \"community-operators-72gcl\" (UID: \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\") " pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:20 crc kubenswrapper[4632]: I0313 11:48:20.003821 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d87f4ca1-c949-445e-86d5-ec3f446e07d7-utilities\") pod \"community-operators-72gcl\" (UID: \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\") " pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:20 crc kubenswrapper[4632]: I0313 11:48:20.024121 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7sqz2\" (UniqueName: \"kubernetes.io/projected/d87f4ca1-c949-445e-86d5-ec3f446e07d7-kube-api-access-7sqz2\") pod \"community-operators-72gcl\" (UID: \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\") " pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:20 crc kubenswrapper[4632]: I0313 11:48:20.101865 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:20 crc kubenswrapper[4632]: I0313 11:48:20.673283 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-72gcl"] Mar 13 11:48:21 crc kubenswrapper[4632]: I0313 11:48:21.682409 4632 generic.go:334] "Generic (PLEG): container finished" podID="d87f4ca1-c949-445e-86d5-ec3f446e07d7" containerID="2194314bd57a0d0a7f1493e557a6ebb9931f0f83fb3a3632d522d74b98d03d0f" exitCode=0 Mar 13 11:48:21 crc kubenswrapper[4632]: I0313 11:48:21.682488 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72gcl" event={"ID":"d87f4ca1-c949-445e-86d5-ec3f446e07d7","Type":"ContainerDied","Data":"2194314bd57a0d0a7f1493e557a6ebb9931f0f83fb3a3632d522d74b98d03d0f"} Mar 13 11:48:21 crc kubenswrapper[4632]: I0313 11:48:21.682652 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72gcl" event={"ID":"d87f4ca1-c949-445e-86d5-ec3f446e07d7","Type":"ContainerStarted","Data":"9b74ae1b97730cf4914fb6b8ac551cf06716d9c442d8192bb2995649e62562de"} Mar 13 11:48:22 crc kubenswrapper[4632]: I0313 11:48:22.592388 4632 scope.go:117] "RemoveContainer" containerID="6ff82271933ceb662a6f6b867ecc2729be9d4acd3b4299ec77fdefa80de44bf3" Mar 13 11:48:23 crc kubenswrapper[4632]: I0313 11:48:23.707680 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72gcl" event={"ID":"d87f4ca1-c949-445e-86d5-ec3f446e07d7","Type":"ContainerStarted","Data":"5cab776fc2c7b37147eb6269a29d34f464d6a44dbfff425d5956d4872660b7ab"} Mar 13 11:48:25 crc kubenswrapper[4632]: I0313 11:48:25.726153 4632 generic.go:334] "Generic (PLEG): container finished" podID="d87f4ca1-c949-445e-86d5-ec3f446e07d7" containerID="5cab776fc2c7b37147eb6269a29d34f464d6a44dbfff425d5956d4872660b7ab" exitCode=0 Mar 13 11:48:25 crc kubenswrapper[4632]: I0313 11:48:25.726232 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72gcl" event={"ID":"d87f4ca1-c949-445e-86d5-ec3f446e07d7","Type":"ContainerDied","Data":"5cab776fc2c7b37147eb6269a29d34f464d6a44dbfff425d5956d4872660b7ab"} Mar 13 11:48:27 crc kubenswrapper[4632]: I0313 11:48:27.745867 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72gcl" event={"ID":"d87f4ca1-c949-445e-86d5-ec3f446e07d7","Type":"ContainerStarted","Data":"989b7b9ba9a1536075bf1e4277ed84ded29045da7f5e204ef2c5670f8f34cbe5"} Mar 13 11:48:27 crc kubenswrapper[4632]: I0313 11:48:27.773259 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-72gcl" podStartSLOduration=3.616099 podStartE2EDuration="8.773235275s" podCreationTimestamp="2026-03-13 11:48:19 +0000 UTC" firstStartedPulling="2026-03-13 11:48:21.685359524 +0000 UTC m=+6275.707889657" lastFinishedPulling="2026-03-13 11:48:26.842495799 +0000 UTC m=+6280.865025932" observedRunningTime="2026-03-13 11:48:27.76936213 +0000 UTC m=+6281.791892283" watchObservedRunningTime="2026-03-13 11:48:27.773235275 +0000 UTC m=+6281.795765408" Mar 13 11:48:30 crc kubenswrapper[4632]: I0313 11:48:30.102711 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:30 crc kubenswrapper[4632]: I0313 11:48:30.103372 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:31 crc kubenswrapper[4632]: I0313 11:48:31.167221 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-72gcl" podUID="d87f4ca1-c949-445e-86d5-ec3f446e07d7" containerName="registry-server" probeResult="failure" output=< Mar 13 11:48:31 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:48:31 crc kubenswrapper[4632]: > Mar 13 11:48:40 crc kubenswrapper[4632]: I0313 11:48:40.148578 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:40 crc kubenswrapper[4632]: I0313 11:48:40.197785 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:40 crc kubenswrapper[4632]: I0313 11:48:40.407058 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-72gcl"] Mar 13 11:48:40 crc kubenswrapper[4632]: I0313 11:48:40.461192 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:48:40 crc kubenswrapper[4632]: I0313 11:48:40.461254 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:48:40 crc kubenswrapper[4632]: I0313 11:48:40.461305 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 11:48:40 crc kubenswrapper[4632]: I0313 11:48:40.482292 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a91f451a2842f8b1b73b10a99ff94ea342a17276601161b96bf6802b9f5327a9"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 11:48:40 crc kubenswrapper[4632]: I0313 11:48:40.482652 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://a91f451a2842f8b1b73b10a99ff94ea342a17276601161b96bf6802b9f5327a9" gracePeriod=600 Mar 13 11:48:40 crc kubenswrapper[4632]: I0313 11:48:40.877845 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="a91f451a2842f8b1b73b10a99ff94ea342a17276601161b96bf6802b9f5327a9" exitCode=0 Mar 13 11:48:40 crc kubenswrapper[4632]: I0313 11:48:40.877954 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"a91f451a2842f8b1b73b10a99ff94ea342a17276601161b96bf6802b9f5327a9"} Mar 13 11:48:40 crc kubenswrapper[4632]: I0313 11:48:40.878422 4632 scope.go:117] "RemoveContainer" containerID="8e0b51539a4ce69896fef2ee7c7e710d1eb74e5257b7d06373268059e30a34f6" Mar 13 11:48:41 crc kubenswrapper[4632]: I0313 11:48:41.894077 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6"} Mar 13 11:48:41 crc kubenswrapper[4632]: I0313 11:48:41.894233 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-72gcl" podUID="d87f4ca1-c949-445e-86d5-ec3f446e07d7" containerName="registry-server" containerID="cri-o://989b7b9ba9a1536075bf1e4277ed84ded29045da7f5e204ef2c5670f8f34cbe5" gracePeriod=2 Mar 13 11:48:42 crc kubenswrapper[4632]: I0313 11:48:42.837172 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:42 crc kubenswrapper[4632]: I0313 11:48:42.906044 4632 generic.go:334] "Generic (PLEG): container finished" podID="d87f4ca1-c949-445e-86d5-ec3f446e07d7" containerID="989b7b9ba9a1536075bf1e4277ed84ded29045da7f5e204ef2c5670f8f34cbe5" exitCode=0 Mar 13 11:48:42 crc kubenswrapper[4632]: I0313 11:48:42.906129 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-72gcl" Mar 13 11:48:42 crc kubenswrapper[4632]: I0313 11:48:42.906158 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72gcl" event={"ID":"d87f4ca1-c949-445e-86d5-ec3f446e07d7","Type":"ContainerDied","Data":"989b7b9ba9a1536075bf1e4277ed84ded29045da7f5e204ef2c5670f8f34cbe5"} Mar 13 11:48:42 crc kubenswrapper[4632]: I0313 11:48:42.907524 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-72gcl" event={"ID":"d87f4ca1-c949-445e-86d5-ec3f446e07d7","Type":"ContainerDied","Data":"9b74ae1b97730cf4914fb6b8ac551cf06716d9c442d8192bb2995649e62562de"} Mar 13 11:48:42 crc kubenswrapper[4632]: I0313 11:48:42.907592 4632 scope.go:117] "RemoveContainer" containerID="989b7b9ba9a1536075bf1e4277ed84ded29045da7f5e204ef2c5670f8f34cbe5" Mar 13 11:48:42 crc kubenswrapper[4632]: I0313 11:48:42.932746 4632 scope.go:117] "RemoveContainer" containerID="5cab776fc2c7b37147eb6269a29d34f464d6a44dbfff425d5956d4872660b7ab" Mar 13 11:48:42 crc kubenswrapper[4632]: I0313 11:48:42.957764 4632 scope.go:117] "RemoveContainer" containerID="2194314bd57a0d0a7f1493e557a6ebb9931f0f83fb3a3632d522d74b98d03d0f" Mar 13 11:48:42 crc kubenswrapper[4632]: I0313 11:48:42.969327 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d87f4ca1-c949-445e-86d5-ec3f446e07d7-utilities\") pod \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\" (UID: \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\") " Mar 13 11:48:42 crc kubenswrapper[4632]: I0313 11:48:42.969689 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7sqz2\" (UniqueName: \"kubernetes.io/projected/d87f4ca1-c949-445e-86d5-ec3f446e07d7-kube-api-access-7sqz2\") pod \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\" (UID: \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\") " Mar 13 11:48:42 crc kubenswrapper[4632]: I0313 11:48:42.970019 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d87f4ca1-c949-445e-86d5-ec3f446e07d7-catalog-content\") pod \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\" (UID: \"d87f4ca1-c949-445e-86d5-ec3f446e07d7\") " Mar 13 11:48:42 crc kubenswrapper[4632]: I0313 11:48:42.970159 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d87f4ca1-c949-445e-86d5-ec3f446e07d7-utilities" (OuterVolumeSpecName: "utilities") pod "d87f4ca1-c949-445e-86d5-ec3f446e07d7" (UID: "d87f4ca1-c949-445e-86d5-ec3f446e07d7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:48:42 crc kubenswrapper[4632]: I0313 11:48:42.970711 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d87f4ca1-c949-445e-86d5-ec3f446e07d7-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:48:42 crc kubenswrapper[4632]: I0313 11:48:42.976268 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d87f4ca1-c949-445e-86d5-ec3f446e07d7-kube-api-access-7sqz2" (OuterVolumeSpecName: "kube-api-access-7sqz2") pod "d87f4ca1-c949-445e-86d5-ec3f446e07d7" (UID: "d87f4ca1-c949-445e-86d5-ec3f446e07d7"). InnerVolumeSpecName "kube-api-access-7sqz2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:48:43 crc kubenswrapper[4632]: I0313 11:48:43.025560 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d87f4ca1-c949-445e-86d5-ec3f446e07d7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d87f4ca1-c949-445e-86d5-ec3f446e07d7" (UID: "d87f4ca1-c949-445e-86d5-ec3f446e07d7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:48:43 crc kubenswrapper[4632]: I0313 11:48:43.042668 4632 scope.go:117] "RemoveContainer" containerID="989b7b9ba9a1536075bf1e4277ed84ded29045da7f5e204ef2c5670f8f34cbe5" Mar 13 11:48:43 crc kubenswrapper[4632]: E0313 11:48:43.043684 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"989b7b9ba9a1536075bf1e4277ed84ded29045da7f5e204ef2c5670f8f34cbe5\": container with ID starting with 989b7b9ba9a1536075bf1e4277ed84ded29045da7f5e204ef2c5670f8f34cbe5 not found: ID does not exist" containerID="989b7b9ba9a1536075bf1e4277ed84ded29045da7f5e204ef2c5670f8f34cbe5" Mar 13 11:48:43 crc kubenswrapper[4632]: I0313 11:48:43.043764 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"989b7b9ba9a1536075bf1e4277ed84ded29045da7f5e204ef2c5670f8f34cbe5"} err="failed to get container status \"989b7b9ba9a1536075bf1e4277ed84ded29045da7f5e204ef2c5670f8f34cbe5\": rpc error: code = NotFound desc = could not find container \"989b7b9ba9a1536075bf1e4277ed84ded29045da7f5e204ef2c5670f8f34cbe5\": container with ID starting with 989b7b9ba9a1536075bf1e4277ed84ded29045da7f5e204ef2c5670f8f34cbe5 not found: ID does not exist" Mar 13 11:48:43 crc kubenswrapper[4632]: I0313 11:48:43.043803 4632 scope.go:117] "RemoveContainer" containerID="5cab776fc2c7b37147eb6269a29d34f464d6a44dbfff425d5956d4872660b7ab" Mar 13 11:48:43 crc kubenswrapper[4632]: E0313 11:48:43.044226 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cab776fc2c7b37147eb6269a29d34f464d6a44dbfff425d5956d4872660b7ab\": container with ID starting with 5cab776fc2c7b37147eb6269a29d34f464d6a44dbfff425d5956d4872660b7ab not found: ID does not exist" containerID="5cab776fc2c7b37147eb6269a29d34f464d6a44dbfff425d5956d4872660b7ab" Mar 13 11:48:43 crc kubenswrapper[4632]: I0313 11:48:43.044256 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cab776fc2c7b37147eb6269a29d34f464d6a44dbfff425d5956d4872660b7ab"} err="failed to get container status \"5cab776fc2c7b37147eb6269a29d34f464d6a44dbfff425d5956d4872660b7ab\": rpc error: code = NotFound desc = could not find container \"5cab776fc2c7b37147eb6269a29d34f464d6a44dbfff425d5956d4872660b7ab\": container with ID starting with 5cab776fc2c7b37147eb6269a29d34f464d6a44dbfff425d5956d4872660b7ab not found: ID does not exist" Mar 13 11:48:43 crc kubenswrapper[4632]: I0313 11:48:43.044275 4632 scope.go:117] "RemoveContainer" containerID="2194314bd57a0d0a7f1493e557a6ebb9931f0f83fb3a3632d522d74b98d03d0f" Mar 13 11:48:43 crc kubenswrapper[4632]: E0313 11:48:43.044809 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2194314bd57a0d0a7f1493e557a6ebb9931f0f83fb3a3632d522d74b98d03d0f\": container with ID starting with 2194314bd57a0d0a7f1493e557a6ebb9931f0f83fb3a3632d522d74b98d03d0f not found: ID does not exist" containerID="2194314bd57a0d0a7f1493e557a6ebb9931f0f83fb3a3632d522d74b98d03d0f" Mar 13 11:48:43 crc kubenswrapper[4632]: I0313 11:48:43.044846 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2194314bd57a0d0a7f1493e557a6ebb9931f0f83fb3a3632d522d74b98d03d0f"} err="failed to get container status \"2194314bd57a0d0a7f1493e557a6ebb9931f0f83fb3a3632d522d74b98d03d0f\": rpc error: code = NotFound desc = could not find container \"2194314bd57a0d0a7f1493e557a6ebb9931f0f83fb3a3632d522d74b98d03d0f\": container with ID starting with 2194314bd57a0d0a7f1493e557a6ebb9931f0f83fb3a3632d522d74b98d03d0f not found: ID does not exist" Mar 13 11:48:43 crc kubenswrapper[4632]: I0313 11:48:43.072420 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d87f4ca1-c949-445e-86d5-ec3f446e07d7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:48:43 crc kubenswrapper[4632]: I0313 11:48:43.072457 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7sqz2\" (UniqueName: \"kubernetes.io/projected/d87f4ca1-c949-445e-86d5-ec3f446e07d7-kube-api-access-7sqz2\") on node \"crc\" DevicePath \"\"" Mar 13 11:48:43 crc kubenswrapper[4632]: I0313 11:48:43.245880 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-72gcl"] Mar 13 11:48:43 crc kubenswrapper[4632]: I0313 11:48:43.254351 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-72gcl"] Mar 13 11:48:44 crc kubenswrapper[4632]: I0313 11:48:44.054737 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d87f4ca1-c949-445e-86d5-ec3f446e07d7" path="/var/lib/kubelet/pods/d87f4ca1-c949-445e-86d5-ec3f446e07d7/volumes" Mar 13 11:50:00 crc kubenswrapper[4632]: I0313 11:50:00.163401 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556710-t4gtk"] Mar 13 11:50:00 crc kubenswrapper[4632]: E0313 11:50:00.164240 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d87f4ca1-c949-445e-86d5-ec3f446e07d7" containerName="extract-content" Mar 13 11:50:00 crc kubenswrapper[4632]: I0313 11:50:00.164253 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d87f4ca1-c949-445e-86d5-ec3f446e07d7" containerName="extract-content" Mar 13 11:50:00 crc kubenswrapper[4632]: E0313 11:50:00.164270 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d87f4ca1-c949-445e-86d5-ec3f446e07d7" containerName="extract-utilities" Mar 13 11:50:00 crc kubenswrapper[4632]: I0313 11:50:00.164355 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d87f4ca1-c949-445e-86d5-ec3f446e07d7" containerName="extract-utilities" Mar 13 11:50:00 crc kubenswrapper[4632]: E0313 11:50:00.164371 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d87f4ca1-c949-445e-86d5-ec3f446e07d7" containerName="registry-server" Mar 13 11:50:00 crc kubenswrapper[4632]: I0313 11:50:00.164380 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d87f4ca1-c949-445e-86d5-ec3f446e07d7" containerName="registry-server" Mar 13 11:50:00 crc kubenswrapper[4632]: I0313 11:50:00.164639 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="d87f4ca1-c949-445e-86d5-ec3f446e07d7" containerName="registry-server" Mar 13 11:50:00 crc kubenswrapper[4632]: I0313 11:50:00.165755 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556710-t4gtk" Mar 13 11:50:00 crc kubenswrapper[4632]: I0313 11:50:00.179515 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556710-t4gtk"] Mar 13 11:50:00 crc kubenswrapper[4632]: I0313 11:50:00.179884 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:50:00 crc kubenswrapper[4632]: I0313 11:50:00.180196 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:50:00 crc kubenswrapper[4632]: I0313 11:50:00.180343 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:50:00 crc kubenswrapper[4632]: I0313 11:50:00.323032 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4h429\" (UniqueName: \"kubernetes.io/projected/38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca-kube-api-access-4h429\") pod \"auto-csr-approver-29556710-t4gtk\" (UID: \"38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca\") " pod="openshift-infra/auto-csr-approver-29556710-t4gtk" Mar 13 11:50:00 crc kubenswrapper[4632]: I0313 11:50:00.425478 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4h429\" (UniqueName: \"kubernetes.io/projected/38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca-kube-api-access-4h429\") pod \"auto-csr-approver-29556710-t4gtk\" (UID: \"38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca\") " pod="openshift-infra/auto-csr-approver-29556710-t4gtk" Mar 13 11:50:00 crc kubenswrapper[4632]: I0313 11:50:00.459763 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4h429\" (UniqueName: \"kubernetes.io/projected/38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca-kube-api-access-4h429\") pod \"auto-csr-approver-29556710-t4gtk\" (UID: \"38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca\") " pod="openshift-infra/auto-csr-approver-29556710-t4gtk" Mar 13 11:50:00 crc kubenswrapper[4632]: I0313 11:50:00.494677 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556710-t4gtk" Mar 13 11:50:01 crc kubenswrapper[4632]: I0313 11:50:01.002392 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556710-t4gtk"] Mar 13 11:50:01 crc kubenswrapper[4632]: I0313 11:50:01.012300 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 11:50:01 crc kubenswrapper[4632]: W0313 11:50:01.012135 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38d73c58_f065_4efc_9fe2_b6c0ed9fa5ca.slice/crio-a0ffd34e6d6d3b29265d6aeaac0f5a22d37f711ef23a833e058ddb0a16cb2146 WatchSource:0}: Error finding container a0ffd34e6d6d3b29265d6aeaac0f5a22d37f711ef23a833e058ddb0a16cb2146: Status 404 returned error can't find the container with id a0ffd34e6d6d3b29265d6aeaac0f5a22d37f711ef23a833e058ddb0a16cb2146 Mar 13 11:50:01 crc kubenswrapper[4632]: I0313 11:50:01.659270 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556710-t4gtk" event={"ID":"38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca","Type":"ContainerStarted","Data":"a0ffd34e6d6d3b29265d6aeaac0f5a22d37f711ef23a833e058ddb0a16cb2146"} Mar 13 11:50:02 crc kubenswrapper[4632]: I0313 11:50:02.672038 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556710-t4gtk" event={"ID":"38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca","Type":"ContainerStarted","Data":"57e073e1e04617c49dfbb2c194d77f02cda77aac917eb626f73490bd0abacbcb"} Mar 13 11:50:02 crc kubenswrapper[4632]: I0313 11:50:02.702932 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556710-t4gtk" podStartSLOduration=1.5159475059999998 podStartE2EDuration="2.702910372s" podCreationTimestamp="2026-03-13 11:50:00 +0000 UTC" firstStartedPulling="2026-03-13 11:50:01.012078427 +0000 UTC m=+6375.034608560" lastFinishedPulling="2026-03-13 11:50:02.199041293 +0000 UTC m=+6376.221571426" observedRunningTime="2026-03-13 11:50:02.691547343 +0000 UTC m=+6376.714077486" watchObservedRunningTime="2026-03-13 11:50:02.702910372 +0000 UTC m=+6376.725440515" Mar 13 11:50:03 crc kubenswrapper[4632]: I0313 11:50:03.682633 4632 generic.go:334] "Generic (PLEG): container finished" podID="38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca" containerID="57e073e1e04617c49dfbb2c194d77f02cda77aac917eb626f73490bd0abacbcb" exitCode=0 Mar 13 11:50:03 crc kubenswrapper[4632]: I0313 11:50:03.682750 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556710-t4gtk" event={"ID":"38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca","Type":"ContainerDied","Data":"57e073e1e04617c49dfbb2c194d77f02cda77aac917eb626f73490bd0abacbcb"} Mar 13 11:50:05 crc kubenswrapper[4632]: I0313 11:50:05.056733 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556710-t4gtk" Mar 13 11:50:05 crc kubenswrapper[4632]: I0313 11:50:05.240101 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4h429\" (UniqueName: \"kubernetes.io/projected/38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca-kube-api-access-4h429\") pod \"38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca\" (UID: \"38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca\") " Mar 13 11:50:05 crc kubenswrapper[4632]: I0313 11:50:05.252284 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca-kube-api-access-4h429" (OuterVolumeSpecName: "kube-api-access-4h429") pod "38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca" (UID: "38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca"). InnerVolumeSpecName "kube-api-access-4h429". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:50:05 crc kubenswrapper[4632]: I0313 11:50:05.343219 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4h429\" (UniqueName: \"kubernetes.io/projected/38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca-kube-api-access-4h429\") on node \"crc\" DevicePath \"\"" Mar 13 11:50:05 crc kubenswrapper[4632]: I0313 11:50:05.703002 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556710-t4gtk" event={"ID":"38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca","Type":"ContainerDied","Data":"a0ffd34e6d6d3b29265d6aeaac0f5a22d37f711ef23a833e058ddb0a16cb2146"} Mar 13 11:50:05 crc kubenswrapper[4632]: I0313 11:50:05.703350 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0ffd34e6d6d3b29265d6aeaac0f5a22d37f711ef23a833e058ddb0a16cb2146" Mar 13 11:50:05 crc kubenswrapper[4632]: I0313 11:50:05.703045 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556710-t4gtk" Mar 13 11:50:05 crc kubenswrapper[4632]: I0313 11:50:05.763267 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556704-r82jw"] Mar 13 11:50:05 crc kubenswrapper[4632]: I0313 11:50:05.780758 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556704-r82jw"] Mar 13 11:50:06 crc kubenswrapper[4632]: I0313 11:50:06.066856 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6300cb33-fba3-4d08-948b-0c6584d2ef26" path="/var/lib/kubelet/pods/6300cb33-fba3-4d08-948b-0c6584d2ef26/volumes" Mar 13 11:50:22 crc kubenswrapper[4632]: I0313 11:50:22.805367 4632 scope.go:117] "RemoveContainer" containerID="be61603d8895006f002bb067cf8ee34fe273ab45dfec9b8aa73261bbdd0ea048" Mar 13 11:50:40 crc kubenswrapper[4632]: I0313 11:50:40.461622 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:50:40 crc kubenswrapper[4632]: I0313 11:50:40.462300 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.021917 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-t9ff9"] Mar 13 11:50:44 crc kubenswrapper[4632]: E0313 11:50:44.023044 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca" containerName="oc" Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.023059 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca" containerName="oc" Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.023312 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca" containerName="oc" Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.025564 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.068753 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t9ff9"] Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.161672 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvw6j\" (UniqueName: \"kubernetes.io/projected/c4fa6838-8789-4c78-873b-26a25f0abdf1-kube-api-access-hvw6j\") pod \"certified-operators-t9ff9\" (UID: \"c4fa6838-8789-4c78-873b-26a25f0abdf1\") " pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.161737 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4fa6838-8789-4c78-873b-26a25f0abdf1-utilities\") pod \"certified-operators-t9ff9\" (UID: \"c4fa6838-8789-4c78-873b-26a25f0abdf1\") " pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.161850 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4fa6838-8789-4c78-873b-26a25f0abdf1-catalog-content\") pod \"certified-operators-t9ff9\" (UID: \"c4fa6838-8789-4c78-873b-26a25f0abdf1\") " pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.264117 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4fa6838-8789-4c78-873b-26a25f0abdf1-catalog-content\") pod \"certified-operators-t9ff9\" (UID: \"c4fa6838-8789-4c78-873b-26a25f0abdf1\") " pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.264208 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvw6j\" (UniqueName: \"kubernetes.io/projected/c4fa6838-8789-4c78-873b-26a25f0abdf1-kube-api-access-hvw6j\") pod \"certified-operators-t9ff9\" (UID: \"c4fa6838-8789-4c78-873b-26a25f0abdf1\") " pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.264241 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4fa6838-8789-4c78-873b-26a25f0abdf1-utilities\") pod \"certified-operators-t9ff9\" (UID: \"c4fa6838-8789-4c78-873b-26a25f0abdf1\") " pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.264682 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4fa6838-8789-4c78-873b-26a25f0abdf1-utilities\") pod \"certified-operators-t9ff9\" (UID: \"c4fa6838-8789-4c78-873b-26a25f0abdf1\") " pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.264859 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4fa6838-8789-4c78-873b-26a25f0abdf1-catalog-content\") pod \"certified-operators-t9ff9\" (UID: \"c4fa6838-8789-4c78-873b-26a25f0abdf1\") " pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.389650 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvw6j\" (UniqueName: \"kubernetes.io/projected/c4fa6838-8789-4c78-873b-26a25f0abdf1-kube-api-access-hvw6j\") pod \"certified-operators-t9ff9\" (UID: \"c4fa6838-8789-4c78-873b-26a25f0abdf1\") " pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:50:44 crc kubenswrapper[4632]: I0313 11:50:44.664380 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:50:45 crc kubenswrapper[4632]: I0313 11:50:45.455285 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-t9ff9"] Mar 13 11:50:45 crc kubenswrapper[4632]: I0313 11:50:45.536889 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9ff9" event={"ID":"c4fa6838-8789-4c78-873b-26a25f0abdf1","Type":"ContainerStarted","Data":"80d8cf95439ed179efd6dc42eb5f2fb0a7e3f615b39f025e0434d011b74ad3da"} Mar 13 11:50:46 crc kubenswrapper[4632]: I0313 11:50:46.547975 4632 generic.go:334] "Generic (PLEG): container finished" podID="c4fa6838-8789-4c78-873b-26a25f0abdf1" containerID="a5633eaa5f191129873caf3bceb3e511fa8394fc5d911663daf8420e23f54b2c" exitCode=0 Mar 13 11:50:46 crc kubenswrapper[4632]: I0313 11:50:46.548134 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9ff9" event={"ID":"c4fa6838-8789-4c78-873b-26a25f0abdf1","Type":"ContainerDied","Data":"a5633eaa5f191129873caf3bceb3e511fa8394fc5d911663daf8420e23f54b2c"} Mar 13 11:50:48 crc kubenswrapper[4632]: I0313 11:50:48.570454 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9ff9" event={"ID":"c4fa6838-8789-4c78-873b-26a25f0abdf1","Type":"ContainerStarted","Data":"80a53111cc96cd63ea8104fbe328156cf485cbea590d1fee2e9ca120cca06ac5"} Mar 13 11:50:51 crc kubenswrapper[4632]: I0313 11:50:51.598435 4632 generic.go:334] "Generic (PLEG): container finished" podID="c4fa6838-8789-4c78-873b-26a25f0abdf1" containerID="80a53111cc96cd63ea8104fbe328156cf485cbea590d1fee2e9ca120cca06ac5" exitCode=0 Mar 13 11:50:51 crc kubenswrapper[4632]: I0313 11:50:51.598507 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9ff9" event={"ID":"c4fa6838-8789-4c78-873b-26a25f0abdf1","Type":"ContainerDied","Data":"80a53111cc96cd63ea8104fbe328156cf485cbea590d1fee2e9ca120cca06ac5"} Mar 13 11:50:52 crc kubenswrapper[4632]: I0313 11:50:52.611720 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9ff9" event={"ID":"c4fa6838-8789-4c78-873b-26a25f0abdf1","Type":"ContainerStarted","Data":"0f58089f35227117cdd24ba80408858b56b232cdfd763e852159342974bc1d04"} Mar 13 11:50:52 crc kubenswrapper[4632]: I0313 11:50:52.665057 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-t9ff9" podStartSLOduration=4.008162918 podStartE2EDuration="9.66503438s" podCreationTimestamp="2026-03-13 11:50:43 +0000 UTC" firstStartedPulling="2026-03-13 11:50:46.550250558 +0000 UTC m=+6420.572780701" lastFinishedPulling="2026-03-13 11:50:52.20712203 +0000 UTC m=+6426.229652163" observedRunningTime="2026-03-13 11:50:52.659522864 +0000 UTC m=+6426.682053007" watchObservedRunningTime="2026-03-13 11:50:52.66503438 +0000 UTC m=+6426.687564513" Mar 13 11:50:54 crc kubenswrapper[4632]: I0313 11:50:54.665133 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:50:54 crc kubenswrapper[4632]: I0313 11:50:54.665614 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:50:55 crc kubenswrapper[4632]: I0313 11:50:55.228670 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hh6sg"] Mar 13 11:50:55 crc kubenswrapper[4632]: I0313 11:50:55.230693 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:50:55 crc kubenswrapper[4632]: I0313 11:50:55.251196 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hh6sg"] Mar 13 11:50:55 crc kubenswrapper[4632]: I0313 11:50:55.290487 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-catalog-content\") pod \"redhat-operators-hh6sg\" (UID: \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\") " pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:50:55 crc kubenswrapper[4632]: I0313 11:50:55.290746 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-utilities\") pod \"redhat-operators-hh6sg\" (UID: \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\") " pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:50:55 crc kubenswrapper[4632]: I0313 11:50:55.290825 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2vhf\" (UniqueName: \"kubernetes.io/projected/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-kube-api-access-n2vhf\") pod \"redhat-operators-hh6sg\" (UID: \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\") " pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:50:55 crc kubenswrapper[4632]: I0313 11:50:55.392490 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2vhf\" (UniqueName: \"kubernetes.io/projected/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-kube-api-access-n2vhf\") pod \"redhat-operators-hh6sg\" (UID: \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\") " pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:50:55 crc kubenswrapper[4632]: I0313 11:50:55.392643 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-catalog-content\") pod \"redhat-operators-hh6sg\" (UID: \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\") " pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:50:55 crc kubenswrapper[4632]: I0313 11:50:55.392670 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-utilities\") pod \"redhat-operators-hh6sg\" (UID: \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\") " pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:50:55 crc kubenswrapper[4632]: I0313 11:50:55.397021 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-utilities\") pod \"redhat-operators-hh6sg\" (UID: \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\") " pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:50:55 crc kubenswrapper[4632]: I0313 11:50:55.397312 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-catalog-content\") pod \"redhat-operators-hh6sg\" (UID: \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\") " pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:50:55 crc kubenswrapper[4632]: I0313 11:50:55.422467 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2vhf\" (UniqueName: \"kubernetes.io/projected/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-kube-api-access-n2vhf\") pod \"redhat-operators-hh6sg\" (UID: \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\") " pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:50:55 crc kubenswrapper[4632]: I0313 11:50:55.555647 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:50:55 crc kubenswrapper[4632]: I0313 11:50:55.723051 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-t9ff9" podUID="c4fa6838-8789-4c78-873b-26a25f0abdf1" containerName="registry-server" probeResult="failure" output=< Mar 13 11:50:55 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:50:55 crc kubenswrapper[4632]: > Mar 13 11:50:56 crc kubenswrapper[4632]: I0313 11:50:56.741585 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hh6sg"] Mar 13 11:50:57 crc kubenswrapper[4632]: I0313 11:50:57.657663 4632 generic.go:334] "Generic (PLEG): container finished" podID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerID="9c6386e4fe91a3027f76952e4436deca9559a2bb03eb25f5a25ef1623adbd2a9" exitCode=0 Mar 13 11:50:57 crc kubenswrapper[4632]: I0313 11:50:57.657782 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hh6sg" event={"ID":"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9","Type":"ContainerDied","Data":"9c6386e4fe91a3027f76952e4436deca9559a2bb03eb25f5a25ef1623adbd2a9"} Mar 13 11:50:57 crc kubenswrapper[4632]: I0313 11:50:57.658050 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hh6sg" event={"ID":"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9","Type":"ContainerStarted","Data":"0d0fb98fbc9e26b81dbba0ff1bf940b997a137857387dead6a902c8814d568eb"} Mar 13 11:50:58 crc kubenswrapper[4632]: I0313 11:50:58.673394 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hh6sg" event={"ID":"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9","Type":"ContainerStarted","Data":"b2ce3710fc9f4762a6cd2dddd49956965e52724f5194ed7f857b989905d5d86e"} Mar 13 11:51:05 crc kubenswrapper[4632]: I0313 11:51:05.717295 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-t9ff9" podUID="c4fa6838-8789-4c78-873b-26a25f0abdf1" containerName="registry-server" probeResult="failure" output=< Mar 13 11:51:05 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:51:05 crc kubenswrapper[4632]: > Mar 13 11:51:05 crc kubenswrapper[4632]: I0313 11:51:05.746374 4632 generic.go:334] "Generic (PLEG): container finished" podID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerID="b2ce3710fc9f4762a6cd2dddd49956965e52724f5194ed7f857b989905d5d86e" exitCode=0 Mar 13 11:51:05 crc kubenswrapper[4632]: I0313 11:51:05.746418 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hh6sg" event={"ID":"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9","Type":"ContainerDied","Data":"b2ce3710fc9f4762a6cd2dddd49956965e52724f5194ed7f857b989905d5d86e"} Mar 13 11:51:06 crc kubenswrapper[4632]: I0313 11:51:06.757278 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hh6sg" event={"ID":"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9","Type":"ContainerStarted","Data":"bfcfca447a7d94e7df3e9d778de055bafa913106d63c3b943bb83e244830618d"} Mar 13 11:51:06 crc kubenswrapper[4632]: I0313 11:51:06.786494 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hh6sg" podStartSLOduration=3.191258481 podStartE2EDuration="11.786465862s" podCreationTimestamp="2026-03-13 11:50:55 +0000 UTC" firstStartedPulling="2026-03-13 11:50:57.659862755 +0000 UTC m=+6431.682392888" lastFinishedPulling="2026-03-13 11:51:06.255070136 +0000 UTC m=+6440.277600269" observedRunningTime="2026-03-13 11:51:06.775023691 +0000 UTC m=+6440.797553824" watchObservedRunningTime="2026-03-13 11:51:06.786465862 +0000 UTC m=+6440.808995995" Mar 13 11:51:10 crc kubenswrapper[4632]: I0313 11:51:10.461616 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:51:10 crc kubenswrapper[4632]: I0313 11:51:10.462005 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:51:15 crc kubenswrapper[4632]: I0313 11:51:15.556879 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:51:15 crc kubenswrapper[4632]: I0313 11:51:15.557483 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:51:15 crc kubenswrapper[4632]: I0313 11:51:15.712524 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-t9ff9" podUID="c4fa6838-8789-4c78-873b-26a25f0abdf1" containerName="registry-server" probeResult="failure" output=< Mar 13 11:51:15 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:51:15 crc kubenswrapper[4632]: > Mar 13 11:51:16 crc kubenswrapper[4632]: I0313 11:51:16.612008 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hh6sg" podUID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerName="registry-server" probeResult="failure" output=< Mar 13 11:51:16 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:51:16 crc kubenswrapper[4632]: > Mar 13 11:51:24 crc kubenswrapper[4632]: I0313 11:51:24.743061 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:51:24 crc kubenswrapper[4632]: I0313 11:51:24.844173 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:51:24 crc kubenswrapper[4632]: I0313 11:51:24.992744 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t9ff9"] Mar 13 11:51:25 crc kubenswrapper[4632]: I0313 11:51:25.947980 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-t9ff9" podUID="c4fa6838-8789-4c78-873b-26a25f0abdf1" containerName="registry-server" containerID="cri-o://0f58089f35227117cdd24ba80408858b56b232cdfd763e852159342974bc1d04" gracePeriod=2 Mar 13 11:51:26 crc kubenswrapper[4632]: I0313 11:51:26.604715 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hh6sg" podUID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerName="registry-server" probeResult="failure" output=< Mar 13 11:51:26 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:51:26 crc kubenswrapper[4632]: > Mar 13 11:51:26 crc kubenswrapper[4632]: I0313 11:51:26.746740 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:51:26 crc kubenswrapper[4632]: I0313 11:51:26.948461 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4fa6838-8789-4c78-873b-26a25f0abdf1-catalog-content\") pod \"c4fa6838-8789-4c78-873b-26a25f0abdf1\" (UID: \"c4fa6838-8789-4c78-873b-26a25f0abdf1\") " Mar 13 11:51:26 crc kubenswrapper[4632]: I0313 11:51:26.949702 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvw6j\" (UniqueName: \"kubernetes.io/projected/c4fa6838-8789-4c78-873b-26a25f0abdf1-kube-api-access-hvw6j\") pod \"c4fa6838-8789-4c78-873b-26a25f0abdf1\" (UID: \"c4fa6838-8789-4c78-873b-26a25f0abdf1\") " Mar 13 11:51:26 crc kubenswrapper[4632]: I0313 11:51:26.950534 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4fa6838-8789-4c78-873b-26a25f0abdf1-utilities\") pod \"c4fa6838-8789-4c78-873b-26a25f0abdf1\" (UID: \"c4fa6838-8789-4c78-873b-26a25f0abdf1\") " Mar 13 11:51:26 crc kubenswrapper[4632]: I0313 11:51:26.951009 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4fa6838-8789-4c78-873b-26a25f0abdf1-utilities" (OuterVolumeSpecName: "utilities") pod "c4fa6838-8789-4c78-873b-26a25f0abdf1" (UID: "c4fa6838-8789-4c78-873b-26a25f0abdf1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:51:26 crc kubenswrapper[4632]: I0313 11:51:26.951759 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c4fa6838-8789-4c78-873b-26a25f0abdf1-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:51:26 crc kubenswrapper[4632]: I0313 11:51:26.958877 4632 generic.go:334] "Generic (PLEG): container finished" podID="c4fa6838-8789-4c78-873b-26a25f0abdf1" containerID="0f58089f35227117cdd24ba80408858b56b232cdfd763e852159342974bc1d04" exitCode=0 Mar 13 11:51:26 crc kubenswrapper[4632]: I0313 11:51:26.959284 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9ff9" event={"ID":"c4fa6838-8789-4c78-873b-26a25f0abdf1","Type":"ContainerDied","Data":"0f58089f35227117cdd24ba80408858b56b232cdfd763e852159342974bc1d04"} Mar 13 11:51:26 crc kubenswrapper[4632]: I0313 11:51:26.959358 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-t9ff9" event={"ID":"c4fa6838-8789-4c78-873b-26a25f0abdf1","Type":"ContainerDied","Data":"80d8cf95439ed179efd6dc42eb5f2fb0a7e3f615b39f025e0434d011b74ad3da"} Mar 13 11:51:26 crc kubenswrapper[4632]: I0313 11:51:26.959388 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-t9ff9" Mar 13 11:51:26 crc kubenswrapper[4632]: I0313 11:51:26.960699 4632 scope.go:117] "RemoveContainer" containerID="0f58089f35227117cdd24ba80408858b56b232cdfd763e852159342974bc1d04" Mar 13 11:51:26 crc kubenswrapper[4632]: I0313 11:51:26.969190 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4fa6838-8789-4c78-873b-26a25f0abdf1-kube-api-access-hvw6j" (OuterVolumeSpecName: "kube-api-access-hvw6j") pod "c4fa6838-8789-4c78-873b-26a25f0abdf1" (UID: "c4fa6838-8789-4c78-873b-26a25f0abdf1"). InnerVolumeSpecName "kube-api-access-hvw6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:51:27 crc kubenswrapper[4632]: I0313 11:51:27.044225 4632 scope.go:117] "RemoveContainer" containerID="80a53111cc96cd63ea8104fbe328156cf485cbea590d1fee2e9ca120cca06ac5" Mar 13 11:51:27 crc kubenswrapper[4632]: I0313 11:51:27.046685 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c4fa6838-8789-4c78-873b-26a25f0abdf1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c4fa6838-8789-4c78-873b-26a25f0abdf1" (UID: "c4fa6838-8789-4c78-873b-26a25f0abdf1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:51:27 crc kubenswrapper[4632]: I0313 11:51:27.053642 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c4fa6838-8789-4c78-873b-26a25f0abdf1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:51:27 crc kubenswrapper[4632]: I0313 11:51:27.053875 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvw6j\" (UniqueName: \"kubernetes.io/projected/c4fa6838-8789-4c78-873b-26a25f0abdf1-kube-api-access-hvw6j\") on node \"crc\" DevicePath \"\"" Mar 13 11:51:27 crc kubenswrapper[4632]: I0313 11:51:27.080013 4632 scope.go:117] "RemoveContainer" containerID="a5633eaa5f191129873caf3bceb3e511fa8394fc5d911663daf8420e23f54b2c" Mar 13 11:51:27 crc kubenswrapper[4632]: I0313 11:51:27.110042 4632 scope.go:117] "RemoveContainer" containerID="0f58089f35227117cdd24ba80408858b56b232cdfd763e852159342974bc1d04" Mar 13 11:51:27 crc kubenswrapper[4632]: E0313 11:51:27.114974 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f58089f35227117cdd24ba80408858b56b232cdfd763e852159342974bc1d04\": container with ID starting with 0f58089f35227117cdd24ba80408858b56b232cdfd763e852159342974bc1d04 not found: ID does not exist" containerID="0f58089f35227117cdd24ba80408858b56b232cdfd763e852159342974bc1d04" Mar 13 11:51:27 crc kubenswrapper[4632]: I0313 11:51:27.116337 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f58089f35227117cdd24ba80408858b56b232cdfd763e852159342974bc1d04"} err="failed to get container status \"0f58089f35227117cdd24ba80408858b56b232cdfd763e852159342974bc1d04\": rpc error: code = NotFound desc = could not find container \"0f58089f35227117cdd24ba80408858b56b232cdfd763e852159342974bc1d04\": container with ID starting with 0f58089f35227117cdd24ba80408858b56b232cdfd763e852159342974bc1d04 not found: ID does not exist" Mar 13 11:51:27 crc kubenswrapper[4632]: I0313 11:51:27.116381 4632 scope.go:117] "RemoveContainer" containerID="80a53111cc96cd63ea8104fbe328156cf485cbea590d1fee2e9ca120cca06ac5" Mar 13 11:51:27 crc kubenswrapper[4632]: E0313 11:51:27.117110 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80a53111cc96cd63ea8104fbe328156cf485cbea590d1fee2e9ca120cca06ac5\": container with ID starting with 80a53111cc96cd63ea8104fbe328156cf485cbea590d1fee2e9ca120cca06ac5 not found: ID does not exist" containerID="80a53111cc96cd63ea8104fbe328156cf485cbea590d1fee2e9ca120cca06ac5" Mar 13 11:51:27 crc kubenswrapper[4632]: I0313 11:51:27.117163 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80a53111cc96cd63ea8104fbe328156cf485cbea590d1fee2e9ca120cca06ac5"} err="failed to get container status \"80a53111cc96cd63ea8104fbe328156cf485cbea590d1fee2e9ca120cca06ac5\": rpc error: code = NotFound desc = could not find container \"80a53111cc96cd63ea8104fbe328156cf485cbea590d1fee2e9ca120cca06ac5\": container with ID starting with 80a53111cc96cd63ea8104fbe328156cf485cbea590d1fee2e9ca120cca06ac5 not found: ID does not exist" Mar 13 11:51:27 crc kubenswrapper[4632]: I0313 11:51:27.117194 4632 scope.go:117] "RemoveContainer" containerID="a5633eaa5f191129873caf3bceb3e511fa8394fc5d911663daf8420e23f54b2c" Mar 13 11:51:27 crc kubenswrapper[4632]: E0313 11:51:27.117649 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5633eaa5f191129873caf3bceb3e511fa8394fc5d911663daf8420e23f54b2c\": container with ID starting with a5633eaa5f191129873caf3bceb3e511fa8394fc5d911663daf8420e23f54b2c not found: ID does not exist" containerID="a5633eaa5f191129873caf3bceb3e511fa8394fc5d911663daf8420e23f54b2c" Mar 13 11:51:27 crc kubenswrapper[4632]: I0313 11:51:27.117684 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5633eaa5f191129873caf3bceb3e511fa8394fc5d911663daf8420e23f54b2c"} err="failed to get container status \"a5633eaa5f191129873caf3bceb3e511fa8394fc5d911663daf8420e23f54b2c\": rpc error: code = NotFound desc = could not find container \"a5633eaa5f191129873caf3bceb3e511fa8394fc5d911663daf8420e23f54b2c\": container with ID starting with a5633eaa5f191129873caf3bceb3e511fa8394fc5d911663daf8420e23f54b2c not found: ID does not exist" Mar 13 11:51:27 crc kubenswrapper[4632]: I0313 11:51:27.299151 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-t9ff9"] Mar 13 11:51:27 crc kubenswrapper[4632]: I0313 11:51:27.334371 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-t9ff9"] Mar 13 11:51:28 crc kubenswrapper[4632]: I0313 11:51:28.056117 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4fa6838-8789-4c78-873b-26a25f0abdf1" path="/var/lib/kubelet/pods/c4fa6838-8789-4c78-873b-26a25f0abdf1/volumes" Mar 13 11:51:36 crc kubenswrapper[4632]: I0313 11:51:36.613577 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hh6sg" podUID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerName="registry-server" probeResult="failure" output=< Mar 13 11:51:36 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:51:36 crc kubenswrapper[4632]: > Mar 13 11:51:40 crc kubenswrapper[4632]: I0313 11:51:40.460955 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:51:40 crc kubenswrapper[4632]: I0313 11:51:40.461460 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:51:40 crc kubenswrapper[4632]: I0313 11:51:40.461524 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 11:51:40 crc kubenswrapper[4632]: I0313 11:51:40.462302 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 11:51:40 crc kubenswrapper[4632]: I0313 11:51:40.462352 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" gracePeriod=600 Mar 13 11:51:40 crc kubenswrapper[4632]: E0313 11:51:40.586564 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:51:41 crc kubenswrapper[4632]: I0313 11:51:41.095667 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" exitCode=0 Mar 13 11:51:41 crc kubenswrapper[4632]: I0313 11:51:41.095734 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6"} Mar 13 11:51:41 crc kubenswrapper[4632]: I0313 11:51:41.095769 4632 scope.go:117] "RemoveContainer" containerID="a91f451a2842f8b1b73b10a99ff94ea342a17276601161b96bf6802b9f5327a9" Mar 13 11:51:41 crc kubenswrapper[4632]: I0313 11:51:41.096520 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:51:41 crc kubenswrapper[4632]: E0313 11:51:41.096806 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:51:46 crc kubenswrapper[4632]: I0313 11:51:46.617396 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hh6sg" podUID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerName="registry-server" probeResult="failure" output=< Mar 13 11:51:46 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:51:46 crc kubenswrapper[4632]: > Mar 13 11:51:52 crc kubenswrapper[4632]: I0313 11:51:52.044808 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:51:52 crc kubenswrapper[4632]: E0313 11:51:52.045637 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:51:55 crc kubenswrapper[4632]: I0313 11:51:55.608057 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:51:55 crc kubenswrapper[4632]: I0313 11:51:55.666842 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:51:55 crc kubenswrapper[4632]: I0313 11:51:55.850537 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hh6sg"] Mar 13 11:51:57 crc kubenswrapper[4632]: I0313 11:51:57.258999 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hh6sg" podUID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerName="registry-server" containerID="cri-o://bfcfca447a7d94e7df3e9d778de055bafa913106d63c3b943bb83e244830618d" gracePeriod=2 Mar 13 11:51:57 crc kubenswrapper[4632]: I0313 11:51:57.898707 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:51:57 crc kubenswrapper[4632]: I0313 11:51:57.996444 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2vhf\" (UniqueName: \"kubernetes.io/projected/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-kube-api-access-n2vhf\") pod \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\" (UID: \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\") " Mar 13 11:51:57 crc kubenswrapper[4632]: I0313 11:51:57.996589 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-utilities\") pod \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\" (UID: \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\") " Mar 13 11:51:57 crc kubenswrapper[4632]: I0313 11:51:57.996717 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-catalog-content\") pod \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\" (UID: \"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9\") " Mar 13 11:51:57 crc kubenswrapper[4632]: I0313 11:51:57.997604 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-utilities" (OuterVolumeSpecName: "utilities") pod "b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" (UID: "b9077b2a-7fb9-405f-8fb4-b472d5ac00a9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.005177 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-kube-api-access-n2vhf" (OuterVolumeSpecName: "kube-api-access-n2vhf") pod "b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" (UID: "b9077b2a-7fb9-405f-8fb4-b472d5ac00a9"). InnerVolumeSpecName "kube-api-access-n2vhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.100480 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2vhf\" (UniqueName: \"kubernetes.io/projected/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-kube-api-access-n2vhf\") on node \"crc\" DevicePath \"\"" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.100529 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.117783 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" (UID: "b9077b2a-7fb9-405f-8fb4-b472d5ac00a9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.202617 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.272002 4632 generic.go:334] "Generic (PLEG): container finished" podID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerID="bfcfca447a7d94e7df3e9d778de055bafa913106d63c3b943bb83e244830618d" exitCode=0 Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.272081 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hh6sg" event={"ID":"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9","Type":"ContainerDied","Data":"bfcfca447a7d94e7df3e9d778de055bafa913106d63c3b943bb83e244830618d"} Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.272115 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hh6sg" event={"ID":"b9077b2a-7fb9-405f-8fb4-b472d5ac00a9","Type":"ContainerDied","Data":"0d0fb98fbc9e26b81dbba0ff1bf940b997a137857387dead6a902c8814d568eb"} Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.272135 4632 scope.go:117] "RemoveContainer" containerID="bfcfca447a7d94e7df3e9d778de055bafa913106d63c3b943bb83e244830618d" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.272080 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hh6sg" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.313726 4632 scope.go:117] "RemoveContainer" containerID="b2ce3710fc9f4762a6cd2dddd49956965e52724f5194ed7f857b989905d5d86e" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.317666 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hh6sg"] Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.333181 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hh6sg"] Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.337993 4632 scope.go:117] "RemoveContainer" containerID="9c6386e4fe91a3027f76952e4436deca9559a2bb03eb25f5a25ef1623adbd2a9" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.385723 4632 scope.go:117] "RemoveContainer" containerID="bfcfca447a7d94e7df3e9d778de055bafa913106d63c3b943bb83e244830618d" Mar 13 11:51:58 crc kubenswrapper[4632]: E0313 11:51:58.386545 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfcfca447a7d94e7df3e9d778de055bafa913106d63c3b943bb83e244830618d\": container with ID starting with bfcfca447a7d94e7df3e9d778de055bafa913106d63c3b943bb83e244830618d not found: ID does not exist" containerID="bfcfca447a7d94e7df3e9d778de055bafa913106d63c3b943bb83e244830618d" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.386578 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfcfca447a7d94e7df3e9d778de055bafa913106d63c3b943bb83e244830618d"} err="failed to get container status \"bfcfca447a7d94e7df3e9d778de055bafa913106d63c3b943bb83e244830618d\": rpc error: code = NotFound desc = could not find container \"bfcfca447a7d94e7df3e9d778de055bafa913106d63c3b943bb83e244830618d\": container with ID starting with bfcfca447a7d94e7df3e9d778de055bafa913106d63c3b943bb83e244830618d not found: ID does not exist" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.386599 4632 scope.go:117] "RemoveContainer" containerID="b2ce3710fc9f4762a6cd2dddd49956965e52724f5194ed7f857b989905d5d86e" Mar 13 11:51:58 crc kubenswrapper[4632]: E0313 11:51:58.387144 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2ce3710fc9f4762a6cd2dddd49956965e52724f5194ed7f857b989905d5d86e\": container with ID starting with b2ce3710fc9f4762a6cd2dddd49956965e52724f5194ed7f857b989905d5d86e not found: ID does not exist" containerID="b2ce3710fc9f4762a6cd2dddd49956965e52724f5194ed7f857b989905d5d86e" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.387231 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2ce3710fc9f4762a6cd2dddd49956965e52724f5194ed7f857b989905d5d86e"} err="failed to get container status \"b2ce3710fc9f4762a6cd2dddd49956965e52724f5194ed7f857b989905d5d86e\": rpc error: code = NotFound desc = could not find container \"b2ce3710fc9f4762a6cd2dddd49956965e52724f5194ed7f857b989905d5d86e\": container with ID starting with b2ce3710fc9f4762a6cd2dddd49956965e52724f5194ed7f857b989905d5d86e not found: ID does not exist" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.387269 4632 scope.go:117] "RemoveContainer" containerID="9c6386e4fe91a3027f76952e4436deca9559a2bb03eb25f5a25ef1623adbd2a9" Mar 13 11:51:58 crc kubenswrapper[4632]: E0313 11:51:58.387800 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9c6386e4fe91a3027f76952e4436deca9559a2bb03eb25f5a25ef1623adbd2a9\": container with ID starting with 9c6386e4fe91a3027f76952e4436deca9559a2bb03eb25f5a25ef1623adbd2a9 not found: ID does not exist" containerID="9c6386e4fe91a3027f76952e4436deca9559a2bb03eb25f5a25ef1623adbd2a9" Mar 13 11:51:58 crc kubenswrapper[4632]: I0313 11:51:58.387835 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9c6386e4fe91a3027f76952e4436deca9559a2bb03eb25f5a25ef1623adbd2a9"} err="failed to get container status \"9c6386e4fe91a3027f76952e4436deca9559a2bb03eb25f5a25ef1623adbd2a9\": rpc error: code = NotFound desc = could not find container \"9c6386e4fe91a3027f76952e4436deca9559a2bb03eb25f5a25ef1623adbd2a9\": container with ID starting with 9c6386e4fe91a3027f76952e4436deca9559a2bb03eb25f5a25ef1623adbd2a9 not found: ID does not exist" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.056104 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" path="/var/lib/kubelet/pods/b9077b2a-7fb9-405f-8fb4-b472d5ac00a9/volumes" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.204372 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556712-jjq26"] Mar 13 11:52:00 crc kubenswrapper[4632]: E0313 11:52:00.204811 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4fa6838-8789-4c78-873b-26a25f0abdf1" containerName="extract-content" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.205743 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4fa6838-8789-4c78-873b-26a25f0abdf1" containerName="extract-content" Mar 13 11:52:00 crc kubenswrapper[4632]: E0313 11:52:00.205795 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerName="extract-utilities" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.205805 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerName="extract-utilities" Mar 13 11:52:00 crc kubenswrapper[4632]: E0313 11:52:00.205828 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4fa6838-8789-4c78-873b-26a25f0abdf1" containerName="extract-utilities" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.205837 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4fa6838-8789-4c78-873b-26a25f0abdf1" containerName="extract-utilities" Mar 13 11:52:00 crc kubenswrapper[4632]: E0313 11:52:00.205852 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerName="extract-content" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.205861 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerName="extract-content" Mar 13 11:52:00 crc kubenswrapper[4632]: E0313 11:52:00.205879 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerName="registry-server" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.205887 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerName="registry-server" Mar 13 11:52:00 crc kubenswrapper[4632]: E0313 11:52:00.205902 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4fa6838-8789-4c78-873b-26a25f0abdf1" containerName="registry-server" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.205909 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4fa6838-8789-4c78-873b-26a25f0abdf1" containerName="registry-server" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.206153 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4fa6838-8789-4c78-873b-26a25f0abdf1" containerName="registry-server" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.206168 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9077b2a-7fb9-405f-8fb4-b472d5ac00a9" containerName="registry-server" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.206864 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556712-jjq26" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.222762 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.224929 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.226383 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.233654 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556712-jjq26"] Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.344370 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmz6r\" (UniqueName: \"kubernetes.io/projected/02db6f2d-ef7c-4444-9776-603e9c44c55a-kube-api-access-vmz6r\") pod \"auto-csr-approver-29556712-jjq26\" (UID: \"02db6f2d-ef7c-4444-9776-603e9c44c55a\") " pod="openshift-infra/auto-csr-approver-29556712-jjq26" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.446525 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmz6r\" (UniqueName: \"kubernetes.io/projected/02db6f2d-ef7c-4444-9776-603e9c44c55a-kube-api-access-vmz6r\") pod \"auto-csr-approver-29556712-jjq26\" (UID: \"02db6f2d-ef7c-4444-9776-603e9c44c55a\") " pod="openshift-infra/auto-csr-approver-29556712-jjq26" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.474323 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmz6r\" (UniqueName: \"kubernetes.io/projected/02db6f2d-ef7c-4444-9776-603e9c44c55a-kube-api-access-vmz6r\") pod \"auto-csr-approver-29556712-jjq26\" (UID: \"02db6f2d-ef7c-4444-9776-603e9c44c55a\") " pod="openshift-infra/auto-csr-approver-29556712-jjq26" Mar 13 11:52:00 crc kubenswrapper[4632]: I0313 11:52:00.530019 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556712-jjq26" Mar 13 11:52:01 crc kubenswrapper[4632]: I0313 11:52:01.150440 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556712-jjq26"] Mar 13 11:52:01 crc kubenswrapper[4632]: I0313 11:52:01.308279 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556712-jjq26" event={"ID":"02db6f2d-ef7c-4444-9776-603e9c44c55a","Type":"ContainerStarted","Data":"e06748d4089f667badd71fb0750ee1ffdc4180e878cbff38f612301d9128cb9a"} Mar 13 11:52:03 crc kubenswrapper[4632]: I0313 11:52:03.327644 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556712-jjq26" event={"ID":"02db6f2d-ef7c-4444-9776-603e9c44c55a","Type":"ContainerStarted","Data":"1b3640b2bd3a5d0dabd874439006e72fb30cf0784909546f4e8957109f6ffcf0"} Mar 13 11:52:03 crc kubenswrapper[4632]: I0313 11:52:03.345252 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556712-jjq26" podStartSLOduration=2.573870574 podStartE2EDuration="3.34523341s" podCreationTimestamp="2026-03-13 11:52:00 +0000 UTC" firstStartedPulling="2026-03-13 11:52:01.163430234 +0000 UTC m=+6495.185960367" lastFinishedPulling="2026-03-13 11:52:01.93479308 +0000 UTC m=+6495.957323203" observedRunningTime="2026-03-13 11:52:03.339284404 +0000 UTC m=+6497.361814537" watchObservedRunningTime="2026-03-13 11:52:03.34523341 +0000 UTC m=+6497.367763543" Mar 13 11:52:04 crc kubenswrapper[4632]: I0313 11:52:04.340033 4632 generic.go:334] "Generic (PLEG): container finished" podID="02db6f2d-ef7c-4444-9776-603e9c44c55a" containerID="1b3640b2bd3a5d0dabd874439006e72fb30cf0784909546f4e8957109f6ffcf0" exitCode=0 Mar 13 11:52:04 crc kubenswrapper[4632]: I0313 11:52:04.340256 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556712-jjq26" event={"ID":"02db6f2d-ef7c-4444-9776-603e9c44c55a","Type":"ContainerDied","Data":"1b3640b2bd3a5d0dabd874439006e72fb30cf0784909546f4e8957109f6ffcf0"} Mar 13 11:52:05 crc kubenswrapper[4632]: I0313 11:52:05.914503 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556712-jjq26" Mar 13 11:52:06 crc kubenswrapper[4632]: I0313 11:52:06.063318 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmz6r\" (UniqueName: \"kubernetes.io/projected/02db6f2d-ef7c-4444-9776-603e9c44c55a-kube-api-access-vmz6r\") pod \"02db6f2d-ef7c-4444-9776-603e9c44c55a\" (UID: \"02db6f2d-ef7c-4444-9776-603e9c44c55a\") " Mar 13 11:52:06 crc kubenswrapper[4632]: I0313 11:52:06.069874 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02db6f2d-ef7c-4444-9776-603e9c44c55a-kube-api-access-vmz6r" (OuterVolumeSpecName: "kube-api-access-vmz6r") pod "02db6f2d-ef7c-4444-9776-603e9c44c55a" (UID: "02db6f2d-ef7c-4444-9776-603e9c44c55a"). InnerVolumeSpecName "kube-api-access-vmz6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:52:06 crc kubenswrapper[4632]: I0313 11:52:06.167419 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmz6r\" (UniqueName: \"kubernetes.io/projected/02db6f2d-ef7c-4444-9776-603e9c44c55a-kube-api-access-vmz6r\") on node \"crc\" DevicePath \"\"" Mar 13 11:52:06 crc kubenswrapper[4632]: I0313 11:52:06.360544 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556712-jjq26" event={"ID":"02db6f2d-ef7c-4444-9776-603e9c44c55a","Type":"ContainerDied","Data":"e06748d4089f667badd71fb0750ee1ffdc4180e878cbff38f612301d9128cb9a"} Mar 13 11:52:06 crc kubenswrapper[4632]: I0313 11:52:06.360597 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e06748d4089f667badd71fb0750ee1ffdc4180e878cbff38f612301d9128cb9a" Mar 13 11:52:06 crc kubenswrapper[4632]: I0313 11:52:06.360669 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556712-jjq26" Mar 13 11:52:06 crc kubenswrapper[4632]: I0313 11:52:06.488487 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556706-psc7t"] Mar 13 11:52:06 crc kubenswrapper[4632]: I0313 11:52:06.496696 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556706-psc7t"] Mar 13 11:52:07 crc kubenswrapper[4632]: I0313 11:52:07.045416 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:52:07 crc kubenswrapper[4632]: E0313 11:52:07.045747 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:52:08 crc kubenswrapper[4632]: I0313 11:52:08.057120 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8980f067-488f-497f-8ba7-5ee2d3069d62" path="/var/lib/kubelet/pods/8980f067-488f-497f-8ba7-5ee2d3069d62/volumes" Mar 13 11:52:19 crc kubenswrapper[4632]: I0313 11:52:19.044154 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:52:19 crc kubenswrapper[4632]: E0313 11:52:19.044963 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:52:23 crc kubenswrapper[4632]: I0313 11:52:23.160077 4632 scope.go:117] "RemoveContainer" containerID="49f7bf435fba27e68a413e86a923b4ddacb7432c6b3ec46cefd0935c8e2aecc2" Mar 13 11:52:31 crc kubenswrapper[4632]: I0313 11:52:31.043881 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:52:31 crc kubenswrapper[4632]: E0313 11:52:31.044572 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:52:44 crc kubenswrapper[4632]: I0313 11:52:44.044766 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:52:44 crc kubenswrapper[4632]: E0313 11:52:44.045501 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:52:58 crc kubenswrapper[4632]: I0313 11:52:58.044260 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:52:58 crc kubenswrapper[4632]: E0313 11:52:58.045049 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:53:10 crc kubenswrapper[4632]: I0313 11:53:10.044315 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:53:10 crc kubenswrapper[4632]: E0313 11:53:10.045237 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:53:25 crc kubenswrapper[4632]: I0313 11:53:25.045275 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:53:25 crc kubenswrapper[4632]: E0313 11:53:25.046516 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:53:40 crc kubenswrapper[4632]: I0313 11:53:40.045319 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:53:40 crc kubenswrapper[4632]: E0313 11:53:40.047025 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:53:52 crc kubenswrapper[4632]: I0313 11:53:52.047323 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:53:52 crc kubenswrapper[4632]: E0313 11:53:52.049092 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:54:00 crc kubenswrapper[4632]: I0313 11:54:00.159745 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556714-cnqc7"] Mar 13 11:54:00 crc kubenswrapper[4632]: E0313 11:54:00.160845 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02db6f2d-ef7c-4444-9776-603e9c44c55a" containerName="oc" Mar 13 11:54:00 crc kubenswrapper[4632]: I0313 11:54:00.160863 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="02db6f2d-ef7c-4444-9776-603e9c44c55a" containerName="oc" Mar 13 11:54:00 crc kubenswrapper[4632]: I0313 11:54:00.161150 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="02db6f2d-ef7c-4444-9776-603e9c44c55a" containerName="oc" Mar 13 11:54:00 crc kubenswrapper[4632]: I0313 11:54:00.162026 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556714-cnqc7" Mar 13 11:54:00 crc kubenswrapper[4632]: I0313 11:54:00.164836 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:54:00 crc kubenswrapper[4632]: I0313 11:54:00.164921 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:54:00 crc kubenswrapper[4632]: I0313 11:54:00.164919 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:54:00 crc kubenswrapper[4632]: I0313 11:54:00.197266 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556714-cnqc7"] Mar 13 11:54:00 crc kubenswrapper[4632]: I0313 11:54:00.284558 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmbkr\" (UniqueName: \"kubernetes.io/projected/88fb48fa-6650-4b22-b44f-d8c6f839489e-kube-api-access-rmbkr\") pod \"auto-csr-approver-29556714-cnqc7\" (UID: \"88fb48fa-6650-4b22-b44f-d8c6f839489e\") " pod="openshift-infra/auto-csr-approver-29556714-cnqc7" Mar 13 11:54:00 crc kubenswrapper[4632]: I0313 11:54:00.387072 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmbkr\" (UniqueName: \"kubernetes.io/projected/88fb48fa-6650-4b22-b44f-d8c6f839489e-kube-api-access-rmbkr\") pod \"auto-csr-approver-29556714-cnqc7\" (UID: \"88fb48fa-6650-4b22-b44f-d8c6f839489e\") " pod="openshift-infra/auto-csr-approver-29556714-cnqc7" Mar 13 11:54:00 crc kubenswrapper[4632]: I0313 11:54:00.408515 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmbkr\" (UniqueName: \"kubernetes.io/projected/88fb48fa-6650-4b22-b44f-d8c6f839489e-kube-api-access-rmbkr\") pod \"auto-csr-approver-29556714-cnqc7\" (UID: \"88fb48fa-6650-4b22-b44f-d8c6f839489e\") " pod="openshift-infra/auto-csr-approver-29556714-cnqc7" Mar 13 11:54:00 crc kubenswrapper[4632]: I0313 11:54:00.482310 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556714-cnqc7" Mar 13 11:54:01 crc kubenswrapper[4632]: I0313 11:54:01.038803 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556714-cnqc7"] Mar 13 11:54:01 crc kubenswrapper[4632]: I0313 11:54:01.580662 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556714-cnqc7" event={"ID":"88fb48fa-6650-4b22-b44f-d8c6f839489e","Type":"ContainerStarted","Data":"a08cce79366424c102d56b429240137bb4b00fe174f555bd016d51d91a7ff38d"} Mar 13 11:54:03 crc kubenswrapper[4632]: I0313 11:54:03.605527 4632 generic.go:334] "Generic (PLEG): container finished" podID="88fb48fa-6650-4b22-b44f-d8c6f839489e" containerID="a65145278e359710e5ff339e23940020997c56d82631e8e73d581b3ec62c80b2" exitCode=0 Mar 13 11:54:03 crc kubenswrapper[4632]: I0313 11:54:03.605933 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556714-cnqc7" event={"ID":"88fb48fa-6650-4b22-b44f-d8c6f839489e","Type":"ContainerDied","Data":"a65145278e359710e5ff339e23940020997c56d82631e8e73d581b3ec62c80b2"} Mar 13 11:54:04 crc kubenswrapper[4632]: I0313 11:54:04.975697 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556714-cnqc7" Mar 13 11:54:05 crc kubenswrapper[4632]: I0313 11:54:05.045053 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:54:05 crc kubenswrapper[4632]: E0313 11:54:05.045328 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:54:05 crc kubenswrapper[4632]: I0313 11:54:05.101767 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmbkr\" (UniqueName: \"kubernetes.io/projected/88fb48fa-6650-4b22-b44f-d8c6f839489e-kube-api-access-rmbkr\") pod \"88fb48fa-6650-4b22-b44f-d8c6f839489e\" (UID: \"88fb48fa-6650-4b22-b44f-d8c6f839489e\") " Mar 13 11:54:05 crc kubenswrapper[4632]: I0313 11:54:05.112449 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88fb48fa-6650-4b22-b44f-d8c6f839489e-kube-api-access-rmbkr" (OuterVolumeSpecName: "kube-api-access-rmbkr") pod "88fb48fa-6650-4b22-b44f-d8c6f839489e" (UID: "88fb48fa-6650-4b22-b44f-d8c6f839489e"). InnerVolumeSpecName "kube-api-access-rmbkr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:54:05 crc kubenswrapper[4632]: I0313 11:54:05.204049 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmbkr\" (UniqueName: \"kubernetes.io/projected/88fb48fa-6650-4b22-b44f-d8c6f839489e-kube-api-access-rmbkr\") on node \"crc\" DevicePath \"\"" Mar 13 11:54:05 crc kubenswrapper[4632]: I0313 11:54:05.628218 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556714-cnqc7" event={"ID":"88fb48fa-6650-4b22-b44f-d8c6f839489e","Type":"ContainerDied","Data":"a08cce79366424c102d56b429240137bb4b00fe174f555bd016d51d91a7ff38d"} Mar 13 11:54:05 crc kubenswrapper[4632]: I0313 11:54:05.628267 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a08cce79366424c102d56b429240137bb4b00fe174f555bd016d51d91a7ff38d" Mar 13 11:54:05 crc kubenswrapper[4632]: I0313 11:54:05.628329 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556714-cnqc7" Mar 13 11:54:06 crc kubenswrapper[4632]: I0313 11:54:06.080227 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556708-f6hsv"] Mar 13 11:54:06 crc kubenswrapper[4632]: I0313 11:54:06.091534 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556708-f6hsv"] Mar 13 11:54:08 crc kubenswrapper[4632]: I0313 11:54:08.068132 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5464d278-31e9-45aa-9e87-78ef3e96115e" path="/var/lib/kubelet/pods/5464d278-31e9-45aa-9e87-78ef3e96115e/volumes" Mar 13 11:54:18 crc kubenswrapper[4632]: I0313 11:54:18.050592 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:54:18 crc kubenswrapper[4632]: E0313 11:54:18.051390 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:54:23 crc kubenswrapper[4632]: I0313 11:54:23.297683 4632 scope.go:117] "RemoveContainer" containerID="6f4e30c3bf10310c255b10e3f6602511c866fd5af961f5f486fed69de586adb4" Mar 13 11:54:33 crc kubenswrapper[4632]: I0313 11:54:33.044833 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:54:33 crc kubenswrapper[4632]: E0313 11:54:33.045602 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:54:48 crc kubenswrapper[4632]: I0313 11:54:48.044338 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:54:48 crc kubenswrapper[4632]: E0313 11:54:48.045318 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:55:00 crc kubenswrapper[4632]: I0313 11:55:00.044883 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:55:00 crc kubenswrapper[4632]: E0313 11:55:00.046770 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:55:13 crc kubenswrapper[4632]: I0313 11:55:13.044617 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:55:13 crc kubenswrapper[4632]: E0313 11:55:13.045588 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:55:25 crc kubenswrapper[4632]: I0313 11:55:25.044104 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:55:25 crc kubenswrapper[4632]: E0313 11:55:25.044740 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:55:39 crc kubenswrapper[4632]: I0313 11:55:39.044043 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:55:39 crc kubenswrapper[4632]: E0313 11:55:39.044859 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:55:51 crc kubenswrapper[4632]: I0313 11:55:51.044336 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:55:51 crc kubenswrapper[4632]: E0313 11:55:51.045101 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:56:00 crc kubenswrapper[4632]: I0313 11:56:00.145836 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556716-2gt9h"] Mar 13 11:56:00 crc kubenswrapper[4632]: E0313 11:56:00.146761 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88fb48fa-6650-4b22-b44f-d8c6f839489e" containerName="oc" Mar 13 11:56:00 crc kubenswrapper[4632]: I0313 11:56:00.146774 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="88fb48fa-6650-4b22-b44f-d8c6f839489e" containerName="oc" Mar 13 11:56:00 crc kubenswrapper[4632]: I0313 11:56:00.147011 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="88fb48fa-6650-4b22-b44f-d8c6f839489e" containerName="oc" Mar 13 11:56:00 crc kubenswrapper[4632]: I0313 11:56:00.147644 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556716-2gt9h" Mar 13 11:56:00 crc kubenswrapper[4632]: I0313 11:56:00.150526 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:56:00 crc kubenswrapper[4632]: I0313 11:56:00.150895 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:56:00 crc kubenswrapper[4632]: I0313 11:56:00.151448 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:56:00 crc kubenswrapper[4632]: I0313 11:56:00.167757 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556716-2gt9h"] Mar 13 11:56:00 crc kubenswrapper[4632]: I0313 11:56:00.297354 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp4vw\" (UniqueName: \"kubernetes.io/projected/38cafee7-6e61-46de-b58b-48b8f7d41bf6-kube-api-access-cp4vw\") pod \"auto-csr-approver-29556716-2gt9h\" (UID: \"38cafee7-6e61-46de-b58b-48b8f7d41bf6\") " pod="openshift-infra/auto-csr-approver-29556716-2gt9h" Mar 13 11:56:00 crc kubenswrapper[4632]: I0313 11:56:00.399373 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp4vw\" (UniqueName: \"kubernetes.io/projected/38cafee7-6e61-46de-b58b-48b8f7d41bf6-kube-api-access-cp4vw\") pod \"auto-csr-approver-29556716-2gt9h\" (UID: \"38cafee7-6e61-46de-b58b-48b8f7d41bf6\") " pod="openshift-infra/auto-csr-approver-29556716-2gt9h" Mar 13 11:56:00 crc kubenswrapper[4632]: I0313 11:56:00.423599 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp4vw\" (UniqueName: \"kubernetes.io/projected/38cafee7-6e61-46de-b58b-48b8f7d41bf6-kube-api-access-cp4vw\") pod \"auto-csr-approver-29556716-2gt9h\" (UID: \"38cafee7-6e61-46de-b58b-48b8f7d41bf6\") " pod="openshift-infra/auto-csr-approver-29556716-2gt9h" Mar 13 11:56:00 crc kubenswrapper[4632]: I0313 11:56:00.468705 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556716-2gt9h" Mar 13 11:56:00 crc kubenswrapper[4632]: I0313 11:56:00.986980 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556716-2gt9h"] Mar 13 11:56:00 crc kubenswrapper[4632]: I0313 11:56:00.987521 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 11:56:01 crc kubenswrapper[4632]: I0313 11:56:01.761303 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556716-2gt9h" event={"ID":"38cafee7-6e61-46de-b58b-48b8f7d41bf6","Type":"ContainerStarted","Data":"f3857abe0b359cb3a75944205d422467fa9c52bab1373b4c2e3b63d8fc99dfb7"} Mar 13 11:56:02 crc kubenswrapper[4632]: I0313 11:56:02.773904 4632 generic.go:334] "Generic (PLEG): container finished" podID="38cafee7-6e61-46de-b58b-48b8f7d41bf6" containerID="334a9f675c9c77aba9558302bf96e3547c17123adf9873e85b3c3871bccb4465" exitCode=0 Mar 13 11:56:02 crc kubenswrapper[4632]: I0313 11:56:02.774191 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556716-2gt9h" event={"ID":"38cafee7-6e61-46de-b58b-48b8f7d41bf6","Type":"ContainerDied","Data":"334a9f675c9c77aba9558302bf96e3547c17123adf9873e85b3c3871bccb4465"} Mar 13 11:56:03 crc kubenswrapper[4632]: I0313 11:56:03.044528 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:56:03 crc kubenswrapper[4632]: E0313 11:56:03.045029 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:56:04 crc kubenswrapper[4632]: I0313 11:56:04.190696 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556716-2gt9h" Mar 13 11:56:04 crc kubenswrapper[4632]: I0313 11:56:04.218730 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cp4vw\" (UniqueName: \"kubernetes.io/projected/38cafee7-6e61-46de-b58b-48b8f7d41bf6-kube-api-access-cp4vw\") pod \"38cafee7-6e61-46de-b58b-48b8f7d41bf6\" (UID: \"38cafee7-6e61-46de-b58b-48b8f7d41bf6\") " Mar 13 11:56:04 crc kubenswrapper[4632]: I0313 11:56:04.233447 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/38cafee7-6e61-46de-b58b-48b8f7d41bf6-kube-api-access-cp4vw" (OuterVolumeSpecName: "kube-api-access-cp4vw") pod "38cafee7-6e61-46de-b58b-48b8f7d41bf6" (UID: "38cafee7-6e61-46de-b58b-48b8f7d41bf6"). InnerVolumeSpecName "kube-api-access-cp4vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:56:04 crc kubenswrapper[4632]: I0313 11:56:04.320407 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cp4vw\" (UniqueName: \"kubernetes.io/projected/38cafee7-6e61-46de-b58b-48b8f7d41bf6-kube-api-access-cp4vw\") on node \"crc\" DevicePath \"\"" Mar 13 11:56:04 crc kubenswrapper[4632]: I0313 11:56:04.798179 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556716-2gt9h" event={"ID":"38cafee7-6e61-46de-b58b-48b8f7d41bf6","Type":"ContainerDied","Data":"f3857abe0b359cb3a75944205d422467fa9c52bab1373b4c2e3b63d8fc99dfb7"} Mar 13 11:56:04 crc kubenswrapper[4632]: I0313 11:56:04.798221 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556716-2gt9h" Mar 13 11:56:04 crc kubenswrapper[4632]: I0313 11:56:04.798247 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f3857abe0b359cb3a75944205d422467fa9c52bab1373b4c2e3b63d8fc99dfb7" Mar 13 11:56:05 crc kubenswrapper[4632]: I0313 11:56:05.273152 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556710-t4gtk"] Mar 13 11:56:05 crc kubenswrapper[4632]: I0313 11:56:05.282358 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556710-t4gtk"] Mar 13 11:56:06 crc kubenswrapper[4632]: I0313 11:56:06.055115 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca" path="/var/lib/kubelet/pods/38d73c58-f065-4efc-9fe2-b6c0ed9fa5ca/volumes" Mar 13 11:56:15 crc kubenswrapper[4632]: I0313 11:56:15.044338 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:56:15 crc kubenswrapper[4632]: E0313 11:56:15.045175 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:56:23 crc kubenswrapper[4632]: I0313 11:56:23.439967 4632 scope.go:117] "RemoveContainer" containerID="57e073e1e04617c49dfbb2c194d77f02cda77aac917eb626f73490bd0abacbcb" Mar 13 11:56:26 crc kubenswrapper[4632]: I0313 11:56:26.044705 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:56:26 crc kubenswrapper[4632]: E0313 11:56:26.045446 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 11:56:41 crc kubenswrapper[4632]: I0313 11:56:41.044273 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 11:56:42 crc kubenswrapper[4632]: I0313 11:56:42.133285 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"f1337fb64ab38c0f489a591d3b3f173d13428642427113f1891b2f17a626304e"} Mar 13 11:58:00 crc kubenswrapper[4632]: I0313 11:58:00.177079 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556718-xxtsr"] Mar 13 11:58:00 crc kubenswrapper[4632]: E0313 11:58:00.178183 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="38cafee7-6e61-46de-b58b-48b8f7d41bf6" containerName="oc" Mar 13 11:58:00 crc kubenswrapper[4632]: I0313 11:58:00.178199 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="38cafee7-6e61-46de-b58b-48b8f7d41bf6" containerName="oc" Mar 13 11:58:00 crc kubenswrapper[4632]: I0313 11:58:00.178403 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="38cafee7-6e61-46de-b58b-48b8f7d41bf6" containerName="oc" Mar 13 11:58:00 crc kubenswrapper[4632]: I0313 11:58:00.180134 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556718-xxtsr" Mar 13 11:58:00 crc kubenswrapper[4632]: I0313 11:58:00.184853 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 11:58:00 crc kubenswrapper[4632]: I0313 11:58:00.184860 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 11:58:00 crc kubenswrapper[4632]: I0313 11:58:00.187338 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 11:58:00 crc kubenswrapper[4632]: I0313 11:58:00.187557 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556718-xxtsr"] Mar 13 11:58:00 crc kubenswrapper[4632]: I0313 11:58:00.325490 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8275x\" (UniqueName: \"kubernetes.io/projected/beb533c7-a735-47fa-b5fa-67b1bcba9787-kube-api-access-8275x\") pod \"auto-csr-approver-29556718-xxtsr\" (UID: \"beb533c7-a735-47fa-b5fa-67b1bcba9787\") " pod="openshift-infra/auto-csr-approver-29556718-xxtsr" Mar 13 11:58:00 crc kubenswrapper[4632]: I0313 11:58:00.427696 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8275x\" (UniqueName: \"kubernetes.io/projected/beb533c7-a735-47fa-b5fa-67b1bcba9787-kube-api-access-8275x\") pod \"auto-csr-approver-29556718-xxtsr\" (UID: \"beb533c7-a735-47fa-b5fa-67b1bcba9787\") " pod="openshift-infra/auto-csr-approver-29556718-xxtsr" Mar 13 11:58:00 crc kubenswrapper[4632]: I0313 11:58:00.496778 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8275x\" (UniqueName: \"kubernetes.io/projected/beb533c7-a735-47fa-b5fa-67b1bcba9787-kube-api-access-8275x\") pod \"auto-csr-approver-29556718-xxtsr\" (UID: \"beb533c7-a735-47fa-b5fa-67b1bcba9787\") " pod="openshift-infra/auto-csr-approver-29556718-xxtsr" Mar 13 11:58:00 crc kubenswrapper[4632]: I0313 11:58:00.508843 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556718-xxtsr" Mar 13 11:58:02 crc kubenswrapper[4632]: I0313 11:58:02.170004 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556718-xxtsr"] Mar 13 11:58:02 crc kubenswrapper[4632]: I0313 11:58:02.882999 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556718-xxtsr" event={"ID":"beb533c7-a735-47fa-b5fa-67b1bcba9787","Type":"ContainerStarted","Data":"0d527ee16e85671e3e98d42a2ffcefa3d907752f70215b9be0fc57b705f6902c"} Mar 13 11:58:05 crc kubenswrapper[4632]: I0313 11:58:05.911803 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556718-xxtsr" event={"ID":"beb533c7-a735-47fa-b5fa-67b1bcba9787","Type":"ContainerStarted","Data":"de5a0b9383a1bdabde0e1290cb2d2e2341dbc3e19f3a7e552782ac9f0501a7ce"} Mar 13 11:58:05 crc kubenswrapper[4632]: I0313 11:58:05.931981 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556718-xxtsr" podStartSLOduration=5.137630648 podStartE2EDuration="5.931958709s" podCreationTimestamp="2026-03-13 11:58:00 +0000 UTC" firstStartedPulling="2026-03-13 11:58:02.204507146 +0000 UTC m=+6856.227037279" lastFinishedPulling="2026-03-13 11:58:02.998835207 +0000 UTC m=+6857.021365340" observedRunningTime="2026-03-13 11:58:05.924899244 +0000 UTC m=+6859.947429377" watchObservedRunningTime="2026-03-13 11:58:05.931958709 +0000 UTC m=+6859.954488852" Mar 13 11:58:09 crc kubenswrapper[4632]: I0313 11:58:09.989054 4632 generic.go:334] "Generic (PLEG): container finished" podID="beb533c7-a735-47fa-b5fa-67b1bcba9787" containerID="de5a0b9383a1bdabde0e1290cb2d2e2341dbc3e19f3a7e552782ac9f0501a7ce" exitCode=0 Mar 13 11:58:09 crc kubenswrapper[4632]: I0313 11:58:09.989117 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556718-xxtsr" event={"ID":"beb533c7-a735-47fa-b5fa-67b1bcba9787","Type":"ContainerDied","Data":"de5a0b9383a1bdabde0e1290cb2d2e2341dbc3e19f3a7e552782ac9f0501a7ce"} Mar 13 11:58:11 crc kubenswrapper[4632]: I0313 11:58:11.410041 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556718-xxtsr" Mar 13 11:58:11 crc kubenswrapper[4632]: I0313 11:58:11.574173 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8275x\" (UniqueName: \"kubernetes.io/projected/beb533c7-a735-47fa-b5fa-67b1bcba9787-kube-api-access-8275x\") pod \"beb533c7-a735-47fa-b5fa-67b1bcba9787\" (UID: \"beb533c7-a735-47fa-b5fa-67b1bcba9787\") " Mar 13 11:58:11 crc kubenswrapper[4632]: I0313 11:58:11.586326 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/beb533c7-a735-47fa-b5fa-67b1bcba9787-kube-api-access-8275x" (OuterVolumeSpecName: "kube-api-access-8275x") pod "beb533c7-a735-47fa-b5fa-67b1bcba9787" (UID: "beb533c7-a735-47fa-b5fa-67b1bcba9787"). InnerVolumeSpecName "kube-api-access-8275x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:58:11 crc kubenswrapper[4632]: I0313 11:58:11.676198 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8275x\" (UniqueName: \"kubernetes.io/projected/beb533c7-a735-47fa-b5fa-67b1bcba9787-kube-api-access-8275x\") on node \"crc\" DevicePath \"\"" Mar 13 11:58:12 crc kubenswrapper[4632]: I0313 11:58:12.010296 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556718-xxtsr" event={"ID":"beb533c7-a735-47fa-b5fa-67b1bcba9787","Type":"ContainerDied","Data":"0d527ee16e85671e3e98d42a2ffcefa3d907752f70215b9be0fc57b705f6902c"} Mar 13 11:58:12 crc kubenswrapper[4632]: I0313 11:58:12.010336 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d527ee16e85671e3e98d42a2ffcefa3d907752f70215b9be0fc57b705f6902c" Mar 13 11:58:12 crc kubenswrapper[4632]: I0313 11:58:12.010407 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556718-xxtsr" Mar 13 11:58:12 crc kubenswrapper[4632]: I0313 11:58:12.097586 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556712-jjq26"] Mar 13 11:58:12 crc kubenswrapper[4632]: I0313 11:58:12.109011 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556712-jjq26"] Mar 13 11:58:14 crc kubenswrapper[4632]: I0313 11:58:14.059342 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02db6f2d-ef7c-4444-9776-603e9c44c55a" path="/var/lib/kubelet/pods/02db6f2d-ef7c-4444-9776-603e9c44c55a/volumes" Mar 13 11:58:23 crc kubenswrapper[4632]: I0313 11:58:23.843574 4632 scope.go:117] "RemoveContainer" containerID="1b3640b2bd3a5d0dabd874439006e72fb30cf0784909546f4e8957109f6ffcf0" Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.532305 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-drgch"] Mar 13 11:58:37 crc kubenswrapper[4632]: E0313 11:58:37.534403 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="beb533c7-a735-47fa-b5fa-67b1bcba9787" containerName="oc" Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.534497 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="beb533c7-a735-47fa-b5fa-67b1bcba9787" containerName="oc" Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.534799 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="beb533c7-a735-47fa-b5fa-67b1bcba9787" containerName="oc" Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.537686 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.553761 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-drgch"] Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.691547 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nr58\" (UniqueName: \"kubernetes.io/projected/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-kube-api-access-7nr58\") pod \"redhat-marketplace-drgch\" (UID: \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\") " pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.691972 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-catalog-content\") pod \"redhat-marketplace-drgch\" (UID: \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\") " pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.692098 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-utilities\") pod \"redhat-marketplace-drgch\" (UID: \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\") " pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.794267 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nr58\" (UniqueName: \"kubernetes.io/projected/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-kube-api-access-7nr58\") pod \"redhat-marketplace-drgch\" (UID: \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\") " pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.794346 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-catalog-content\") pod \"redhat-marketplace-drgch\" (UID: \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\") " pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.794374 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-utilities\") pod \"redhat-marketplace-drgch\" (UID: \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\") " pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.795103 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-utilities\") pod \"redhat-marketplace-drgch\" (UID: \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\") " pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.795371 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-catalog-content\") pod \"redhat-marketplace-drgch\" (UID: \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\") " pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.829468 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nr58\" (UniqueName: \"kubernetes.io/projected/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-kube-api-access-7nr58\") pod \"redhat-marketplace-drgch\" (UID: \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\") " pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:37 crc kubenswrapper[4632]: I0313 11:58:37.857618 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:38 crc kubenswrapper[4632]: I0313 11:58:38.366816 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-drgch"] Mar 13 11:58:39 crc kubenswrapper[4632]: I0313 11:58:39.281215 4632 generic.go:334] "Generic (PLEG): container finished" podID="6cd708bb-8ca4-4c04-95f1-dfb27bee6832" containerID="0785a5b501b04e1d4e9bc948bd7a7407375731736568cf4ddf2b272fd73d83f6" exitCode=0 Mar 13 11:58:39 crc kubenswrapper[4632]: I0313 11:58:39.281263 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drgch" event={"ID":"6cd708bb-8ca4-4c04-95f1-dfb27bee6832","Type":"ContainerDied","Data":"0785a5b501b04e1d4e9bc948bd7a7407375731736568cf4ddf2b272fd73d83f6"} Mar 13 11:58:39 crc kubenswrapper[4632]: I0313 11:58:39.281442 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drgch" event={"ID":"6cd708bb-8ca4-4c04-95f1-dfb27bee6832","Type":"ContainerStarted","Data":"9679a196913903ccdb588bb8ba90e2da048eff47535c211d5b2101f6e140b063"} Mar 13 11:58:41 crc kubenswrapper[4632]: I0313 11:58:41.301304 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drgch" event={"ID":"6cd708bb-8ca4-4c04-95f1-dfb27bee6832","Type":"ContainerStarted","Data":"5efb7453218a8d0b7abc9c2463a8936543c07c4d4dc2ed95212fd8a5cd84d6d3"} Mar 13 11:58:42 crc kubenswrapper[4632]: I0313 11:58:42.317642 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drgch" event={"ID":"6cd708bb-8ca4-4c04-95f1-dfb27bee6832","Type":"ContainerDied","Data":"5efb7453218a8d0b7abc9c2463a8936543c07c4d4dc2ed95212fd8a5cd84d6d3"} Mar 13 11:58:42 crc kubenswrapper[4632]: I0313 11:58:42.317699 4632 generic.go:334] "Generic (PLEG): container finished" podID="6cd708bb-8ca4-4c04-95f1-dfb27bee6832" containerID="5efb7453218a8d0b7abc9c2463a8936543c07c4d4dc2ed95212fd8a5cd84d6d3" exitCode=0 Mar 13 11:58:43 crc kubenswrapper[4632]: I0313 11:58:43.329111 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drgch" event={"ID":"6cd708bb-8ca4-4c04-95f1-dfb27bee6832","Type":"ContainerStarted","Data":"8c3253a43857fd1fd4bdd5757007dffd53fbea2cdd3ce32c464af84fa8752306"} Mar 13 11:58:43 crc kubenswrapper[4632]: I0313 11:58:43.352695 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-drgch" podStartSLOduration=2.919448954 podStartE2EDuration="6.352674241s" podCreationTimestamp="2026-03-13 11:58:37 +0000 UTC" firstStartedPulling="2026-03-13 11:58:39.284482431 +0000 UTC m=+6893.307012564" lastFinishedPulling="2026-03-13 11:58:42.717707718 +0000 UTC m=+6896.740237851" observedRunningTime="2026-03-13 11:58:43.346016768 +0000 UTC m=+6897.368546921" watchObservedRunningTime="2026-03-13 11:58:43.352674241 +0000 UTC m=+6897.375204394" Mar 13 11:58:47 crc kubenswrapper[4632]: I0313 11:58:47.858594 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:47 crc kubenswrapper[4632]: I0313 11:58:47.859192 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:48 crc kubenswrapper[4632]: I0313 11:58:48.912501 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-drgch" podUID="6cd708bb-8ca4-4c04-95f1-dfb27bee6832" containerName="registry-server" probeResult="failure" output=< Mar 13 11:58:48 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:58:48 crc kubenswrapper[4632]: > Mar 13 11:58:57 crc kubenswrapper[4632]: I0313 11:58:57.909998 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:57 crc kubenswrapper[4632]: I0313 11:58:57.973266 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:58:58 crc kubenswrapper[4632]: I0313 11:58:58.157706 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-drgch"] Mar 13 11:58:59 crc kubenswrapper[4632]: I0313 11:58:59.534831 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-drgch" podUID="6cd708bb-8ca4-4c04-95f1-dfb27bee6832" containerName="registry-server" containerID="cri-o://8c3253a43857fd1fd4bdd5757007dffd53fbea2cdd3ce32c464af84fa8752306" gracePeriod=2 Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.068533 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.181074 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-utilities\") pod \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\" (UID: \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\") " Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.181167 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nr58\" (UniqueName: \"kubernetes.io/projected/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-kube-api-access-7nr58\") pod \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\" (UID: \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\") " Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.181277 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-catalog-content\") pod \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\" (UID: \"6cd708bb-8ca4-4c04-95f1-dfb27bee6832\") " Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.182358 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-utilities" (OuterVolumeSpecName: "utilities") pod "6cd708bb-8ca4-4c04-95f1-dfb27bee6832" (UID: "6cd708bb-8ca4-4c04-95f1-dfb27bee6832"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.183454 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.190660 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-kube-api-access-7nr58" (OuterVolumeSpecName: "kube-api-access-7nr58") pod "6cd708bb-8ca4-4c04-95f1-dfb27bee6832" (UID: "6cd708bb-8ca4-4c04-95f1-dfb27bee6832"). InnerVolumeSpecName "kube-api-access-7nr58". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.213086 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6cd708bb-8ca4-4c04-95f1-dfb27bee6832" (UID: "6cd708bb-8ca4-4c04-95f1-dfb27bee6832"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.285331 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nr58\" (UniqueName: \"kubernetes.io/projected/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-kube-api-access-7nr58\") on node \"crc\" DevicePath \"\"" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.285409 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6cd708bb-8ca4-4c04-95f1-dfb27bee6832-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.545617 4632 generic.go:334] "Generic (PLEG): container finished" podID="6cd708bb-8ca4-4c04-95f1-dfb27bee6832" containerID="8c3253a43857fd1fd4bdd5757007dffd53fbea2cdd3ce32c464af84fa8752306" exitCode=0 Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.545668 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drgch" event={"ID":"6cd708bb-8ca4-4c04-95f1-dfb27bee6832","Type":"ContainerDied","Data":"8c3253a43857fd1fd4bdd5757007dffd53fbea2cdd3ce32c464af84fa8752306"} Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.545698 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-drgch" event={"ID":"6cd708bb-8ca4-4c04-95f1-dfb27bee6832","Type":"ContainerDied","Data":"9679a196913903ccdb588bb8ba90e2da048eff47535c211d5b2101f6e140b063"} Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.545720 4632 scope.go:117] "RemoveContainer" containerID="8c3253a43857fd1fd4bdd5757007dffd53fbea2cdd3ce32c464af84fa8752306" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.545871 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-drgch" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.588295 4632 scope.go:117] "RemoveContainer" containerID="5efb7453218a8d0b7abc9c2463a8936543c07c4d4dc2ed95212fd8a5cd84d6d3" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.590744 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-drgch"] Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.608299 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-drgch"] Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.614514 4632 scope.go:117] "RemoveContainer" containerID="0785a5b501b04e1d4e9bc948bd7a7407375731736568cf4ddf2b272fd73d83f6" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.668467 4632 scope.go:117] "RemoveContainer" containerID="8c3253a43857fd1fd4bdd5757007dffd53fbea2cdd3ce32c464af84fa8752306" Mar 13 11:59:00 crc kubenswrapper[4632]: E0313 11:59:00.670025 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c3253a43857fd1fd4bdd5757007dffd53fbea2cdd3ce32c464af84fa8752306\": container with ID starting with 8c3253a43857fd1fd4bdd5757007dffd53fbea2cdd3ce32c464af84fa8752306 not found: ID does not exist" containerID="8c3253a43857fd1fd4bdd5757007dffd53fbea2cdd3ce32c464af84fa8752306" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.670062 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c3253a43857fd1fd4bdd5757007dffd53fbea2cdd3ce32c464af84fa8752306"} err="failed to get container status \"8c3253a43857fd1fd4bdd5757007dffd53fbea2cdd3ce32c464af84fa8752306\": rpc error: code = NotFound desc = could not find container \"8c3253a43857fd1fd4bdd5757007dffd53fbea2cdd3ce32c464af84fa8752306\": container with ID starting with 8c3253a43857fd1fd4bdd5757007dffd53fbea2cdd3ce32c464af84fa8752306 not found: ID does not exist" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.670087 4632 scope.go:117] "RemoveContainer" containerID="5efb7453218a8d0b7abc9c2463a8936543c07c4d4dc2ed95212fd8a5cd84d6d3" Mar 13 11:59:00 crc kubenswrapper[4632]: E0313 11:59:00.670576 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5efb7453218a8d0b7abc9c2463a8936543c07c4d4dc2ed95212fd8a5cd84d6d3\": container with ID starting with 5efb7453218a8d0b7abc9c2463a8936543c07c4d4dc2ed95212fd8a5cd84d6d3 not found: ID does not exist" containerID="5efb7453218a8d0b7abc9c2463a8936543c07c4d4dc2ed95212fd8a5cd84d6d3" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.670601 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5efb7453218a8d0b7abc9c2463a8936543c07c4d4dc2ed95212fd8a5cd84d6d3"} err="failed to get container status \"5efb7453218a8d0b7abc9c2463a8936543c07c4d4dc2ed95212fd8a5cd84d6d3\": rpc error: code = NotFound desc = could not find container \"5efb7453218a8d0b7abc9c2463a8936543c07c4d4dc2ed95212fd8a5cd84d6d3\": container with ID starting with 5efb7453218a8d0b7abc9c2463a8936543c07c4d4dc2ed95212fd8a5cd84d6d3 not found: ID does not exist" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.670616 4632 scope.go:117] "RemoveContainer" containerID="0785a5b501b04e1d4e9bc948bd7a7407375731736568cf4ddf2b272fd73d83f6" Mar 13 11:59:00 crc kubenswrapper[4632]: E0313 11:59:00.672187 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0785a5b501b04e1d4e9bc948bd7a7407375731736568cf4ddf2b272fd73d83f6\": container with ID starting with 0785a5b501b04e1d4e9bc948bd7a7407375731736568cf4ddf2b272fd73d83f6 not found: ID does not exist" containerID="0785a5b501b04e1d4e9bc948bd7a7407375731736568cf4ddf2b272fd73d83f6" Mar 13 11:59:00 crc kubenswrapper[4632]: I0313 11:59:00.672216 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0785a5b501b04e1d4e9bc948bd7a7407375731736568cf4ddf2b272fd73d83f6"} err="failed to get container status \"0785a5b501b04e1d4e9bc948bd7a7407375731736568cf4ddf2b272fd73d83f6\": rpc error: code = NotFound desc = could not find container \"0785a5b501b04e1d4e9bc948bd7a7407375731736568cf4ddf2b272fd73d83f6\": container with ID starting with 0785a5b501b04e1d4e9bc948bd7a7407375731736568cf4ddf2b272fd73d83f6 not found: ID does not exist" Mar 13 11:59:02 crc kubenswrapper[4632]: I0313 11:59:02.067392 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cd708bb-8ca4-4c04-95f1-dfb27bee6832" path="/var/lib/kubelet/pods/6cd708bb-8ca4-4c04-95f1-dfb27bee6832/volumes" Mar 13 11:59:10 crc kubenswrapper[4632]: I0313 11:59:10.461425 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:59:10 crc kubenswrapper[4632]: I0313 11:59:10.461859 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.665099 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-wsx2g"] Mar 13 11:59:16 crc kubenswrapper[4632]: E0313 11:59:16.666290 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd708bb-8ca4-4c04-95f1-dfb27bee6832" containerName="registry-server" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.666314 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd708bb-8ca4-4c04-95f1-dfb27bee6832" containerName="registry-server" Mar 13 11:59:16 crc kubenswrapper[4632]: E0313 11:59:16.666357 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd708bb-8ca4-4c04-95f1-dfb27bee6832" containerName="extract-content" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.666367 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd708bb-8ca4-4c04-95f1-dfb27bee6832" containerName="extract-content" Mar 13 11:59:16 crc kubenswrapper[4632]: E0313 11:59:16.666384 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cd708bb-8ca4-4c04-95f1-dfb27bee6832" containerName="extract-utilities" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.666395 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cd708bb-8ca4-4c04-95f1-dfb27bee6832" containerName="extract-utilities" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.666656 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cd708bb-8ca4-4c04-95f1-dfb27bee6832" containerName="registry-server" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.668569 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.707043 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wsx2g"] Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.832425 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shzdx\" (UniqueName: \"kubernetes.io/projected/91e35ce1-c54e-46e0-aa1f-01eed2268826-kube-api-access-shzdx\") pod \"community-operators-wsx2g\" (UID: \"91e35ce1-c54e-46e0-aa1f-01eed2268826\") " pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.832738 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e35ce1-c54e-46e0-aa1f-01eed2268826-catalog-content\") pod \"community-operators-wsx2g\" (UID: \"91e35ce1-c54e-46e0-aa1f-01eed2268826\") " pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.833004 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e35ce1-c54e-46e0-aa1f-01eed2268826-utilities\") pod \"community-operators-wsx2g\" (UID: \"91e35ce1-c54e-46e0-aa1f-01eed2268826\") " pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.935181 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e35ce1-c54e-46e0-aa1f-01eed2268826-utilities\") pod \"community-operators-wsx2g\" (UID: \"91e35ce1-c54e-46e0-aa1f-01eed2268826\") " pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.935254 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shzdx\" (UniqueName: \"kubernetes.io/projected/91e35ce1-c54e-46e0-aa1f-01eed2268826-kube-api-access-shzdx\") pod \"community-operators-wsx2g\" (UID: \"91e35ce1-c54e-46e0-aa1f-01eed2268826\") " pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.935293 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e35ce1-c54e-46e0-aa1f-01eed2268826-catalog-content\") pod \"community-operators-wsx2g\" (UID: \"91e35ce1-c54e-46e0-aa1f-01eed2268826\") " pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.935863 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e35ce1-c54e-46e0-aa1f-01eed2268826-catalog-content\") pod \"community-operators-wsx2g\" (UID: \"91e35ce1-c54e-46e0-aa1f-01eed2268826\") " pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.936248 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e35ce1-c54e-46e0-aa1f-01eed2268826-utilities\") pod \"community-operators-wsx2g\" (UID: \"91e35ce1-c54e-46e0-aa1f-01eed2268826\") " pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.960831 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shzdx\" (UniqueName: \"kubernetes.io/projected/91e35ce1-c54e-46e0-aa1f-01eed2268826-kube-api-access-shzdx\") pod \"community-operators-wsx2g\" (UID: \"91e35ce1-c54e-46e0-aa1f-01eed2268826\") " pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:16 crc kubenswrapper[4632]: I0313 11:59:16.995646 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:17 crc kubenswrapper[4632]: W0313 11:59:17.562626 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91e35ce1_c54e_46e0_aa1f_01eed2268826.slice/crio-64bdcb0f99dfc86dc26ea146e91dc9717e2b551a594e5de0e4aa29b35e623e4d WatchSource:0}: Error finding container 64bdcb0f99dfc86dc26ea146e91dc9717e2b551a594e5de0e4aa29b35e623e4d: Status 404 returned error can't find the container with id 64bdcb0f99dfc86dc26ea146e91dc9717e2b551a594e5de0e4aa29b35e623e4d Mar 13 11:59:17 crc kubenswrapper[4632]: I0313 11:59:17.565933 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-wsx2g"] Mar 13 11:59:17 crc kubenswrapper[4632]: I0313 11:59:17.705484 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wsx2g" event={"ID":"91e35ce1-c54e-46e0-aa1f-01eed2268826","Type":"ContainerStarted","Data":"64bdcb0f99dfc86dc26ea146e91dc9717e2b551a594e5de0e4aa29b35e623e4d"} Mar 13 11:59:18 crc kubenswrapper[4632]: I0313 11:59:18.716311 4632 generic.go:334] "Generic (PLEG): container finished" podID="91e35ce1-c54e-46e0-aa1f-01eed2268826" containerID="fd80794757f43c71fc7435b850cacfdc1cc26db2bff61d8d3f5c45e9fc5f1169" exitCode=0 Mar 13 11:59:18 crc kubenswrapper[4632]: I0313 11:59:18.716421 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wsx2g" event={"ID":"91e35ce1-c54e-46e0-aa1f-01eed2268826","Type":"ContainerDied","Data":"fd80794757f43c71fc7435b850cacfdc1cc26db2bff61d8d3f5c45e9fc5f1169"} Mar 13 11:59:20 crc kubenswrapper[4632]: I0313 11:59:20.734987 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wsx2g" event={"ID":"91e35ce1-c54e-46e0-aa1f-01eed2268826","Type":"ContainerStarted","Data":"5e4654df138739a880c0cf4bc5f6fd259cb10ae8d73f8303664c6a10dbea639a"} Mar 13 11:59:23 crc kubenswrapper[4632]: I0313 11:59:23.767554 4632 generic.go:334] "Generic (PLEG): container finished" podID="91e35ce1-c54e-46e0-aa1f-01eed2268826" containerID="5e4654df138739a880c0cf4bc5f6fd259cb10ae8d73f8303664c6a10dbea639a" exitCode=0 Mar 13 11:59:23 crc kubenswrapper[4632]: I0313 11:59:23.767762 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wsx2g" event={"ID":"91e35ce1-c54e-46e0-aa1f-01eed2268826","Type":"ContainerDied","Data":"5e4654df138739a880c0cf4bc5f6fd259cb10ae8d73f8303664c6a10dbea639a"} Mar 13 11:59:24 crc kubenswrapper[4632]: I0313 11:59:24.779337 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wsx2g" event={"ID":"91e35ce1-c54e-46e0-aa1f-01eed2268826","Type":"ContainerStarted","Data":"ac32c8bb9b1ef33627f10d2b87bef22688c20d954eb247669d8659c5d11a8b81"} Mar 13 11:59:24 crc kubenswrapper[4632]: I0313 11:59:24.800622 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-wsx2g" podStartSLOduration=3.11447268 podStartE2EDuration="8.800601172s" podCreationTimestamp="2026-03-13 11:59:16 +0000 UTC" firstStartedPulling="2026-03-13 11:59:18.718513864 +0000 UTC m=+6932.741043997" lastFinishedPulling="2026-03-13 11:59:24.404642356 +0000 UTC m=+6938.427172489" observedRunningTime="2026-03-13 11:59:24.796342968 +0000 UTC m=+6938.818873101" watchObservedRunningTime="2026-03-13 11:59:24.800601172 +0000 UTC m=+6938.823131305" Mar 13 11:59:26 crc kubenswrapper[4632]: I0313 11:59:26.996380 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:26 crc kubenswrapper[4632]: I0313 11:59:26.996713 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:28 crc kubenswrapper[4632]: I0313 11:59:28.053229 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-wsx2g" podUID="91e35ce1-c54e-46e0-aa1f-01eed2268826" containerName="registry-server" probeResult="failure" output=< Mar 13 11:59:28 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:59:28 crc kubenswrapper[4632]: > Mar 13 11:59:38 crc kubenswrapper[4632]: I0313 11:59:38.050664 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-wsx2g" podUID="91e35ce1-c54e-46e0-aa1f-01eed2268826" containerName="registry-server" probeResult="failure" output=< Mar 13 11:59:38 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:59:38 crc kubenswrapper[4632]: > Mar 13 11:59:40 crc kubenswrapper[4632]: I0313 11:59:40.461390 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 11:59:40 crc kubenswrapper[4632]: I0313 11:59:40.461718 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 11:59:48 crc kubenswrapper[4632]: I0313 11:59:48.053458 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-wsx2g" podUID="91e35ce1-c54e-46e0-aa1f-01eed2268826" containerName="registry-server" probeResult="failure" output=< Mar 13 11:59:48 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 11:59:48 crc kubenswrapper[4632]: > Mar 13 11:59:57 crc kubenswrapper[4632]: I0313 11:59:57.048509 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:57 crc kubenswrapper[4632]: I0313 11:59:57.117367 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:57 crc kubenswrapper[4632]: I0313 11:59:57.302127 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wsx2g"] Mar 13 11:59:58 crc kubenswrapper[4632]: I0313 11:59:58.095478 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-wsx2g" podUID="91e35ce1-c54e-46e0-aa1f-01eed2268826" containerName="registry-server" containerID="cri-o://ac32c8bb9b1ef33627f10d2b87bef22688c20d954eb247669d8659c5d11a8b81" gracePeriod=2 Mar 13 11:59:58 crc kubenswrapper[4632]: I0313 11:59:58.585488 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:58 crc kubenswrapper[4632]: I0313 11:59:58.738904 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shzdx\" (UniqueName: \"kubernetes.io/projected/91e35ce1-c54e-46e0-aa1f-01eed2268826-kube-api-access-shzdx\") pod \"91e35ce1-c54e-46e0-aa1f-01eed2268826\" (UID: \"91e35ce1-c54e-46e0-aa1f-01eed2268826\") " Mar 13 11:59:58 crc kubenswrapper[4632]: I0313 11:59:58.739029 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e35ce1-c54e-46e0-aa1f-01eed2268826-catalog-content\") pod \"91e35ce1-c54e-46e0-aa1f-01eed2268826\" (UID: \"91e35ce1-c54e-46e0-aa1f-01eed2268826\") " Mar 13 11:59:58 crc kubenswrapper[4632]: I0313 11:59:58.739178 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e35ce1-c54e-46e0-aa1f-01eed2268826-utilities\") pod \"91e35ce1-c54e-46e0-aa1f-01eed2268826\" (UID: \"91e35ce1-c54e-46e0-aa1f-01eed2268826\") " Mar 13 11:59:58 crc kubenswrapper[4632]: I0313 11:59:58.740221 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91e35ce1-c54e-46e0-aa1f-01eed2268826-utilities" (OuterVolumeSpecName: "utilities") pod "91e35ce1-c54e-46e0-aa1f-01eed2268826" (UID: "91e35ce1-c54e-46e0-aa1f-01eed2268826"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:59:58 crc kubenswrapper[4632]: I0313 11:59:58.748247 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/91e35ce1-c54e-46e0-aa1f-01eed2268826-kube-api-access-shzdx" (OuterVolumeSpecName: "kube-api-access-shzdx") pod "91e35ce1-c54e-46e0-aa1f-01eed2268826" (UID: "91e35ce1-c54e-46e0-aa1f-01eed2268826"). InnerVolumeSpecName "kube-api-access-shzdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 11:59:58 crc kubenswrapper[4632]: I0313 11:59:58.801934 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/91e35ce1-c54e-46e0-aa1f-01eed2268826-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "91e35ce1-c54e-46e0-aa1f-01eed2268826" (UID: "91e35ce1-c54e-46e0-aa1f-01eed2268826"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 11:59:58 crc kubenswrapper[4632]: I0313 11:59:58.843518 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/91e35ce1-c54e-46e0-aa1f-01eed2268826-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 11:59:58 crc kubenswrapper[4632]: I0313 11:59:58.844121 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shzdx\" (UniqueName: \"kubernetes.io/projected/91e35ce1-c54e-46e0-aa1f-01eed2268826-kube-api-access-shzdx\") on node \"crc\" DevicePath \"\"" Mar 13 11:59:58 crc kubenswrapper[4632]: I0313 11:59:58.844142 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/91e35ce1-c54e-46e0-aa1f-01eed2268826-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.106346 4632 generic.go:334] "Generic (PLEG): container finished" podID="91e35ce1-c54e-46e0-aa1f-01eed2268826" containerID="ac32c8bb9b1ef33627f10d2b87bef22688c20d954eb247669d8659c5d11a8b81" exitCode=0 Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.106402 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wsx2g" event={"ID":"91e35ce1-c54e-46e0-aa1f-01eed2268826","Type":"ContainerDied","Data":"ac32c8bb9b1ef33627f10d2b87bef22688c20d954eb247669d8659c5d11a8b81"} Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.106436 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-wsx2g" event={"ID":"91e35ce1-c54e-46e0-aa1f-01eed2268826","Type":"ContainerDied","Data":"64bdcb0f99dfc86dc26ea146e91dc9717e2b551a594e5de0e4aa29b35e623e4d"} Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.106459 4632 scope.go:117] "RemoveContainer" containerID="ac32c8bb9b1ef33627f10d2b87bef22688c20d954eb247669d8659c5d11a8b81" Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.106634 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-wsx2g" Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.138542 4632 scope.go:117] "RemoveContainer" containerID="5e4654df138739a880c0cf4bc5f6fd259cb10ae8d73f8303664c6a10dbea639a" Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.163300 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-wsx2g"] Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.167030 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-wsx2g"] Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.175569 4632 scope.go:117] "RemoveContainer" containerID="fd80794757f43c71fc7435b850cacfdc1cc26db2bff61d8d3f5c45e9fc5f1169" Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.212752 4632 scope.go:117] "RemoveContainer" containerID="ac32c8bb9b1ef33627f10d2b87bef22688c20d954eb247669d8659c5d11a8b81" Mar 13 11:59:59 crc kubenswrapper[4632]: E0313 11:59:59.213349 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac32c8bb9b1ef33627f10d2b87bef22688c20d954eb247669d8659c5d11a8b81\": container with ID starting with ac32c8bb9b1ef33627f10d2b87bef22688c20d954eb247669d8659c5d11a8b81 not found: ID does not exist" containerID="ac32c8bb9b1ef33627f10d2b87bef22688c20d954eb247669d8659c5d11a8b81" Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.213429 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac32c8bb9b1ef33627f10d2b87bef22688c20d954eb247669d8659c5d11a8b81"} err="failed to get container status \"ac32c8bb9b1ef33627f10d2b87bef22688c20d954eb247669d8659c5d11a8b81\": rpc error: code = NotFound desc = could not find container \"ac32c8bb9b1ef33627f10d2b87bef22688c20d954eb247669d8659c5d11a8b81\": container with ID starting with ac32c8bb9b1ef33627f10d2b87bef22688c20d954eb247669d8659c5d11a8b81 not found: ID does not exist" Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.213462 4632 scope.go:117] "RemoveContainer" containerID="5e4654df138739a880c0cf4bc5f6fd259cb10ae8d73f8303664c6a10dbea639a" Mar 13 11:59:59 crc kubenswrapper[4632]: E0313 11:59:59.213921 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e4654df138739a880c0cf4bc5f6fd259cb10ae8d73f8303664c6a10dbea639a\": container with ID starting with 5e4654df138739a880c0cf4bc5f6fd259cb10ae8d73f8303664c6a10dbea639a not found: ID does not exist" containerID="5e4654df138739a880c0cf4bc5f6fd259cb10ae8d73f8303664c6a10dbea639a" Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.213976 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e4654df138739a880c0cf4bc5f6fd259cb10ae8d73f8303664c6a10dbea639a"} err="failed to get container status \"5e4654df138739a880c0cf4bc5f6fd259cb10ae8d73f8303664c6a10dbea639a\": rpc error: code = NotFound desc = could not find container \"5e4654df138739a880c0cf4bc5f6fd259cb10ae8d73f8303664c6a10dbea639a\": container with ID starting with 5e4654df138739a880c0cf4bc5f6fd259cb10ae8d73f8303664c6a10dbea639a not found: ID does not exist" Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.214004 4632 scope.go:117] "RemoveContainer" containerID="fd80794757f43c71fc7435b850cacfdc1cc26db2bff61d8d3f5c45e9fc5f1169" Mar 13 11:59:59 crc kubenswrapper[4632]: E0313 11:59:59.214352 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd80794757f43c71fc7435b850cacfdc1cc26db2bff61d8d3f5c45e9fc5f1169\": container with ID starting with fd80794757f43c71fc7435b850cacfdc1cc26db2bff61d8d3f5c45e9fc5f1169 not found: ID does not exist" containerID="fd80794757f43c71fc7435b850cacfdc1cc26db2bff61d8d3f5c45e9fc5f1169" Mar 13 11:59:59 crc kubenswrapper[4632]: I0313 11:59:59.214438 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd80794757f43c71fc7435b850cacfdc1cc26db2bff61d8d3f5c45e9fc5f1169"} err="failed to get container status \"fd80794757f43c71fc7435b850cacfdc1cc26db2bff61d8d3f5c45e9fc5f1169\": rpc error: code = NotFound desc = could not find container \"fd80794757f43c71fc7435b850cacfdc1cc26db2bff61d8d3f5c45e9fc5f1169\": container with ID starting with fd80794757f43c71fc7435b850cacfdc1cc26db2bff61d8d3f5c45e9fc5f1169 not found: ID does not exist" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.056525 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="91e35ce1-c54e-46e0-aa1f-01eed2268826" path="/var/lib/kubelet/pods/91e35ce1-c54e-46e0-aa1f-01eed2268826/volumes" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.167380 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556720-2985g"] Mar 13 12:00:00 crc kubenswrapper[4632]: E0313 12:00:00.167932 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e35ce1-c54e-46e0-aa1f-01eed2268826" containerName="extract-utilities" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.167977 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e35ce1-c54e-46e0-aa1f-01eed2268826" containerName="extract-utilities" Mar 13 12:00:00 crc kubenswrapper[4632]: E0313 12:00:00.168010 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e35ce1-c54e-46e0-aa1f-01eed2268826" containerName="registry-server" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.168022 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e35ce1-c54e-46e0-aa1f-01eed2268826" containerName="registry-server" Mar 13 12:00:00 crc kubenswrapper[4632]: E0313 12:00:00.168059 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="91e35ce1-c54e-46e0-aa1f-01eed2268826" containerName="extract-content" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.168068 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="91e35ce1-c54e-46e0-aa1f-01eed2268826" containerName="extract-content" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.170144 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="91e35ce1-c54e-46e0-aa1f-01eed2268826" containerName="registry-server" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.171241 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556720-2985g" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.185406 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556720-2985g"] Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.186366 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.187999 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.188093 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.271502 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7"] Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.273569 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.276188 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.276387 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.279780 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z5vxf\" (UniqueName: \"kubernetes.io/projected/614f8dd7-8a57-4b22-b741-63c3ed563216-kube-api-access-z5vxf\") pod \"auto-csr-approver-29556720-2985g\" (UID: \"614f8dd7-8a57-4b22-b741-63c3ed563216\") " pod="openshift-infra/auto-csr-approver-29556720-2985g" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.295111 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7"] Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.382688 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-secret-volume\") pod \"collect-profiles-29556720-9qfd7\" (UID: \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.382972 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z5vxf\" (UniqueName: \"kubernetes.io/projected/614f8dd7-8a57-4b22-b741-63c3ed563216-kube-api-access-z5vxf\") pod \"auto-csr-approver-29556720-2985g\" (UID: \"614f8dd7-8a57-4b22-b741-63c3ed563216\") " pod="openshift-infra/auto-csr-approver-29556720-2985g" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.383072 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-config-volume\") pod \"collect-profiles-29556720-9qfd7\" (UID: \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.383219 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmzs7\" (UniqueName: \"kubernetes.io/projected/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-kube-api-access-jmzs7\") pod \"collect-profiles-29556720-9qfd7\" (UID: \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.437364 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z5vxf\" (UniqueName: \"kubernetes.io/projected/614f8dd7-8a57-4b22-b741-63c3ed563216-kube-api-access-z5vxf\") pod \"auto-csr-approver-29556720-2985g\" (UID: \"614f8dd7-8a57-4b22-b741-63c3ed563216\") " pod="openshift-infra/auto-csr-approver-29556720-2985g" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.486816 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-secret-volume\") pod \"collect-profiles-29556720-9qfd7\" (UID: \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.487051 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-config-volume\") pod \"collect-profiles-29556720-9qfd7\" (UID: \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.487210 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmzs7\" (UniqueName: \"kubernetes.io/projected/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-kube-api-access-jmzs7\") pod \"collect-profiles-29556720-9qfd7\" (UID: \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.490133 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-config-volume\") pod \"collect-profiles-29556720-9qfd7\" (UID: \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.490196 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556720-2985g" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.502222 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-secret-volume\") pod \"collect-profiles-29556720-9qfd7\" (UID: \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.533046 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmzs7\" (UniqueName: \"kubernetes.io/projected/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-kube-api-access-jmzs7\") pod \"collect-profiles-29556720-9qfd7\" (UID: \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" Mar 13 12:00:00 crc kubenswrapper[4632]: I0313 12:00:00.592174 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" Mar 13 12:00:01 crc kubenswrapper[4632]: I0313 12:00:01.365872 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7"] Mar 13 12:00:01 crc kubenswrapper[4632]: W0313 12:00:01.369024 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod453f2bd4_a723_4b7f_9b06_05d75e8df7b8.slice/crio-e9f5d693baba17da6616fa665348293847d1c5fe75db701538d10a5b75834253 WatchSource:0}: Error finding container e9f5d693baba17da6616fa665348293847d1c5fe75db701538d10a5b75834253: Status 404 returned error can't find the container with id e9f5d693baba17da6616fa665348293847d1c5fe75db701538d10a5b75834253 Mar 13 12:00:01 crc kubenswrapper[4632]: W0313 12:00:01.506527 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod614f8dd7_8a57_4b22_b741_63c3ed563216.slice/crio-9e7bb3a0a8747b8c4cac419fde52b47e802a05c159ca271aa9d70f819a6c2958 WatchSource:0}: Error finding container 9e7bb3a0a8747b8c4cac419fde52b47e802a05c159ca271aa9d70f819a6c2958: Status 404 returned error can't find the container with id 9e7bb3a0a8747b8c4cac419fde52b47e802a05c159ca271aa9d70f819a6c2958 Mar 13 12:00:01 crc kubenswrapper[4632]: I0313 12:00:01.511159 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556720-2985g"] Mar 13 12:00:02 crc kubenswrapper[4632]: I0313 12:00:02.139036 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" event={"ID":"453f2bd4-a723-4b7f-9b06-05d75e8df7b8","Type":"ContainerStarted","Data":"513ba32a6f64209e9e7a4b86369065ec16320243702d6e9f6899a7182c651338"} Mar 13 12:00:02 crc kubenswrapper[4632]: I0313 12:00:02.139421 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" event={"ID":"453f2bd4-a723-4b7f-9b06-05d75e8df7b8","Type":"ContainerStarted","Data":"e9f5d693baba17da6616fa665348293847d1c5fe75db701538d10a5b75834253"} Mar 13 12:00:02 crc kubenswrapper[4632]: I0313 12:00:02.141185 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556720-2985g" event={"ID":"614f8dd7-8a57-4b22-b741-63c3ed563216","Type":"ContainerStarted","Data":"9e7bb3a0a8747b8c4cac419fde52b47e802a05c159ca271aa9d70f819a6c2958"} Mar 13 12:00:02 crc kubenswrapper[4632]: I0313 12:00:02.163087 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" podStartSLOduration=2.163068723 podStartE2EDuration="2.163068723s" podCreationTimestamp="2026-03-13 12:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 12:00:02.161445753 +0000 UTC m=+6976.183975906" watchObservedRunningTime="2026-03-13 12:00:02.163068723 +0000 UTC m=+6976.185598856" Mar 13 12:00:03 crc kubenswrapper[4632]: I0313 12:00:03.152870 4632 generic.go:334] "Generic (PLEG): container finished" podID="453f2bd4-a723-4b7f-9b06-05d75e8df7b8" containerID="513ba32a6f64209e9e7a4b86369065ec16320243702d6e9f6899a7182c651338" exitCode=0 Mar 13 12:00:03 crc kubenswrapper[4632]: I0313 12:00:03.153231 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" event={"ID":"453f2bd4-a723-4b7f-9b06-05d75e8df7b8","Type":"ContainerDied","Data":"513ba32a6f64209e9e7a4b86369065ec16320243702d6e9f6899a7182c651338"} Mar 13 12:00:04 crc kubenswrapper[4632]: I0313 12:00:04.545418 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" Mar 13 12:00:04 crc kubenswrapper[4632]: I0313 12:00:04.600488 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-secret-volume\") pod \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\" (UID: \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\") " Mar 13 12:00:04 crc kubenswrapper[4632]: I0313 12:00:04.600555 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-config-volume\") pod \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\" (UID: \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\") " Mar 13 12:00:04 crc kubenswrapper[4632]: I0313 12:00:04.600616 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jmzs7\" (UniqueName: \"kubernetes.io/projected/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-kube-api-access-jmzs7\") pod \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\" (UID: \"453f2bd4-a723-4b7f-9b06-05d75e8df7b8\") " Mar 13 12:00:04 crc kubenswrapper[4632]: I0313 12:00:04.602773 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-config-volume" (OuterVolumeSpecName: "config-volume") pod "453f2bd4-a723-4b7f-9b06-05d75e8df7b8" (UID: "453f2bd4-a723-4b7f-9b06-05d75e8df7b8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 12:00:04 crc kubenswrapper[4632]: I0313 12:00:04.610117 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-kube-api-access-jmzs7" (OuterVolumeSpecName: "kube-api-access-jmzs7") pod "453f2bd4-a723-4b7f-9b06-05d75e8df7b8" (UID: "453f2bd4-a723-4b7f-9b06-05d75e8df7b8"). InnerVolumeSpecName "kube-api-access-jmzs7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:00:04 crc kubenswrapper[4632]: I0313 12:00:04.617893 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "453f2bd4-a723-4b7f-9b06-05d75e8df7b8" (UID: "453f2bd4-a723-4b7f-9b06-05d75e8df7b8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:00:04 crc kubenswrapper[4632]: I0313 12:00:04.703203 4632 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 12:00:04 crc kubenswrapper[4632]: I0313 12:00:04.703244 4632 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 12:00:04 crc kubenswrapper[4632]: I0313 12:00:04.703254 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jmzs7\" (UniqueName: \"kubernetes.io/projected/453f2bd4-a723-4b7f-9b06-05d75e8df7b8-kube-api-access-jmzs7\") on node \"crc\" DevicePath \"\"" Mar 13 12:00:05 crc kubenswrapper[4632]: I0313 12:00:05.174862 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" event={"ID":"453f2bd4-a723-4b7f-9b06-05d75e8df7b8","Type":"ContainerDied","Data":"e9f5d693baba17da6616fa665348293847d1c5fe75db701538d10a5b75834253"} Mar 13 12:00:05 crc kubenswrapper[4632]: I0313 12:00:05.175167 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9f5d693baba17da6616fa665348293847d1c5fe75db701538d10a5b75834253" Mar 13 12:00:05 crc kubenswrapper[4632]: I0313 12:00:05.174894 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7" Mar 13 12:00:05 crc kubenswrapper[4632]: I0313 12:00:05.177193 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556720-2985g" event={"ID":"614f8dd7-8a57-4b22-b741-63c3ed563216","Type":"ContainerStarted","Data":"d1793c511542a0a35aa5afc5e36e94033f5a38dca84400ac841ea4e47dc426f7"} Mar 13 12:00:05 crc kubenswrapper[4632]: I0313 12:00:05.202746 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556720-2985g" podStartSLOduration=1.870040328 podStartE2EDuration="5.202723103s" podCreationTimestamp="2026-03-13 12:00:00 +0000 UTC" firstStartedPulling="2026-03-13 12:00:01.508786245 +0000 UTC m=+6975.531316388" lastFinishedPulling="2026-03-13 12:00:04.84146903 +0000 UTC m=+6978.863999163" observedRunningTime="2026-03-13 12:00:05.192273336 +0000 UTC m=+6979.214803469" watchObservedRunningTime="2026-03-13 12:00:05.202723103 +0000 UTC m=+6979.225253236" Mar 13 12:00:05 crc kubenswrapper[4632]: I0313 12:00:05.620532 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8"] Mar 13 12:00:05 crc kubenswrapper[4632]: I0313 12:00:05.631814 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556675-n64w8"] Mar 13 12:00:06 crc kubenswrapper[4632]: I0313 12:00:06.056293 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9481bb7b-d00a-4ee1-b711-7b90d97907c1" path="/var/lib/kubelet/pods/9481bb7b-d00a-4ee1-b711-7b90d97907c1/volumes" Mar 13 12:00:07 crc kubenswrapper[4632]: I0313 12:00:07.196559 4632 generic.go:334] "Generic (PLEG): container finished" podID="614f8dd7-8a57-4b22-b741-63c3ed563216" containerID="d1793c511542a0a35aa5afc5e36e94033f5a38dca84400ac841ea4e47dc426f7" exitCode=0 Mar 13 12:00:07 crc kubenswrapper[4632]: I0313 12:00:07.196710 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556720-2985g" event={"ID":"614f8dd7-8a57-4b22-b741-63c3ed563216","Type":"ContainerDied","Data":"d1793c511542a0a35aa5afc5e36e94033f5a38dca84400ac841ea4e47dc426f7"} Mar 13 12:00:08 crc kubenswrapper[4632]: I0313 12:00:08.663205 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556720-2985g" Mar 13 12:00:08 crc kubenswrapper[4632]: I0313 12:00:08.678549 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z5vxf\" (UniqueName: \"kubernetes.io/projected/614f8dd7-8a57-4b22-b741-63c3ed563216-kube-api-access-z5vxf\") pod \"614f8dd7-8a57-4b22-b741-63c3ed563216\" (UID: \"614f8dd7-8a57-4b22-b741-63c3ed563216\") " Mar 13 12:00:08 crc kubenswrapper[4632]: I0313 12:00:08.687238 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/614f8dd7-8a57-4b22-b741-63c3ed563216-kube-api-access-z5vxf" (OuterVolumeSpecName: "kube-api-access-z5vxf") pod "614f8dd7-8a57-4b22-b741-63c3ed563216" (UID: "614f8dd7-8a57-4b22-b741-63c3ed563216"). InnerVolumeSpecName "kube-api-access-z5vxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:00:08 crc kubenswrapper[4632]: I0313 12:00:08.780976 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z5vxf\" (UniqueName: \"kubernetes.io/projected/614f8dd7-8a57-4b22-b741-63c3ed563216-kube-api-access-z5vxf\") on node \"crc\" DevicePath \"\"" Mar 13 12:00:09 crc kubenswrapper[4632]: I0313 12:00:09.217867 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556720-2985g" event={"ID":"614f8dd7-8a57-4b22-b741-63c3ed563216","Type":"ContainerDied","Data":"9e7bb3a0a8747b8c4cac419fde52b47e802a05c159ca271aa9d70f819a6c2958"} Mar 13 12:00:09 crc kubenswrapper[4632]: I0313 12:00:09.217908 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e7bb3a0a8747b8c4cac419fde52b47e802a05c159ca271aa9d70f819a6c2958" Mar 13 12:00:09 crc kubenswrapper[4632]: I0313 12:00:09.217964 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556720-2985g" Mar 13 12:00:09 crc kubenswrapper[4632]: I0313 12:00:09.268541 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556714-cnqc7"] Mar 13 12:00:09 crc kubenswrapper[4632]: I0313 12:00:09.277854 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556714-cnqc7"] Mar 13 12:00:10 crc kubenswrapper[4632]: I0313 12:00:10.054620 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88fb48fa-6650-4b22-b44f-d8c6f839489e" path="/var/lib/kubelet/pods/88fb48fa-6650-4b22-b44f-d8c6f839489e/volumes" Mar 13 12:00:10 crc kubenswrapper[4632]: I0313 12:00:10.461073 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:00:10 crc kubenswrapper[4632]: I0313 12:00:10.461185 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:00:10 crc kubenswrapper[4632]: I0313 12:00:10.461256 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 12:00:10 crc kubenswrapper[4632]: I0313 12:00:10.462497 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f1337fb64ab38c0f489a591d3b3f173d13428642427113f1891b2f17a626304e"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 12:00:10 crc kubenswrapper[4632]: I0313 12:00:10.462583 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://f1337fb64ab38c0f489a591d3b3f173d13428642427113f1891b2f17a626304e" gracePeriod=600 Mar 13 12:00:11 crc kubenswrapper[4632]: I0313 12:00:11.254213 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="f1337fb64ab38c0f489a591d3b3f173d13428642427113f1891b2f17a626304e" exitCode=0 Mar 13 12:00:11 crc kubenswrapper[4632]: I0313 12:00:11.254296 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"f1337fb64ab38c0f489a591d3b3f173d13428642427113f1891b2f17a626304e"} Mar 13 12:00:11 crc kubenswrapper[4632]: I0313 12:00:11.254537 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6"} Mar 13 12:00:11 crc kubenswrapper[4632]: I0313 12:00:11.254555 4632 scope.go:117] "RemoveContainer" containerID="09f62a713fe019f208bcff213bc55f14995ec3a8014d027c3bf7cfc3b5b612e6" Mar 13 12:00:23 crc kubenswrapper[4632]: I0313 12:00:23.998478 4632 scope.go:117] "RemoveContainer" containerID="95360e41112a84b3ea4b235c3e7fd03654d6110fccc446520298cff419091ae2" Mar 13 12:00:24 crc kubenswrapper[4632]: I0313 12:00:24.031549 4632 scope.go:117] "RemoveContainer" containerID="a65145278e359710e5ff339e23940020997c56d82631e8e73d581b3ec62c80b2" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.168752 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29556721-l54tk"] Mar 13 12:01:00 crc kubenswrapper[4632]: E0313 12:01:00.169728 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="453f2bd4-a723-4b7f-9b06-05d75e8df7b8" containerName="collect-profiles" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.169744 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="453f2bd4-a723-4b7f-9b06-05d75e8df7b8" containerName="collect-profiles" Mar 13 12:01:00 crc kubenswrapper[4632]: E0313 12:01:00.169766 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="614f8dd7-8a57-4b22-b741-63c3ed563216" containerName="oc" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.169773 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="614f8dd7-8a57-4b22-b741-63c3ed563216" containerName="oc" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.169992 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="453f2bd4-a723-4b7f-9b06-05d75e8df7b8" containerName="collect-profiles" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.170008 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="614f8dd7-8a57-4b22-b741-63c3ed563216" containerName="oc" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.170679 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.185227 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29556721-l54tk"] Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.293645 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d8xzd\" (UniqueName: \"kubernetes.io/projected/c8daf4c2-f012-4d18-b11a-e666e00d6a03-kube-api-access-d8xzd\") pod \"keystone-cron-29556721-l54tk\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.293734 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-combined-ca-bundle\") pod \"keystone-cron-29556721-l54tk\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.293759 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-config-data\") pod \"keystone-cron-29556721-l54tk\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.293840 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-fernet-keys\") pod \"keystone-cron-29556721-l54tk\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.395915 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-fernet-keys\") pod \"keystone-cron-29556721-l54tk\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.396542 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d8xzd\" (UniqueName: \"kubernetes.io/projected/c8daf4c2-f012-4d18-b11a-e666e00d6a03-kube-api-access-d8xzd\") pod \"keystone-cron-29556721-l54tk\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.396684 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-combined-ca-bundle\") pod \"keystone-cron-29556721-l54tk\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.396789 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-config-data\") pod \"keystone-cron-29556721-l54tk\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.403385 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-combined-ca-bundle\") pod \"keystone-cron-29556721-l54tk\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.408393 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-fernet-keys\") pod \"keystone-cron-29556721-l54tk\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.410118 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-config-data\") pod \"keystone-cron-29556721-l54tk\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.422194 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d8xzd\" (UniqueName: \"kubernetes.io/projected/c8daf4c2-f012-4d18-b11a-e666e00d6a03-kube-api-access-d8xzd\") pod \"keystone-cron-29556721-l54tk\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:00 crc kubenswrapper[4632]: I0313 12:01:00.490734 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:01 crc kubenswrapper[4632]: I0313 12:01:01.070641 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29556721-l54tk"] Mar 13 12:01:01 crc kubenswrapper[4632]: I0313 12:01:01.541618 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556721-l54tk" event={"ID":"c8daf4c2-f012-4d18-b11a-e666e00d6a03","Type":"ContainerStarted","Data":"0a8c37300ecb47a610879d64ee89a84f73a8a208385e27ede8cead0b63ad2bee"} Mar 13 12:01:01 crc kubenswrapper[4632]: I0313 12:01:01.542019 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556721-l54tk" event={"ID":"c8daf4c2-f012-4d18-b11a-e666e00d6a03","Type":"ContainerStarted","Data":"0391d4900a2de8d700528f304be2206114d264ce6de429a1a1cb7769b6adf632"} Mar 13 12:01:01 crc kubenswrapper[4632]: I0313 12:01:01.563052 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29556721-l54tk" podStartSLOduration=1.563030205 podStartE2EDuration="1.563030205s" podCreationTimestamp="2026-03-13 12:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 12:01:01.557785129 +0000 UTC m=+7035.580315282" watchObservedRunningTime="2026-03-13 12:01:01.563030205 +0000 UTC m=+7035.585560338" Mar 13 12:01:08 crc kubenswrapper[4632]: I0313 12:01:08.602619 4632 generic.go:334] "Generic (PLEG): container finished" podID="c8daf4c2-f012-4d18-b11a-e666e00d6a03" containerID="0a8c37300ecb47a610879d64ee89a84f73a8a208385e27ede8cead0b63ad2bee" exitCode=0 Mar 13 12:01:08 crc kubenswrapper[4632]: I0313 12:01:08.602698 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556721-l54tk" event={"ID":"c8daf4c2-f012-4d18-b11a-e666e00d6a03","Type":"ContainerDied","Data":"0a8c37300ecb47a610879d64ee89a84f73a8a208385e27ede8cead0b63ad2bee"} Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.060825 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.193132 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-fernet-keys\") pod \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.193219 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d8xzd\" (UniqueName: \"kubernetes.io/projected/c8daf4c2-f012-4d18-b11a-e666e00d6a03-kube-api-access-d8xzd\") pod \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.193356 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-combined-ca-bundle\") pod \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.193400 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-config-data\") pod \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\" (UID: \"c8daf4c2-f012-4d18-b11a-e666e00d6a03\") " Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.210561 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "c8daf4c2-f012-4d18-b11a-e666e00d6a03" (UID: "c8daf4c2-f012-4d18-b11a-e666e00d6a03"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.210641 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c8daf4c2-f012-4d18-b11a-e666e00d6a03-kube-api-access-d8xzd" (OuterVolumeSpecName: "kube-api-access-d8xzd") pod "c8daf4c2-f012-4d18-b11a-e666e00d6a03" (UID: "c8daf4c2-f012-4d18-b11a-e666e00d6a03"). InnerVolumeSpecName "kube-api-access-d8xzd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.229100 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c8daf4c2-f012-4d18-b11a-e666e00d6a03" (UID: "c8daf4c2-f012-4d18-b11a-e666e00d6a03"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.269181 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-config-data" (OuterVolumeSpecName: "config-data") pod "c8daf4c2-f012-4d18-b11a-e666e00d6a03" (UID: "c8daf4c2-f012-4d18-b11a-e666e00d6a03"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.299892 4632 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.300177 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d8xzd\" (UniqueName: \"kubernetes.io/projected/c8daf4c2-f012-4d18-b11a-e666e00d6a03-kube-api-access-d8xzd\") on node \"crc\" DevicePath \"\"" Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.300283 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.300359 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8daf4c2-f012-4d18-b11a-e666e00d6a03-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.628754 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556721-l54tk" event={"ID":"c8daf4c2-f012-4d18-b11a-e666e00d6a03","Type":"ContainerDied","Data":"0391d4900a2de8d700528f304be2206114d264ce6de429a1a1cb7769b6adf632"} Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.629042 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0391d4900a2de8d700528f304be2206114d264ce6de429a1a1cb7769b6adf632" Mar 13 12:01:10 crc kubenswrapper[4632]: I0313 12:01:10.628848 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556721-l54tk" Mar 13 12:01:35 crc kubenswrapper[4632]: I0313 12:01:35.999169 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-dgt68"] Mar 13 12:01:36 crc kubenswrapper[4632]: E0313 12:01:36.000387 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c8daf4c2-f012-4d18-b11a-e666e00d6a03" containerName="keystone-cron" Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.000411 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c8daf4c2-f012-4d18-b11a-e666e00d6a03" containerName="keystone-cron" Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.000671 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c8daf4c2-f012-4d18-b11a-e666e00d6a03" containerName="keystone-cron" Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.002647 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.009670 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dgt68"] Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.042395 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccc5cad2-2d89-458e-826f-12b47e70afd6-catalog-content\") pod \"certified-operators-dgt68\" (UID: \"ccc5cad2-2d89-458e-826f-12b47e70afd6\") " pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.042700 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccc5cad2-2d89-458e-826f-12b47e70afd6-utilities\") pod \"certified-operators-dgt68\" (UID: \"ccc5cad2-2d89-458e-826f-12b47e70afd6\") " pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.043001 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6df8w\" (UniqueName: \"kubernetes.io/projected/ccc5cad2-2d89-458e-826f-12b47e70afd6-kube-api-access-6df8w\") pod \"certified-operators-dgt68\" (UID: \"ccc5cad2-2d89-458e-826f-12b47e70afd6\") " pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.145284 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccc5cad2-2d89-458e-826f-12b47e70afd6-utilities\") pod \"certified-operators-dgt68\" (UID: \"ccc5cad2-2d89-458e-826f-12b47e70afd6\") " pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.145465 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6df8w\" (UniqueName: \"kubernetes.io/projected/ccc5cad2-2d89-458e-826f-12b47e70afd6-kube-api-access-6df8w\") pod \"certified-operators-dgt68\" (UID: \"ccc5cad2-2d89-458e-826f-12b47e70afd6\") " pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.145707 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccc5cad2-2d89-458e-826f-12b47e70afd6-catalog-content\") pod \"certified-operators-dgt68\" (UID: \"ccc5cad2-2d89-458e-826f-12b47e70afd6\") " pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.146788 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccc5cad2-2d89-458e-826f-12b47e70afd6-utilities\") pod \"certified-operators-dgt68\" (UID: \"ccc5cad2-2d89-458e-826f-12b47e70afd6\") " pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.147313 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccc5cad2-2d89-458e-826f-12b47e70afd6-catalog-content\") pod \"certified-operators-dgt68\" (UID: \"ccc5cad2-2d89-458e-826f-12b47e70afd6\") " pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.167172 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6df8w\" (UniqueName: \"kubernetes.io/projected/ccc5cad2-2d89-458e-826f-12b47e70afd6-kube-api-access-6df8w\") pod \"certified-operators-dgt68\" (UID: \"ccc5cad2-2d89-458e-826f-12b47e70afd6\") " pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.325200 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.835089 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-dgt68"] Mar 13 12:01:36 crc kubenswrapper[4632]: I0313 12:01:36.883659 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgt68" event={"ID":"ccc5cad2-2d89-458e-826f-12b47e70afd6","Type":"ContainerStarted","Data":"eb7530bf3d8d12731d0a9602228b54fda149b37c977f13fb24f5acdf816e2ca5"} Mar 13 12:01:37 crc kubenswrapper[4632]: I0313 12:01:37.897784 4632 generic.go:334] "Generic (PLEG): container finished" podID="ccc5cad2-2d89-458e-826f-12b47e70afd6" containerID="fe41022c2ad0a86e947222937b1ec880188ea79ad756a089f29f058200df6736" exitCode=0 Mar 13 12:01:37 crc kubenswrapper[4632]: I0313 12:01:37.897846 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgt68" event={"ID":"ccc5cad2-2d89-458e-826f-12b47e70afd6","Type":"ContainerDied","Data":"fe41022c2ad0a86e947222937b1ec880188ea79ad756a089f29f058200df6736"} Mar 13 12:01:37 crc kubenswrapper[4632]: I0313 12:01:37.901996 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 12:01:39 crc kubenswrapper[4632]: I0313 12:01:39.917705 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgt68" event={"ID":"ccc5cad2-2d89-458e-826f-12b47e70afd6","Type":"ContainerStarted","Data":"21ced36348c43768801a7eb8ff5c40cfd7107b222ad4e8d17045732a44d8e9a7"} Mar 13 12:01:42 crc kubenswrapper[4632]: I0313 12:01:42.951532 4632 generic.go:334] "Generic (PLEG): container finished" podID="ccc5cad2-2d89-458e-826f-12b47e70afd6" containerID="21ced36348c43768801a7eb8ff5c40cfd7107b222ad4e8d17045732a44d8e9a7" exitCode=0 Mar 13 12:01:42 crc kubenswrapper[4632]: I0313 12:01:42.951608 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgt68" event={"ID":"ccc5cad2-2d89-458e-826f-12b47e70afd6","Type":"ContainerDied","Data":"21ced36348c43768801a7eb8ff5c40cfd7107b222ad4e8d17045732a44d8e9a7"} Mar 13 12:01:43 crc kubenswrapper[4632]: I0313 12:01:43.964811 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgt68" event={"ID":"ccc5cad2-2d89-458e-826f-12b47e70afd6","Type":"ContainerStarted","Data":"e227e3bd9f407b1dd5a86c9decd8ab735f6ca6d90d2a001533af3524de0eecdf"} Mar 13 12:01:44 crc kubenswrapper[4632]: I0313 12:01:44.009717 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-dgt68" podStartSLOduration=3.5380606930000003 podStartE2EDuration="9.009694821s" podCreationTimestamp="2026-03-13 12:01:35 +0000 UTC" firstStartedPulling="2026-03-13 12:01:37.90084969 +0000 UTC m=+7071.923379823" lastFinishedPulling="2026-03-13 12:01:43.372483818 +0000 UTC m=+7077.395013951" observedRunningTime="2026-03-13 12:01:44.003468161 +0000 UTC m=+7078.025998314" watchObservedRunningTime="2026-03-13 12:01:44.009694821 +0000 UTC m=+7078.032224954" Mar 13 12:01:46 crc kubenswrapper[4632]: I0313 12:01:46.325841 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:01:46 crc kubenswrapper[4632]: I0313 12:01:46.325908 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:01:47 crc kubenswrapper[4632]: I0313 12:01:47.384618 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-dgt68" podUID="ccc5cad2-2d89-458e-826f-12b47e70afd6" containerName="registry-server" probeResult="failure" output=< Mar 13 12:01:47 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:01:47 crc kubenswrapper[4632]: > Mar 13 12:01:57 crc kubenswrapper[4632]: I0313 12:01:57.383391 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-dgt68" podUID="ccc5cad2-2d89-458e-826f-12b47e70afd6" containerName="registry-server" probeResult="failure" output=< Mar 13 12:01:57 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:01:57 crc kubenswrapper[4632]: > Mar 13 12:02:00 crc kubenswrapper[4632]: I0313 12:02:00.151446 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556722-r4pp4"] Mar 13 12:02:00 crc kubenswrapper[4632]: I0313 12:02:00.153466 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556722-r4pp4" Mar 13 12:02:00 crc kubenswrapper[4632]: I0313 12:02:00.155546 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:02:00 crc kubenswrapper[4632]: I0313 12:02:00.155952 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:02:00 crc kubenswrapper[4632]: I0313 12:02:00.162581 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556722-r4pp4"] Mar 13 12:02:00 crc kubenswrapper[4632]: I0313 12:02:00.163243 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:02:00 crc kubenswrapper[4632]: I0313 12:02:00.300734 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zgf5\" (UniqueName: \"kubernetes.io/projected/5ae03001-344b-4e5e-baf2-c8171109eb1a-kube-api-access-8zgf5\") pod \"auto-csr-approver-29556722-r4pp4\" (UID: \"5ae03001-344b-4e5e-baf2-c8171109eb1a\") " pod="openshift-infra/auto-csr-approver-29556722-r4pp4" Mar 13 12:02:00 crc kubenswrapper[4632]: I0313 12:02:00.402553 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zgf5\" (UniqueName: \"kubernetes.io/projected/5ae03001-344b-4e5e-baf2-c8171109eb1a-kube-api-access-8zgf5\") pod \"auto-csr-approver-29556722-r4pp4\" (UID: \"5ae03001-344b-4e5e-baf2-c8171109eb1a\") " pod="openshift-infra/auto-csr-approver-29556722-r4pp4" Mar 13 12:02:00 crc kubenswrapper[4632]: I0313 12:02:00.428408 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zgf5\" (UniqueName: \"kubernetes.io/projected/5ae03001-344b-4e5e-baf2-c8171109eb1a-kube-api-access-8zgf5\") pod \"auto-csr-approver-29556722-r4pp4\" (UID: \"5ae03001-344b-4e5e-baf2-c8171109eb1a\") " pod="openshift-infra/auto-csr-approver-29556722-r4pp4" Mar 13 12:02:00 crc kubenswrapper[4632]: I0313 12:02:00.477202 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556722-r4pp4" Mar 13 12:02:01 crc kubenswrapper[4632]: I0313 12:02:01.088686 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556722-r4pp4"] Mar 13 12:02:01 crc kubenswrapper[4632]: I0313 12:02:01.128192 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556722-r4pp4" event={"ID":"5ae03001-344b-4e5e-baf2-c8171109eb1a","Type":"ContainerStarted","Data":"dd78aa588ce6352a368cfc59900ce80669e0738f4b75bcdd822f3fb6256d9f3a"} Mar 13 12:02:03 crc kubenswrapper[4632]: I0313 12:02:03.151684 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556722-r4pp4" event={"ID":"5ae03001-344b-4e5e-baf2-c8171109eb1a","Type":"ContainerStarted","Data":"45cc8231b2d3d1ca8cec2f7a7da9147a3a370632b65df6ead7d138b8c0f615b3"} Mar 13 12:02:03 crc kubenswrapper[4632]: I0313 12:02:03.173233 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556722-r4pp4" podStartSLOduration=2.122412012 podStartE2EDuration="3.173210619s" podCreationTimestamp="2026-03-13 12:02:00 +0000 UTC" firstStartedPulling="2026-03-13 12:02:01.092317683 +0000 UTC m=+7095.114847816" lastFinishedPulling="2026-03-13 12:02:02.1431163 +0000 UTC m=+7096.165646423" observedRunningTime="2026-03-13 12:02:03.166893257 +0000 UTC m=+7097.189423380" watchObservedRunningTime="2026-03-13 12:02:03.173210619 +0000 UTC m=+7097.195740752" Mar 13 12:02:05 crc kubenswrapper[4632]: I0313 12:02:05.172419 4632 generic.go:334] "Generic (PLEG): container finished" podID="5ae03001-344b-4e5e-baf2-c8171109eb1a" containerID="45cc8231b2d3d1ca8cec2f7a7da9147a3a370632b65df6ead7d138b8c0f615b3" exitCode=0 Mar 13 12:02:05 crc kubenswrapper[4632]: I0313 12:02:05.172519 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556722-r4pp4" event={"ID":"5ae03001-344b-4e5e-baf2-c8171109eb1a","Type":"ContainerDied","Data":"45cc8231b2d3d1ca8cec2f7a7da9147a3a370632b65df6ead7d138b8c0f615b3"} Mar 13 12:02:06 crc kubenswrapper[4632]: I0313 12:02:06.380842 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:02:06 crc kubenswrapper[4632]: I0313 12:02:06.439302 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:02:06 crc kubenswrapper[4632]: I0313 12:02:06.970160 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556722-r4pp4" Mar 13 12:02:07 crc kubenswrapper[4632]: I0313 12:02:07.055267 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zgf5\" (UniqueName: \"kubernetes.io/projected/5ae03001-344b-4e5e-baf2-c8171109eb1a-kube-api-access-8zgf5\") pod \"5ae03001-344b-4e5e-baf2-c8171109eb1a\" (UID: \"5ae03001-344b-4e5e-baf2-c8171109eb1a\") " Mar 13 12:02:07 crc kubenswrapper[4632]: I0313 12:02:07.067319 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ae03001-344b-4e5e-baf2-c8171109eb1a-kube-api-access-8zgf5" (OuterVolumeSpecName: "kube-api-access-8zgf5") pod "5ae03001-344b-4e5e-baf2-c8171109eb1a" (UID: "5ae03001-344b-4e5e-baf2-c8171109eb1a"). InnerVolumeSpecName "kube-api-access-8zgf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:02:07 crc kubenswrapper[4632]: I0313 12:02:07.159401 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zgf5\" (UniqueName: \"kubernetes.io/projected/5ae03001-344b-4e5e-baf2-c8171109eb1a-kube-api-access-8zgf5\") on node \"crc\" DevicePath \"\"" Mar 13 12:02:07 crc kubenswrapper[4632]: I0313 12:02:07.196337 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556722-r4pp4" event={"ID":"5ae03001-344b-4e5e-baf2-c8171109eb1a","Type":"ContainerDied","Data":"dd78aa588ce6352a368cfc59900ce80669e0738f4b75bcdd822f3fb6256d9f3a"} Mar 13 12:02:07 crc kubenswrapper[4632]: I0313 12:02:07.196414 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dd78aa588ce6352a368cfc59900ce80669e0738f4b75bcdd822f3fb6256d9f3a" Mar 13 12:02:07 crc kubenswrapper[4632]: I0313 12:02:07.196379 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556722-r4pp4" Mar 13 12:02:07 crc kubenswrapper[4632]: I0313 12:02:07.207264 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dgt68"] Mar 13 12:02:07 crc kubenswrapper[4632]: I0313 12:02:07.283253 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556716-2gt9h"] Mar 13 12:02:07 crc kubenswrapper[4632]: I0313 12:02:07.301414 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556716-2gt9h"] Mar 13 12:02:08 crc kubenswrapper[4632]: I0313 12:02:08.058213 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="38cafee7-6e61-46de-b58b-48b8f7d41bf6" path="/var/lib/kubelet/pods/38cafee7-6e61-46de-b58b-48b8f7d41bf6/volumes" Mar 13 12:02:08 crc kubenswrapper[4632]: I0313 12:02:08.204670 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-dgt68" podUID="ccc5cad2-2d89-458e-826f-12b47e70afd6" containerName="registry-server" containerID="cri-o://e227e3bd9f407b1dd5a86c9decd8ab735f6ca6d90d2a001533af3524de0eecdf" gracePeriod=2 Mar 13 12:02:08 crc kubenswrapper[4632]: I0313 12:02:08.715617 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:02:08 crc kubenswrapper[4632]: I0313 12:02:08.791174 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccc5cad2-2d89-458e-826f-12b47e70afd6-utilities\") pod \"ccc5cad2-2d89-458e-826f-12b47e70afd6\" (UID: \"ccc5cad2-2d89-458e-826f-12b47e70afd6\") " Mar 13 12:02:08 crc kubenswrapper[4632]: I0313 12:02:08.791427 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6df8w\" (UniqueName: \"kubernetes.io/projected/ccc5cad2-2d89-458e-826f-12b47e70afd6-kube-api-access-6df8w\") pod \"ccc5cad2-2d89-458e-826f-12b47e70afd6\" (UID: \"ccc5cad2-2d89-458e-826f-12b47e70afd6\") " Mar 13 12:02:08 crc kubenswrapper[4632]: I0313 12:02:08.791532 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccc5cad2-2d89-458e-826f-12b47e70afd6-catalog-content\") pod \"ccc5cad2-2d89-458e-826f-12b47e70afd6\" (UID: \"ccc5cad2-2d89-458e-826f-12b47e70afd6\") " Mar 13 12:02:08 crc kubenswrapper[4632]: I0313 12:02:08.801544 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccc5cad2-2d89-458e-826f-12b47e70afd6-utilities" (OuterVolumeSpecName: "utilities") pod "ccc5cad2-2d89-458e-826f-12b47e70afd6" (UID: "ccc5cad2-2d89-458e-826f-12b47e70afd6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:02:08 crc kubenswrapper[4632]: I0313 12:02:08.808702 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccc5cad2-2d89-458e-826f-12b47e70afd6-kube-api-access-6df8w" (OuterVolumeSpecName: "kube-api-access-6df8w") pod "ccc5cad2-2d89-458e-826f-12b47e70afd6" (UID: "ccc5cad2-2d89-458e-826f-12b47e70afd6"). InnerVolumeSpecName "kube-api-access-6df8w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:02:08 crc kubenswrapper[4632]: I0313 12:02:08.898139 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ccc5cad2-2d89-458e-826f-12b47e70afd6-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:02:08 crc kubenswrapper[4632]: I0313 12:02:08.898174 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6df8w\" (UniqueName: \"kubernetes.io/projected/ccc5cad2-2d89-458e-826f-12b47e70afd6-kube-api-access-6df8w\") on node \"crc\" DevicePath \"\"" Mar 13 12:02:08 crc kubenswrapper[4632]: I0313 12:02:08.920169 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ccc5cad2-2d89-458e-826f-12b47e70afd6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ccc5cad2-2d89-458e-826f-12b47e70afd6" (UID: "ccc5cad2-2d89-458e-826f-12b47e70afd6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:02:08 crc kubenswrapper[4632]: I0313 12:02:08.999711 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ccc5cad2-2d89-458e-826f-12b47e70afd6-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.219319 4632 generic.go:334] "Generic (PLEG): container finished" podID="ccc5cad2-2d89-458e-826f-12b47e70afd6" containerID="e227e3bd9f407b1dd5a86c9decd8ab735f6ca6d90d2a001533af3524de0eecdf" exitCode=0 Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.219390 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-dgt68" Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.219423 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgt68" event={"ID":"ccc5cad2-2d89-458e-826f-12b47e70afd6","Type":"ContainerDied","Data":"e227e3bd9f407b1dd5a86c9decd8ab735f6ca6d90d2a001533af3524de0eecdf"} Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.219881 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-dgt68" event={"ID":"ccc5cad2-2d89-458e-826f-12b47e70afd6","Type":"ContainerDied","Data":"eb7530bf3d8d12731d0a9602228b54fda149b37c977f13fb24f5acdf816e2ca5"} Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.219921 4632 scope.go:117] "RemoveContainer" containerID="e227e3bd9f407b1dd5a86c9decd8ab735f6ca6d90d2a001533af3524de0eecdf" Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.251801 4632 scope.go:117] "RemoveContainer" containerID="21ced36348c43768801a7eb8ff5c40cfd7107b222ad4e8d17045732a44d8e9a7" Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.266038 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-dgt68"] Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.278441 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-dgt68"] Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.322130 4632 scope.go:117] "RemoveContainer" containerID="fe41022c2ad0a86e947222937b1ec880188ea79ad756a089f29f058200df6736" Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.365375 4632 scope.go:117] "RemoveContainer" containerID="e227e3bd9f407b1dd5a86c9decd8ab735f6ca6d90d2a001533af3524de0eecdf" Mar 13 12:02:09 crc kubenswrapper[4632]: E0313 12:02:09.366117 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e227e3bd9f407b1dd5a86c9decd8ab735f6ca6d90d2a001533af3524de0eecdf\": container with ID starting with e227e3bd9f407b1dd5a86c9decd8ab735f6ca6d90d2a001533af3524de0eecdf not found: ID does not exist" containerID="e227e3bd9f407b1dd5a86c9decd8ab735f6ca6d90d2a001533af3524de0eecdf" Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.366160 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e227e3bd9f407b1dd5a86c9decd8ab735f6ca6d90d2a001533af3524de0eecdf"} err="failed to get container status \"e227e3bd9f407b1dd5a86c9decd8ab735f6ca6d90d2a001533af3524de0eecdf\": rpc error: code = NotFound desc = could not find container \"e227e3bd9f407b1dd5a86c9decd8ab735f6ca6d90d2a001533af3524de0eecdf\": container with ID starting with e227e3bd9f407b1dd5a86c9decd8ab735f6ca6d90d2a001533af3524de0eecdf not found: ID does not exist" Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.366189 4632 scope.go:117] "RemoveContainer" containerID="21ced36348c43768801a7eb8ff5c40cfd7107b222ad4e8d17045732a44d8e9a7" Mar 13 12:02:09 crc kubenswrapper[4632]: E0313 12:02:09.366602 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21ced36348c43768801a7eb8ff5c40cfd7107b222ad4e8d17045732a44d8e9a7\": container with ID starting with 21ced36348c43768801a7eb8ff5c40cfd7107b222ad4e8d17045732a44d8e9a7 not found: ID does not exist" containerID="21ced36348c43768801a7eb8ff5c40cfd7107b222ad4e8d17045732a44d8e9a7" Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.366688 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21ced36348c43768801a7eb8ff5c40cfd7107b222ad4e8d17045732a44d8e9a7"} err="failed to get container status \"21ced36348c43768801a7eb8ff5c40cfd7107b222ad4e8d17045732a44d8e9a7\": rpc error: code = NotFound desc = could not find container \"21ced36348c43768801a7eb8ff5c40cfd7107b222ad4e8d17045732a44d8e9a7\": container with ID starting with 21ced36348c43768801a7eb8ff5c40cfd7107b222ad4e8d17045732a44d8e9a7 not found: ID does not exist" Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.366730 4632 scope.go:117] "RemoveContainer" containerID="fe41022c2ad0a86e947222937b1ec880188ea79ad756a089f29f058200df6736" Mar 13 12:02:09 crc kubenswrapper[4632]: E0313 12:02:09.367156 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fe41022c2ad0a86e947222937b1ec880188ea79ad756a089f29f058200df6736\": container with ID starting with fe41022c2ad0a86e947222937b1ec880188ea79ad756a089f29f058200df6736 not found: ID does not exist" containerID="fe41022c2ad0a86e947222937b1ec880188ea79ad756a089f29f058200df6736" Mar 13 12:02:09 crc kubenswrapper[4632]: I0313 12:02:09.367194 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fe41022c2ad0a86e947222937b1ec880188ea79ad756a089f29f058200df6736"} err="failed to get container status \"fe41022c2ad0a86e947222937b1ec880188ea79ad756a089f29f058200df6736\": rpc error: code = NotFound desc = could not find container \"fe41022c2ad0a86e947222937b1ec880188ea79ad756a089f29f058200df6736\": container with ID starting with fe41022c2ad0a86e947222937b1ec880188ea79ad756a089f29f058200df6736 not found: ID does not exist" Mar 13 12:02:10 crc kubenswrapper[4632]: I0313 12:02:10.057550 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccc5cad2-2d89-458e-826f-12b47e70afd6" path="/var/lib/kubelet/pods/ccc5cad2-2d89-458e-826f-12b47e70afd6/volumes" Mar 13 12:02:10 crc kubenswrapper[4632]: I0313 12:02:10.461173 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:02:10 crc kubenswrapper[4632]: I0313 12:02:10.461234 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:02:24 crc kubenswrapper[4632]: I0313 12:02:24.189168 4632 scope.go:117] "RemoveContainer" containerID="334a9f675c9c77aba9558302bf96e3547c17123adf9873e85b3c3871bccb4465" Mar 13 12:02:40 crc kubenswrapper[4632]: I0313 12:02:40.461449 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:02:40 crc kubenswrapper[4632]: I0313 12:02:40.462183 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.193830 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qttkd"] Mar 13 12:02:55 crc kubenswrapper[4632]: E0313 12:02:55.195204 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc5cad2-2d89-458e-826f-12b47e70afd6" containerName="registry-server" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.195243 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc5cad2-2d89-458e-826f-12b47e70afd6" containerName="registry-server" Mar 13 12:02:55 crc kubenswrapper[4632]: E0313 12:02:55.195271 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc5cad2-2d89-458e-826f-12b47e70afd6" containerName="extract-utilities" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.195281 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc5cad2-2d89-458e-826f-12b47e70afd6" containerName="extract-utilities" Mar 13 12:02:55 crc kubenswrapper[4632]: E0313 12:02:55.195298 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ae03001-344b-4e5e-baf2-c8171109eb1a" containerName="oc" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.195307 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ae03001-344b-4e5e-baf2-c8171109eb1a" containerName="oc" Mar 13 12:02:55 crc kubenswrapper[4632]: E0313 12:02:55.195334 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc5cad2-2d89-458e-826f-12b47e70afd6" containerName="extract-content" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.195342 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc5cad2-2d89-458e-826f-12b47e70afd6" containerName="extract-content" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.195586 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ae03001-344b-4e5e-baf2-c8171109eb1a" containerName="oc" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.195613 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccc5cad2-2d89-458e-826f-12b47e70afd6" containerName="registry-server" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.199720 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.217034 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qttkd"] Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.271848 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d15a7682-8687-4495-9ea5-bab97097930e-utilities\") pod \"redhat-operators-qttkd\" (UID: \"d15a7682-8687-4495-9ea5-bab97097930e\") " pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.271928 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d15a7682-8687-4495-9ea5-bab97097930e-catalog-content\") pod \"redhat-operators-qttkd\" (UID: \"d15a7682-8687-4495-9ea5-bab97097930e\") " pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.272152 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px4g4\" (UniqueName: \"kubernetes.io/projected/d15a7682-8687-4495-9ea5-bab97097930e-kube-api-access-px4g4\") pod \"redhat-operators-qttkd\" (UID: \"d15a7682-8687-4495-9ea5-bab97097930e\") " pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.373678 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px4g4\" (UniqueName: \"kubernetes.io/projected/d15a7682-8687-4495-9ea5-bab97097930e-kube-api-access-px4g4\") pod \"redhat-operators-qttkd\" (UID: \"d15a7682-8687-4495-9ea5-bab97097930e\") " pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.373840 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d15a7682-8687-4495-9ea5-bab97097930e-utilities\") pod \"redhat-operators-qttkd\" (UID: \"d15a7682-8687-4495-9ea5-bab97097930e\") " pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.373887 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d15a7682-8687-4495-9ea5-bab97097930e-catalog-content\") pod \"redhat-operators-qttkd\" (UID: \"d15a7682-8687-4495-9ea5-bab97097930e\") " pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.374422 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d15a7682-8687-4495-9ea5-bab97097930e-utilities\") pod \"redhat-operators-qttkd\" (UID: \"d15a7682-8687-4495-9ea5-bab97097930e\") " pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.374471 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d15a7682-8687-4495-9ea5-bab97097930e-catalog-content\") pod \"redhat-operators-qttkd\" (UID: \"d15a7682-8687-4495-9ea5-bab97097930e\") " pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.402584 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px4g4\" (UniqueName: \"kubernetes.io/projected/d15a7682-8687-4495-9ea5-bab97097930e-kube-api-access-px4g4\") pod \"redhat-operators-qttkd\" (UID: \"d15a7682-8687-4495-9ea5-bab97097930e\") " pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:02:55 crc kubenswrapper[4632]: I0313 12:02:55.528442 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:02:56 crc kubenswrapper[4632]: I0313 12:02:56.151101 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qttkd"] Mar 13 12:02:56 crc kubenswrapper[4632]: I0313 12:02:56.674957 4632 generic.go:334] "Generic (PLEG): container finished" podID="d15a7682-8687-4495-9ea5-bab97097930e" containerID="28be9f7463878e855f3ddab509b5ad179e18babc64a9ca78c7b8ffe29d0be6e0" exitCode=0 Mar 13 12:02:56 crc kubenswrapper[4632]: I0313 12:02:56.675459 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qttkd" event={"ID":"d15a7682-8687-4495-9ea5-bab97097930e","Type":"ContainerDied","Data":"28be9f7463878e855f3ddab509b5ad179e18babc64a9ca78c7b8ffe29d0be6e0"} Mar 13 12:02:56 crc kubenswrapper[4632]: I0313 12:02:56.675519 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qttkd" event={"ID":"d15a7682-8687-4495-9ea5-bab97097930e","Type":"ContainerStarted","Data":"ccb1372aabc48631479f228688c95abab44ec403b778ae2c4d4db0abf3ee9d7e"} Mar 13 12:02:58 crc kubenswrapper[4632]: I0313 12:02:58.699466 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qttkd" event={"ID":"d15a7682-8687-4495-9ea5-bab97097930e","Type":"ContainerStarted","Data":"0f02737dbf81fc5b27a98d7116776423f246a75c7aa8570b9234458aedbdf876"} Mar 13 12:03:05 crc kubenswrapper[4632]: I0313 12:03:05.769010 4632 generic.go:334] "Generic (PLEG): container finished" podID="d15a7682-8687-4495-9ea5-bab97097930e" containerID="0f02737dbf81fc5b27a98d7116776423f246a75c7aa8570b9234458aedbdf876" exitCode=0 Mar 13 12:03:05 crc kubenswrapper[4632]: I0313 12:03:05.769061 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qttkd" event={"ID":"d15a7682-8687-4495-9ea5-bab97097930e","Type":"ContainerDied","Data":"0f02737dbf81fc5b27a98d7116776423f246a75c7aa8570b9234458aedbdf876"} Mar 13 12:03:06 crc kubenswrapper[4632]: I0313 12:03:06.783505 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qttkd" event={"ID":"d15a7682-8687-4495-9ea5-bab97097930e","Type":"ContainerStarted","Data":"4dfa9a954b7ca6e925f87876683c3628cf9338c717e6ef370787714a55baee80"} Mar 13 12:03:06 crc kubenswrapper[4632]: I0313 12:03:06.811784 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qttkd" podStartSLOduration=2.316522534 podStartE2EDuration="11.811758497s" podCreationTimestamp="2026-03-13 12:02:55 +0000 UTC" firstStartedPulling="2026-03-13 12:02:56.681884153 +0000 UTC m=+7150.704414286" lastFinishedPulling="2026-03-13 12:03:06.177120116 +0000 UTC m=+7160.199650249" observedRunningTime="2026-03-13 12:03:06.803395366 +0000 UTC m=+7160.825925509" watchObservedRunningTime="2026-03-13 12:03:06.811758497 +0000 UTC m=+7160.834288630" Mar 13 12:03:10 crc kubenswrapper[4632]: I0313 12:03:10.461682 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:03:10 crc kubenswrapper[4632]: I0313 12:03:10.462142 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:03:10 crc kubenswrapper[4632]: I0313 12:03:10.462191 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 12:03:10 crc kubenswrapper[4632]: I0313 12:03:10.463004 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 12:03:10 crc kubenswrapper[4632]: I0313 12:03:10.463063 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" gracePeriod=600 Mar 13 12:03:10 crc kubenswrapper[4632]: E0313 12:03:10.610295 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:03:10 crc kubenswrapper[4632]: I0313 12:03:10.822637 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" exitCode=0 Mar 13 12:03:10 crc kubenswrapper[4632]: I0313 12:03:10.822684 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6"} Mar 13 12:03:10 crc kubenswrapper[4632]: I0313 12:03:10.822723 4632 scope.go:117] "RemoveContainer" containerID="f1337fb64ab38c0f489a591d3b3f173d13428642427113f1891b2f17a626304e" Mar 13 12:03:10 crc kubenswrapper[4632]: I0313 12:03:10.823420 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:03:10 crc kubenswrapper[4632]: E0313 12:03:10.823772 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:03:15 crc kubenswrapper[4632]: I0313 12:03:15.533286 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:03:15 crc kubenswrapper[4632]: I0313 12:03:15.533613 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:03:16 crc kubenswrapper[4632]: I0313 12:03:16.585201 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qttkd" podUID="d15a7682-8687-4495-9ea5-bab97097930e" containerName="registry-server" probeResult="failure" output=< Mar 13 12:03:16 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:03:16 crc kubenswrapper[4632]: > Mar 13 12:03:25 crc kubenswrapper[4632]: I0313 12:03:25.044871 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:03:25 crc kubenswrapper[4632]: E0313 12:03:25.045635 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:03:26 crc kubenswrapper[4632]: I0313 12:03:26.582756 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qttkd" podUID="d15a7682-8687-4495-9ea5-bab97097930e" containerName="registry-server" probeResult="failure" output=< Mar 13 12:03:26 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:03:26 crc kubenswrapper[4632]: > Mar 13 12:03:36 crc kubenswrapper[4632]: I0313 12:03:36.578770 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qttkd" podUID="d15a7682-8687-4495-9ea5-bab97097930e" containerName="registry-server" probeResult="failure" output=< Mar 13 12:03:36 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:03:36 crc kubenswrapper[4632]: > Mar 13 12:03:40 crc kubenswrapper[4632]: I0313 12:03:40.044784 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:03:40 crc kubenswrapper[4632]: E0313 12:03:40.045723 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:03:46 crc kubenswrapper[4632]: I0313 12:03:46.580823 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-qttkd" podUID="d15a7682-8687-4495-9ea5-bab97097930e" containerName="registry-server" probeResult="failure" output=< Mar 13 12:03:46 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:03:46 crc kubenswrapper[4632]: > Mar 13 12:03:53 crc kubenswrapper[4632]: I0313 12:03:53.045520 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:03:53 crc kubenswrapper[4632]: E0313 12:03:53.046639 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:03:55 crc kubenswrapper[4632]: I0313 12:03:55.599751 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:03:55 crc kubenswrapper[4632]: I0313 12:03:55.659220 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:03:56 crc kubenswrapper[4632]: I0313 12:03:56.376560 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qttkd"] Mar 13 12:03:57 crc kubenswrapper[4632]: I0313 12:03:57.270265 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qttkd" podUID="d15a7682-8687-4495-9ea5-bab97097930e" containerName="registry-server" containerID="cri-o://4dfa9a954b7ca6e925f87876683c3628cf9338c717e6ef370787714a55baee80" gracePeriod=2 Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.246531 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.300457 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qttkd" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.300459 4632 generic.go:334] "Generic (PLEG): container finished" podID="d15a7682-8687-4495-9ea5-bab97097930e" containerID="4dfa9a954b7ca6e925f87876683c3628cf9338c717e6ef370787714a55baee80" exitCode=0 Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.310070 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qttkd" event={"ID":"d15a7682-8687-4495-9ea5-bab97097930e","Type":"ContainerDied","Data":"4dfa9a954b7ca6e925f87876683c3628cf9338c717e6ef370787714a55baee80"} Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.310170 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qttkd" event={"ID":"d15a7682-8687-4495-9ea5-bab97097930e","Type":"ContainerDied","Data":"ccb1372aabc48631479f228688c95abab44ec403b778ae2c4d4db0abf3ee9d7e"} Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.310425 4632 scope.go:117] "RemoveContainer" containerID="4dfa9a954b7ca6e925f87876683c3628cf9338c717e6ef370787714a55baee80" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.334419 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d15a7682-8687-4495-9ea5-bab97097930e-utilities\") pod \"d15a7682-8687-4495-9ea5-bab97097930e\" (UID: \"d15a7682-8687-4495-9ea5-bab97097930e\") " Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.334576 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px4g4\" (UniqueName: \"kubernetes.io/projected/d15a7682-8687-4495-9ea5-bab97097930e-kube-api-access-px4g4\") pod \"d15a7682-8687-4495-9ea5-bab97097930e\" (UID: \"d15a7682-8687-4495-9ea5-bab97097930e\") " Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.334732 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d15a7682-8687-4495-9ea5-bab97097930e-catalog-content\") pod \"d15a7682-8687-4495-9ea5-bab97097930e\" (UID: \"d15a7682-8687-4495-9ea5-bab97097930e\") " Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.336383 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d15a7682-8687-4495-9ea5-bab97097930e-utilities" (OuterVolumeSpecName: "utilities") pod "d15a7682-8687-4495-9ea5-bab97097930e" (UID: "d15a7682-8687-4495-9ea5-bab97097930e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.341494 4632 scope.go:117] "RemoveContainer" containerID="0f02737dbf81fc5b27a98d7116776423f246a75c7aa8570b9234458aedbdf876" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.378866 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d15a7682-8687-4495-9ea5-bab97097930e-kube-api-access-px4g4" (OuterVolumeSpecName: "kube-api-access-px4g4") pod "d15a7682-8687-4495-9ea5-bab97097930e" (UID: "d15a7682-8687-4495-9ea5-bab97097930e"). InnerVolumeSpecName "kube-api-access-px4g4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.384244 4632 scope.go:117] "RemoveContainer" containerID="28be9f7463878e855f3ddab509b5ad179e18babc64a9ca78c7b8ffe29d0be6e0" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.438105 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d15a7682-8687-4495-9ea5-bab97097930e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.438146 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px4g4\" (UniqueName: \"kubernetes.io/projected/d15a7682-8687-4495-9ea5-bab97097930e-kube-api-access-px4g4\") on node \"crc\" DevicePath \"\"" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.471062 4632 scope.go:117] "RemoveContainer" containerID="4dfa9a954b7ca6e925f87876683c3628cf9338c717e6ef370787714a55baee80" Mar 13 12:03:58 crc kubenswrapper[4632]: E0313 12:03:58.472347 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dfa9a954b7ca6e925f87876683c3628cf9338c717e6ef370787714a55baee80\": container with ID starting with 4dfa9a954b7ca6e925f87876683c3628cf9338c717e6ef370787714a55baee80 not found: ID does not exist" containerID="4dfa9a954b7ca6e925f87876683c3628cf9338c717e6ef370787714a55baee80" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.472492 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dfa9a954b7ca6e925f87876683c3628cf9338c717e6ef370787714a55baee80"} err="failed to get container status \"4dfa9a954b7ca6e925f87876683c3628cf9338c717e6ef370787714a55baee80\": rpc error: code = NotFound desc = could not find container \"4dfa9a954b7ca6e925f87876683c3628cf9338c717e6ef370787714a55baee80\": container with ID starting with 4dfa9a954b7ca6e925f87876683c3628cf9338c717e6ef370787714a55baee80 not found: ID does not exist" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.472636 4632 scope.go:117] "RemoveContainer" containerID="0f02737dbf81fc5b27a98d7116776423f246a75c7aa8570b9234458aedbdf876" Mar 13 12:03:58 crc kubenswrapper[4632]: E0313 12:03:58.473318 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f02737dbf81fc5b27a98d7116776423f246a75c7aa8570b9234458aedbdf876\": container with ID starting with 0f02737dbf81fc5b27a98d7116776423f246a75c7aa8570b9234458aedbdf876 not found: ID does not exist" containerID="0f02737dbf81fc5b27a98d7116776423f246a75c7aa8570b9234458aedbdf876" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.473366 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f02737dbf81fc5b27a98d7116776423f246a75c7aa8570b9234458aedbdf876"} err="failed to get container status \"0f02737dbf81fc5b27a98d7116776423f246a75c7aa8570b9234458aedbdf876\": rpc error: code = NotFound desc = could not find container \"0f02737dbf81fc5b27a98d7116776423f246a75c7aa8570b9234458aedbdf876\": container with ID starting with 0f02737dbf81fc5b27a98d7116776423f246a75c7aa8570b9234458aedbdf876 not found: ID does not exist" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.473412 4632 scope.go:117] "RemoveContainer" containerID="28be9f7463878e855f3ddab509b5ad179e18babc64a9ca78c7b8ffe29d0be6e0" Mar 13 12:03:58 crc kubenswrapper[4632]: E0313 12:03:58.475431 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28be9f7463878e855f3ddab509b5ad179e18babc64a9ca78c7b8ffe29d0be6e0\": container with ID starting with 28be9f7463878e855f3ddab509b5ad179e18babc64a9ca78c7b8ffe29d0be6e0 not found: ID does not exist" containerID="28be9f7463878e855f3ddab509b5ad179e18babc64a9ca78c7b8ffe29d0be6e0" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.475603 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28be9f7463878e855f3ddab509b5ad179e18babc64a9ca78c7b8ffe29d0be6e0"} err="failed to get container status \"28be9f7463878e855f3ddab509b5ad179e18babc64a9ca78c7b8ffe29d0be6e0\": rpc error: code = NotFound desc = could not find container \"28be9f7463878e855f3ddab509b5ad179e18babc64a9ca78c7b8ffe29d0be6e0\": container with ID starting with 28be9f7463878e855f3ddab509b5ad179e18babc64a9ca78c7b8ffe29d0be6e0 not found: ID does not exist" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.506159 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d15a7682-8687-4495-9ea5-bab97097930e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d15a7682-8687-4495-9ea5-bab97097930e" (UID: "d15a7682-8687-4495-9ea5-bab97097930e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.539958 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d15a7682-8687-4495-9ea5-bab97097930e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.643746 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qttkd"] Mar 13 12:03:58 crc kubenswrapper[4632]: I0313 12:03:58.652574 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qttkd"] Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.059807 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d15a7682-8687-4495-9ea5-bab97097930e" path="/var/lib/kubelet/pods/d15a7682-8687-4495-9ea5-bab97097930e/volumes" Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.155015 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556724-thbvk"] Mar 13 12:04:00 crc kubenswrapper[4632]: E0313 12:04:00.155522 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15a7682-8687-4495-9ea5-bab97097930e" containerName="extract-content" Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.155547 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15a7682-8687-4495-9ea5-bab97097930e" containerName="extract-content" Mar 13 12:04:00 crc kubenswrapper[4632]: E0313 12:04:00.155583 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15a7682-8687-4495-9ea5-bab97097930e" containerName="extract-utilities" Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.155596 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15a7682-8687-4495-9ea5-bab97097930e" containerName="extract-utilities" Mar 13 12:04:00 crc kubenswrapper[4632]: E0313 12:04:00.155613 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d15a7682-8687-4495-9ea5-bab97097930e" containerName="registry-server" Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.155621 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d15a7682-8687-4495-9ea5-bab97097930e" containerName="registry-server" Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.155882 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="d15a7682-8687-4495-9ea5-bab97097930e" containerName="registry-server" Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.158455 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556724-thbvk" Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.167638 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556724-thbvk"] Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.174066 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.174549 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.176176 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.303504 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbbw5\" (UniqueName: \"kubernetes.io/projected/af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe-kube-api-access-rbbw5\") pod \"auto-csr-approver-29556724-thbvk\" (UID: \"af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe\") " pod="openshift-infra/auto-csr-approver-29556724-thbvk" Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.405273 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbbw5\" (UniqueName: \"kubernetes.io/projected/af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe-kube-api-access-rbbw5\") pod \"auto-csr-approver-29556724-thbvk\" (UID: \"af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe\") " pod="openshift-infra/auto-csr-approver-29556724-thbvk" Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.434717 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbbw5\" (UniqueName: \"kubernetes.io/projected/af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe-kube-api-access-rbbw5\") pod \"auto-csr-approver-29556724-thbvk\" (UID: \"af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe\") " pod="openshift-infra/auto-csr-approver-29556724-thbvk" Mar 13 12:04:00 crc kubenswrapper[4632]: I0313 12:04:00.480852 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556724-thbvk" Mar 13 12:04:01 crc kubenswrapper[4632]: I0313 12:04:01.033773 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556724-thbvk"] Mar 13 12:04:01 crc kubenswrapper[4632]: I0313 12:04:01.329767 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556724-thbvk" event={"ID":"af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe","Type":"ContainerStarted","Data":"2369d2de44ac16f350468ba3eab4c9d9e96546c32f4677ea6d5ace02630e3153"} Mar 13 12:04:03 crc kubenswrapper[4632]: I0313 12:04:03.349041 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556724-thbvk" event={"ID":"af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe","Type":"ContainerStarted","Data":"b2b91ac6b566e0c21758b3baa48d2497ca87c0016cb01ff16589e8a5fd981c2d"} Mar 13 12:04:03 crc kubenswrapper[4632]: I0313 12:04:03.365670 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556724-thbvk" podStartSLOduration=2.211404459 podStartE2EDuration="3.365649729s" podCreationTimestamp="2026-03-13 12:04:00 +0000 UTC" firstStartedPulling="2026-03-13 12:04:01.042835674 +0000 UTC m=+7215.065365807" lastFinishedPulling="2026-03-13 12:04:02.197080944 +0000 UTC m=+7216.219611077" observedRunningTime="2026-03-13 12:04:03.362891293 +0000 UTC m=+7217.385421426" watchObservedRunningTime="2026-03-13 12:04:03.365649729 +0000 UTC m=+7217.388179872" Mar 13 12:04:04 crc kubenswrapper[4632]: I0313 12:04:04.360521 4632 generic.go:334] "Generic (PLEG): container finished" podID="af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe" containerID="b2b91ac6b566e0c21758b3baa48d2497ca87c0016cb01ff16589e8a5fd981c2d" exitCode=0 Mar 13 12:04:04 crc kubenswrapper[4632]: I0313 12:04:04.360578 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556724-thbvk" event={"ID":"af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe","Type":"ContainerDied","Data":"b2b91ac6b566e0c21758b3baa48d2497ca87c0016cb01ff16589e8a5fd981c2d"} Mar 13 12:04:05 crc kubenswrapper[4632]: I0313 12:04:05.875808 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556724-thbvk" Mar 13 12:04:06 crc kubenswrapper[4632]: I0313 12:04:06.013769 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbbw5\" (UniqueName: \"kubernetes.io/projected/af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe-kube-api-access-rbbw5\") pod \"af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe\" (UID: \"af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe\") " Mar 13 12:04:06 crc kubenswrapper[4632]: I0313 12:04:06.027153 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe-kube-api-access-rbbw5" (OuterVolumeSpecName: "kube-api-access-rbbw5") pod "af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe" (UID: "af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe"). InnerVolumeSpecName "kube-api-access-rbbw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:04:06 crc kubenswrapper[4632]: I0313 12:04:06.115988 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbbw5\" (UniqueName: \"kubernetes.io/projected/af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe-kube-api-access-rbbw5\") on node \"crc\" DevicePath \"\"" Mar 13 12:04:06 crc kubenswrapper[4632]: I0313 12:04:06.380766 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556724-thbvk" event={"ID":"af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe","Type":"ContainerDied","Data":"2369d2de44ac16f350468ba3eab4c9d9e96546c32f4677ea6d5ace02630e3153"} Mar 13 12:04:06 crc kubenswrapper[4632]: I0313 12:04:06.380808 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2369d2de44ac16f350468ba3eab4c9d9e96546c32f4677ea6d5ace02630e3153" Mar 13 12:04:06 crc kubenswrapper[4632]: I0313 12:04:06.380805 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556724-thbvk" Mar 13 12:04:06 crc kubenswrapper[4632]: I0313 12:04:06.472705 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556718-xxtsr"] Mar 13 12:04:06 crc kubenswrapper[4632]: I0313 12:04:06.481290 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556718-xxtsr"] Mar 13 12:04:08 crc kubenswrapper[4632]: I0313 12:04:08.049820 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:04:08 crc kubenswrapper[4632]: E0313 12:04:08.050519 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:04:08 crc kubenswrapper[4632]: I0313 12:04:08.060368 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="beb533c7-a735-47fa-b5fa-67b1bcba9787" path="/var/lib/kubelet/pods/beb533c7-a735-47fa-b5fa-67b1bcba9787/volumes" Mar 13 12:04:22 crc kubenswrapper[4632]: I0313 12:04:22.044921 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:04:22 crc kubenswrapper[4632]: E0313 12:04:22.045845 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:04:24 crc kubenswrapper[4632]: I0313 12:04:24.415912 4632 scope.go:117] "RemoveContainer" containerID="de5a0b9383a1bdabde0e1290cb2d2e2341dbc3e19f3a7e552782ac9f0501a7ce" Mar 13 12:04:37 crc kubenswrapper[4632]: I0313 12:04:37.045173 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:04:37 crc kubenswrapper[4632]: E0313 12:04:37.046359 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:04:51 crc kubenswrapper[4632]: I0313 12:04:51.044397 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:04:51 crc kubenswrapper[4632]: E0313 12:04:51.045149 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:05:06 crc kubenswrapper[4632]: I0313 12:05:06.044096 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:05:06 crc kubenswrapper[4632]: E0313 12:05:06.046173 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:05:20 crc kubenswrapper[4632]: I0313 12:05:20.043880 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:05:20 crc kubenswrapper[4632]: E0313 12:05:20.044596 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:05:35 crc kubenswrapper[4632]: I0313 12:05:35.044465 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:05:35 crc kubenswrapper[4632]: E0313 12:05:35.045660 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:05:48 crc kubenswrapper[4632]: I0313 12:05:48.050753 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:05:48 crc kubenswrapper[4632]: E0313 12:05:48.051634 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:06:00 crc kubenswrapper[4632]: I0313 12:06:00.044825 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:06:00 crc kubenswrapper[4632]: E0313 12:06:00.045561 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:06:00 crc kubenswrapper[4632]: I0313 12:06:00.163489 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556726-pjf8m"] Mar 13 12:06:00 crc kubenswrapper[4632]: E0313 12:06:00.164183 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe" containerName="oc" Mar 13 12:06:00 crc kubenswrapper[4632]: I0313 12:06:00.164212 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe" containerName="oc" Mar 13 12:06:00 crc kubenswrapper[4632]: I0313 12:06:00.164485 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe" containerName="oc" Mar 13 12:06:00 crc kubenswrapper[4632]: I0313 12:06:00.166237 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556726-pjf8m" Mar 13 12:06:00 crc kubenswrapper[4632]: I0313 12:06:00.168516 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:06:00 crc kubenswrapper[4632]: I0313 12:06:00.169161 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:06:00 crc kubenswrapper[4632]: I0313 12:06:00.174270 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:06:00 crc kubenswrapper[4632]: I0313 12:06:00.178105 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556726-pjf8m"] Mar 13 12:06:00 crc kubenswrapper[4632]: I0313 12:06:00.301740 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whhpb\" (UniqueName: \"kubernetes.io/projected/2bb028cb-1d8f-4f09-8c73-135a6ce2bb46-kube-api-access-whhpb\") pod \"auto-csr-approver-29556726-pjf8m\" (UID: \"2bb028cb-1d8f-4f09-8c73-135a6ce2bb46\") " pod="openshift-infra/auto-csr-approver-29556726-pjf8m" Mar 13 12:06:00 crc kubenswrapper[4632]: I0313 12:06:00.404479 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whhpb\" (UniqueName: \"kubernetes.io/projected/2bb028cb-1d8f-4f09-8c73-135a6ce2bb46-kube-api-access-whhpb\") pod \"auto-csr-approver-29556726-pjf8m\" (UID: \"2bb028cb-1d8f-4f09-8c73-135a6ce2bb46\") " pod="openshift-infra/auto-csr-approver-29556726-pjf8m" Mar 13 12:06:00 crc kubenswrapper[4632]: I0313 12:06:00.429122 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whhpb\" (UniqueName: \"kubernetes.io/projected/2bb028cb-1d8f-4f09-8c73-135a6ce2bb46-kube-api-access-whhpb\") pod \"auto-csr-approver-29556726-pjf8m\" (UID: \"2bb028cb-1d8f-4f09-8c73-135a6ce2bb46\") " pod="openshift-infra/auto-csr-approver-29556726-pjf8m" Mar 13 12:06:00 crc kubenswrapper[4632]: I0313 12:06:00.499325 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556726-pjf8m" Mar 13 12:06:01 crc kubenswrapper[4632]: I0313 12:06:01.092129 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556726-pjf8m"] Mar 13 12:06:02 crc kubenswrapper[4632]: I0313 12:06:02.013616 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556726-pjf8m" event={"ID":"2bb028cb-1d8f-4f09-8c73-135a6ce2bb46","Type":"ContainerStarted","Data":"92f020c422f37715af0272fc7b122cdd5f09f8d4eb1ed9ce0772749e38c83f9d"} Mar 13 12:06:03 crc kubenswrapper[4632]: I0313 12:06:03.026058 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556726-pjf8m" event={"ID":"2bb028cb-1d8f-4f09-8c73-135a6ce2bb46","Type":"ContainerStarted","Data":"d0bcc7d380da406a96a84a48eb87ce7d69302220aa1982823e25de22c01f77e0"} Mar 13 12:06:03 crc kubenswrapper[4632]: I0313 12:06:03.046285 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556726-pjf8m" podStartSLOduration=2.209526589 podStartE2EDuration="3.046262825s" podCreationTimestamp="2026-03-13 12:06:00 +0000 UTC" firstStartedPulling="2026-03-13 12:06:01.102623216 +0000 UTC m=+7335.125153349" lastFinishedPulling="2026-03-13 12:06:01.939359452 +0000 UTC m=+7335.961889585" observedRunningTime="2026-03-13 12:06:03.040577439 +0000 UTC m=+7337.063107592" watchObservedRunningTime="2026-03-13 12:06:03.046262825 +0000 UTC m=+7337.068792958" Mar 13 12:06:04 crc kubenswrapper[4632]: I0313 12:06:04.038342 4632 generic.go:334] "Generic (PLEG): container finished" podID="2bb028cb-1d8f-4f09-8c73-135a6ce2bb46" containerID="d0bcc7d380da406a96a84a48eb87ce7d69302220aa1982823e25de22c01f77e0" exitCode=0 Mar 13 12:06:04 crc kubenswrapper[4632]: I0313 12:06:04.038460 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556726-pjf8m" event={"ID":"2bb028cb-1d8f-4f09-8c73-135a6ce2bb46","Type":"ContainerDied","Data":"d0bcc7d380da406a96a84a48eb87ce7d69302220aa1982823e25de22c01f77e0"} Mar 13 12:06:05 crc kubenswrapper[4632]: I0313 12:06:05.564510 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556726-pjf8m" Mar 13 12:06:05 crc kubenswrapper[4632]: I0313 12:06:05.704702 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whhpb\" (UniqueName: \"kubernetes.io/projected/2bb028cb-1d8f-4f09-8c73-135a6ce2bb46-kube-api-access-whhpb\") pod \"2bb028cb-1d8f-4f09-8c73-135a6ce2bb46\" (UID: \"2bb028cb-1d8f-4f09-8c73-135a6ce2bb46\") " Mar 13 12:06:05 crc kubenswrapper[4632]: I0313 12:06:05.723330 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2bb028cb-1d8f-4f09-8c73-135a6ce2bb46-kube-api-access-whhpb" (OuterVolumeSpecName: "kube-api-access-whhpb") pod "2bb028cb-1d8f-4f09-8c73-135a6ce2bb46" (UID: "2bb028cb-1d8f-4f09-8c73-135a6ce2bb46"). InnerVolumeSpecName "kube-api-access-whhpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:06:05 crc kubenswrapper[4632]: I0313 12:06:05.807211 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whhpb\" (UniqueName: \"kubernetes.io/projected/2bb028cb-1d8f-4f09-8c73-135a6ce2bb46-kube-api-access-whhpb\") on node \"crc\" DevicePath \"\"" Mar 13 12:06:06 crc kubenswrapper[4632]: I0313 12:06:06.068537 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556726-pjf8m" event={"ID":"2bb028cb-1d8f-4f09-8c73-135a6ce2bb46","Type":"ContainerDied","Data":"92f020c422f37715af0272fc7b122cdd5f09f8d4eb1ed9ce0772749e38c83f9d"} Mar 13 12:06:06 crc kubenswrapper[4632]: I0313 12:06:06.068584 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92f020c422f37715af0272fc7b122cdd5f09f8d4eb1ed9ce0772749e38c83f9d" Mar 13 12:06:06 crc kubenswrapper[4632]: I0313 12:06:06.068654 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556726-pjf8m" Mar 13 12:06:06 crc kubenswrapper[4632]: I0313 12:06:06.128667 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556720-2985g"] Mar 13 12:06:06 crc kubenswrapper[4632]: I0313 12:06:06.140723 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556720-2985g"] Mar 13 12:06:08 crc kubenswrapper[4632]: I0313 12:06:08.062510 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="614f8dd7-8a57-4b22-b741-63c3ed563216" path="/var/lib/kubelet/pods/614f8dd7-8a57-4b22-b741-63c3ed563216/volumes" Mar 13 12:06:13 crc kubenswrapper[4632]: I0313 12:06:13.044250 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:06:13 crc kubenswrapper[4632]: E0313 12:06:13.045086 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:06:24 crc kubenswrapper[4632]: I0313 12:06:24.563604 4632 scope.go:117] "RemoveContainer" containerID="d1793c511542a0a35aa5afc5e36e94033f5a38dca84400ac841ea4e47dc426f7" Mar 13 12:06:25 crc kubenswrapper[4632]: I0313 12:06:25.044328 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:06:25 crc kubenswrapper[4632]: E0313 12:06:25.044640 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:06:37 crc kubenswrapper[4632]: I0313 12:06:37.044245 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:06:37 crc kubenswrapper[4632]: E0313 12:06:37.045671 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:06:52 crc kubenswrapper[4632]: I0313 12:06:52.045123 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:06:52 crc kubenswrapper[4632]: E0313 12:06:52.046211 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:07:04 crc kubenswrapper[4632]: I0313 12:07:04.045027 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:07:04 crc kubenswrapper[4632]: E0313 12:07:04.045924 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:07:15 crc kubenswrapper[4632]: I0313 12:07:15.045459 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:07:15 crc kubenswrapper[4632]: E0313 12:07:15.046433 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:07:28 crc kubenswrapper[4632]: I0313 12:07:28.055012 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:07:28 crc kubenswrapper[4632]: E0313 12:07:28.055773 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:07:40 crc kubenswrapper[4632]: I0313 12:07:40.045150 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:07:40 crc kubenswrapper[4632]: E0313 12:07:40.046055 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:07:51 crc kubenswrapper[4632]: I0313 12:07:51.044441 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:07:51 crc kubenswrapper[4632]: E0313 12:07:51.045389 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:08:00 crc kubenswrapper[4632]: I0313 12:08:00.146305 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556728-lnbz4"] Mar 13 12:08:00 crc kubenswrapper[4632]: E0313 12:08:00.147115 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2bb028cb-1d8f-4f09-8c73-135a6ce2bb46" containerName="oc" Mar 13 12:08:00 crc kubenswrapper[4632]: I0313 12:08:00.147131 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="2bb028cb-1d8f-4f09-8c73-135a6ce2bb46" containerName="oc" Mar 13 12:08:00 crc kubenswrapper[4632]: I0313 12:08:00.147679 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="2bb028cb-1d8f-4f09-8c73-135a6ce2bb46" containerName="oc" Mar 13 12:08:00 crc kubenswrapper[4632]: I0313 12:08:00.148347 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556728-lnbz4" Mar 13 12:08:00 crc kubenswrapper[4632]: I0313 12:08:00.150872 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:08:00 crc kubenswrapper[4632]: I0313 12:08:00.151074 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:08:00 crc kubenswrapper[4632]: I0313 12:08:00.151153 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:08:00 crc kubenswrapper[4632]: I0313 12:08:00.167562 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556728-lnbz4"] Mar 13 12:08:00 crc kubenswrapper[4632]: I0313 12:08:00.242102 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpwzp\" (UniqueName: \"kubernetes.io/projected/4980ff16-68a2-4b11-83d2-9d8ad1fa105c-kube-api-access-wpwzp\") pod \"auto-csr-approver-29556728-lnbz4\" (UID: \"4980ff16-68a2-4b11-83d2-9d8ad1fa105c\") " pod="openshift-infra/auto-csr-approver-29556728-lnbz4" Mar 13 12:08:00 crc kubenswrapper[4632]: I0313 12:08:00.343677 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpwzp\" (UniqueName: \"kubernetes.io/projected/4980ff16-68a2-4b11-83d2-9d8ad1fa105c-kube-api-access-wpwzp\") pod \"auto-csr-approver-29556728-lnbz4\" (UID: \"4980ff16-68a2-4b11-83d2-9d8ad1fa105c\") " pod="openshift-infra/auto-csr-approver-29556728-lnbz4" Mar 13 12:08:00 crc kubenswrapper[4632]: I0313 12:08:00.366437 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpwzp\" (UniqueName: \"kubernetes.io/projected/4980ff16-68a2-4b11-83d2-9d8ad1fa105c-kube-api-access-wpwzp\") pod \"auto-csr-approver-29556728-lnbz4\" (UID: \"4980ff16-68a2-4b11-83d2-9d8ad1fa105c\") " pod="openshift-infra/auto-csr-approver-29556728-lnbz4" Mar 13 12:08:00 crc kubenswrapper[4632]: I0313 12:08:00.468427 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556728-lnbz4" Mar 13 12:08:01 crc kubenswrapper[4632]: I0313 12:08:01.044938 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556728-lnbz4"] Mar 13 12:08:01 crc kubenswrapper[4632]: I0313 12:08:01.055257 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 12:08:01 crc kubenswrapper[4632]: I0313 12:08:01.370230 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556728-lnbz4" event={"ID":"4980ff16-68a2-4b11-83d2-9d8ad1fa105c","Type":"ContainerStarted","Data":"28c418bff6764d978c7a7a37a355de25591b1f316732c8050beecc8e6c5b3e72"} Mar 13 12:08:02 crc kubenswrapper[4632]: I0313 12:08:02.382885 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556728-lnbz4" event={"ID":"4980ff16-68a2-4b11-83d2-9d8ad1fa105c","Type":"ContainerStarted","Data":"f67ad12914918a2b5742053c25b75fbf60ba190fadc812d9c25ad10140a8556c"} Mar 13 12:08:02 crc kubenswrapper[4632]: I0313 12:08:02.403859 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556728-lnbz4" podStartSLOduration=1.462827393 podStartE2EDuration="2.403840713s" podCreationTimestamp="2026-03-13 12:08:00 +0000 UTC" firstStartedPulling="2026-03-13 12:08:01.053303039 +0000 UTC m=+7455.075833172" lastFinishedPulling="2026-03-13 12:08:01.994316359 +0000 UTC m=+7456.016846492" observedRunningTime="2026-03-13 12:08:02.397774708 +0000 UTC m=+7456.420304841" watchObservedRunningTime="2026-03-13 12:08:02.403840713 +0000 UTC m=+7456.426370846" Mar 13 12:08:03 crc kubenswrapper[4632]: I0313 12:08:03.044612 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:08:03 crc kubenswrapper[4632]: E0313 12:08:03.044968 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:08:04 crc kubenswrapper[4632]: I0313 12:08:04.403844 4632 generic.go:334] "Generic (PLEG): container finished" podID="4980ff16-68a2-4b11-83d2-9d8ad1fa105c" containerID="f67ad12914918a2b5742053c25b75fbf60ba190fadc812d9c25ad10140a8556c" exitCode=0 Mar 13 12:08:04 crc kubenswrapper[4632]: I0313 12:08:04.404060 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556728-lnbz4" event={"ID":"4980ff16-68a2-4b11-83d2-9d8ad1fa105c","Type":"ContainerDied","Data":"f67ad12914918a2b5742053c25b75fbf60ba190fadc812d9c25ad10140a8556c"} Mar 13 12:08:05 crc kubenswrapper[4632]: I0313 12:08:05.954105 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556728-lnbz4" Mar 13 12:08:06 crc kubenswrapper[4632]: I0313 12:08:06.116870 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpwzp\" (UniqueName: \"kubernetes.io/projected/4980ff16-68a2-4b11-83d2-9d8ad1fa105c-kube-api-access-wpwzp\") pod \"4980ff16-68a2-4b11-83d2-9d8ad1fa105c\" (UID: \"4980ff16-68a2-4b11-83d2-9d8ad1fa105c\") " Mar 13 12:08:06 crc kubenswrapper[4632]: I0313 12:08:06.127131 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4980ff16-68a2-4b11-83d2-9d8ad1fa105c-kube-api-access-wpwzp" (OuterVolumeSpecName: "kube-api-access-wpwzp") pod "4980ff16-68a2-4b11-83d2-9d8ad1fa105c" (UID: "4980ff16-68a2-4b11-83d2-9d8ad1fa105c"). InnerVolumeSpecName "kube-api-access-wpwzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:08:06 crc kubenswrapper[4632]: I0313 12:08:06.219277 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpwzp\" (UniqueName: \"kubernetes.io/projected/4980ff16-68a2-4b11-83d2-9d8ad1fa105c-kube-api-access-wpwzp\") on node \"crc\" DevicePath \"\"" Mar 13 12:08:06 crc kubenswrapper[4632]: I0313 12:08:06.426340 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556728-lnbz4" event={"ID":"4980ff16-68a2-4b11-83d2-9d8ad1fa105c","Type":"ContainerDied","Data":"28c418bff6764d978c7a7a37a355de25591b1f316732c8050beecc8e6c5b3e72"} Mar 13 12:08:06 crc kubenswrapper[4632]: I0313 12:08:06.426391 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556728-lnbz4" Mar 13 12:08:06 crc kubenswrapper[4632]: I0313 12:08:06.426400 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="28c418bff6764d978c7a7a37a355de25591b1f316732c8050beecc8e6c5b3e72" Mar 13 12:08:06 crc kubenswrapper[4632]: I0313 12:08:06.509714 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556722-r4pp4"] Mar 13 12:08:06 crc kubenswrapper[4632]: I0313 12:08:06.521377 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556722-r4pp4"] Mar 13 12:08:08 crc kubenswrapper[4632]: I0313 12:08:08.061700 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ae03001-344b-4e5e-baf2-c8171109eb1a" path="/var/lib/kubelet/pods/5ae03001-344b-4e5e-baf2-c8171109eb1a/volumes" Mar 13 12:08:16 crc kubenswrapper[4632]: I0313 12:08:16.047537 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:08:16 crc kubenswrapper[4632]: I0313 12:08:16.525957 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"ef6a755da94d8b26aaa61b1a356ec9030e87ec1440f6bdf1f6abec8411efbdd9"} Mar 13 12:08:24 crc kubenswrapper[4632]: I0313 12:08:24.705326 4632 scope.go:117] "RemoveContainer" containerID="45cc8231b2d3d1ca8cec2f7a7da9147a3a370632b65df6ead7d138b8c0f615b3" Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.321089 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-9c4xw"] Mar 13 12:09:08 crc kubenswrapper[4632]: E0313 12:09:08.323646 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4980ff16-68a2-4b11-83d2-9d8ad1fa105c" containerName="oc" Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.323678 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="4980ff16-68a2-4b11-83d2-9d8ad1fa105c" containerName="oc" Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.324811 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="4980ff16-68a2-4b11-83d2-9d8ad1fa105c" containerName="oc" Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.330258 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.373824 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9c4xw"] Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.431151 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96a9f61e-825e-4999-a5e5-931111334a3c-utilities\") pod \"redhat-marketplace-9c4xw\" (UID: \"96a9f61e-825e-4999-a5e5-931111334a3c\") " pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.431302 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96a9f61e-825e-4999-a5e5-931111334a3c-catalog-content\") pod \"redhat-marketplace-9c4xw\" (UID: \"96a9f61e-825e-4999-a5e5-931111334a3c\") " pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.431342 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5mc2\" (UniqueName: \"kubernetes.io/projected/96a9f61e-825e-4999-a5e5-931111334a3c-kube-api-access-x5mc2\") pod \"redhat-marketplace-9c4xw\" (UID: \"96a9f61e-825e-4999-a5e5-931111334a3c\") " pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.532972 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5mc2\" (UniqueName: \"kubernetes.io/projected/96a9f61e-825e-4999-a5e5-931111334a3c-kube-api-access-x5mc2\") pod \"redhat-marketplace-9c4xw\" (UID: \"96a9f61e-825e-4999-a5e5-931111334a3c\") " pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.533196 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96a9f61e-825e-4999-a5e5-931111334a3c-utilities\") pod \"redhat-marketplace-9c4xw\" (UID: \"96a9f61e-825e-4999-a5e5-931111334a3c\") " pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.533292 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96a9f61e-825e-4999-a5e5-931111334a3c-catalog-content\") pod \"redhat-marketplace-9c4xw\" (UID: \"96a9f61e-825e-4999-a5e5-931111334a3c\") " pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.533812 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96a9f61e-825e-4999-a5e5-931111334a3c-utilities\") pod \"redhat-marketplace-9c4xw\" (UID: \"96a9f61e-825e-4999-a5e5-931111334a3c\") " pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.533859 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96a9f61e-825e-4999-a5e5-931111334a3c-catalog-content\") pod \"redhat-marketplace-9c4xw\" (UID: \"96a9f61e-825e-4999-a5e5-931111334a3c\") " pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.561829 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5mc2\" (UniqueName: \"kubernetes.io/projected/96a9f61e-825e-4999-a5e5-931111334a3c-kube-api-access-x5mc2\") pod \"redhat-marketplace-9c4xw\" (UID: \"96a9f61e-825e-4999-a5e5-931111334a3c\") " pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:08 crc kubenswrapper[4632]: I0313 12:09:08.674571 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:09 crc kubenswrapper[4632]: I0313 12:09:09.241015 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-9c4xw"] Mar 13 12:09:10 crc kubenswrapper[4632]: I0313 12:09:10.079404 4632 generic.go:334] "Generic (PLEG): container finished" podID="96a9f61e-825e-4999-a5e5-931111334a3c" containerID="fef21df0ca08e5753cfe1eb56acc3f885fe3e8067984bbd0d17ccfaaae953502" exitCode=0 Mar 13 12:09:10 crc kubenswrapper[4632]: I0313 12:09:10.079525 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9c4xw" event={"ID":"96a9f61e-825e-4999-a5e5-931111334a3c","Type":"ContainerDied","Data":"fef21df0ca08e5753cfe1eb56acc3f885fe3e8067984bbd0d17ccfaaae953502"} Mar 13 12:09:10 crc kubenswrapper[4632]: I0313 12:09:10.079934 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9c4xw" event={"ID":"96a9f61e-825e-4999-a5e5-931111334a3c","Type":"ContainerStarted","Data":"1eef9aa64ac81546604be8de930dc45934789d3cb4d0dc2fb31c3f6608753b84"} Mar 13 12:09:11 crc kubenswrapper[4632]: I0313 12:09:11.092676 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9c4xw" event={"ID":"96a9f61e-825e-4999-a5e5-931111334a3c","Type":"ContainerStarted","Data":"e0fecd98bfa39d9c2c53d83d664bc034530c6ba45b3a9a1361aa2d2fc97a532a"} Mar 13 12:09:13 crc kubenswrapper[4632]: I0313 12:09:13.110243 4632 generic.go:334] "Generic (PLEG): container finished" podID="96a9f61e-825e-4999-a5e5-931111334a3c" containerID="e0fecd98bfa39d9c2c53d83d664bc034530c6ba45b3a9a1361aa2d2fc97a532a" exitCode=0 Mar 13 12:09:13 crc kubenswrapper[4632]: I0313 12:09:13.110439 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9c4xw" event={"ID":"96a9f61e-825e-4999-a5e5-931111334a3c","Type":"ContainerDied","Data":"e0fecd98bfa39d9c2c53d83d664bc034530c6ba45b3a9a1361aa2d2fc97a532a"} Mar 13 12:09:14 crc kubenswrapper[4632]: I0313 12:09:14.122449 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9c4xw" event={"ID":"96a9f61e-825e-4999-a5e5-931111334a3c","Type":"ContainerStarted","Data":"a31594fca63f55ea4e7094a820be7740427ea837bc2d5d54f827d6587be7f8ef"} Mar 13 12:09:14 crc kubenswrapper[4632]: I0313 12:09:14.149447 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-9c4xw" podStartSLOduration=2.612267403 podStartE2EDuration="6.149415943s" podCreationTimestamp="2026-03-13 12:09:08 +0000 UTC" firstStartedPulling="2026-03-13 12:09:10.081603659 +0000 UTC m=+7524.104133792" lastFinishedPulling="2026-03-13 12:09:13.618752199 +0000 UTC m=+7527.641282332" observedRunningTime="2026-03-13 12:09:14.14388219 +0000 UTC m=+7528.166412373" watchObservedRunningTime="2026-03-13 12:09:14.149415943 +0000 UTC m=+7528.171946076" Mar 13 12:09:18 crc kubenswrapper[4632]: I0313 12:09:18.675865 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:18 crc kubenswrapper[4632]: I0313 12:09:18.676671 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:19 crc kubenswrapper[4632]: I0313 12:09:19.727931 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-9c4xw" podUID="96a9f61e-825e-4999-a5e5-931111334a3c" containerName="registry-server" probeResult="failure" output=< Mar 13 12:09:19 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:09:19 crc kubenswrapper[4632]: > Mar 13 12:09:28 crc kubenswrapper[4632]: I0313 12:09:28.730312 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:28 crc kubenswrapper[4632]: I0313 12:09:28.789011 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:28 crc kubenswrapper[4632]: I0313 12:09:28.972755 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9c4xw"] Mar 13 12:09:30 crc kubenswrapper[4632]: I0313 12:09:30.258832 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-9c4xw" podUID="96a9f61e-825e-4999-a5e5-931111334a3c" containerName="registry-server" containerID="cri-o://a31594fca63f55ea4e7094a820be7740427ea837bc2d5d54f827d6587be7f8ef" gracePeriod=2 Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.148833 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.181347 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96a9f61e-825e-4999-a5e5-931111334a3c-utilities\") pod \"96a9f61e-825e-4999-a5e5-931111334a3c\" (UID: \"96a9f61e-825e-4999-a5e5-931111334a3c\") " Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.181440 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5mc2\" (UniqueName: \"kubernetes.io/projected/96a9f61e-825e-4999-a5e5-931111334a3c-kube-api-access-x5mc2\") pod \"96a9f61e-825e-4999-a5e5-931111334a3c\" (UID: \"96a9f61e-825e-4999-a5e5-931111334a3c\") " Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.181462 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96a9f61e-825e-4999-a5e5-931111334a3c-catalog-content\") pod \"96a9f61e-825e-4999-a5e5-931111334a3c\" (UID: \"96a9f61e-825e-4999-a5e5-931111334a3c\") " Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.182650 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96a9f61e-825e-4999-a5e5-931111334a3c-utilities" (OuterVolumeSpecName: "utilities") pod "96a9f61e-825e-4999-a5e5-931111334a3c" (UID: "96a9f61e-825e-4999-a5e5-931111334a3c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.216740 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96a9f61e-825e-4999-a5e5-931111334a3c-kube-api-access-x5mc2" (OuterVolumeSpecName: "kube-api-access-x5mc2") pod "96a9f61e-825e-4999-a5e5-931111334a3c" (UID: "96a9f61e-825e-4999-a5e5-931111334a3c"). InnerVolumeSpecName "kube-api-access-x5mc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.229136 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/96a9f61e-825e-4999-a5e5-931111334a3c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "96a9f61e-825e-4999-a5e5-931111334a3c" (UID: "96a9f61e-825e-4999-a5e5-931111334a3c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.268329 4632 generic.go:334] "Generic (PLEG): container finished" podID="96a9f61e-825e-4999-a5e5-931111334a3c" containerID="a31594fca63f55ea4e7094a820be7740427ea837bc2d5d54f827d6587be7f8ef" exitCode=0 Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.268408 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-9c4xw" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.269306 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9c4xw" event={"ID":"96a9f61e-825e-4999-a5e5-931111334a3c","Type":"ContainerDied","Data":"a31594fca63f55ea4e7094a820be7740427ea837bc2d5d54f827d6587be7f8ef"} Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.269403 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-9c4xw" event={"ID":"96a9f61e-825e-4999-a5e5-931111334a3c","Type":"ContainerDied","Data":"1eef9aa64ac81546604be8de930dc45934789d3cb4d0dc2fb31c3f6608753b84"} Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.269431 4632 scope.go:117] "RemoveContainer" containerID="a31594fca63f55ea4e7094a820be7740427ea837bc2d5d54f827d6587be7f8ef" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.284591 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/96a9f61e-825e-4999-a5e5-931111334a3c-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.284628 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5mc2\" (UniqueName: \"kubernetes.io/projected/96a9f61e-825e-4999-a5e5-931111334a3c-kube-api-access-x5mc2\") on node \"crc\" DevicePath \"\"" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.284637 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/96a9f61e-825e-4999-a5e5-931111334a3c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.325090 4632 scope.go:117] "RemoveContainer" containerID="e0fecd98bfa39d9c2c53d83d664bc034530c6ba45b3a9a1361aa2d2fc97a532a" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.383558 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-9c4xw"] Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.404622 4632 scope.go:117] "RemoveContainer" containerID="fef21df0ca08e5753cfe1eb56acc3f885fe3e8067984bbd0d17ccfaaae953502" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.420301 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-9c4xw"] Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.434190 4632 scope.go:117] "RemoveContainer" containerID="a31594fca63f55ea4e7094a820be7740427ea837bc2d5d54f827d6587be7f8ef" Mar 13 12:09:31 crc kubenswrapper[4632]: E0313 12:09:31.437597 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a31594fca63f55ea4e7094a820be7740427ea837bc2d5d54f827d6587be7f8ef\": container with ID starting with a31594fca63f55ea4e7094a820be7740427ea837bc2d5d54f827d6587be7f8ef not found: ID does not exist" containerID="a31594fca63f55ea4e7094a820be7740427ea837bc2d5d54f827d6587be7f8ef" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.437638 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a31594fca63f55ea4e7094a820be7740427ea837bc2d5d54f827d6587be7f8ef"} err="failed to get container status \"a31594fca63f55ea4e7094a820be7740427ea837bc2d5d54f827d6587be7f8ef\": rpc error: code = NotFound desc = could not find container \"a31594fca63f55ea4e7094a820be7740427ea837bc2d5d54f827d6587be7f8ef\": container with ID starting with a31594fca63f55ea4e7094a820be7740427ea837bc2d5d54f827d6587be7f8ef not found: ID does not exist" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.437659 4632 scope.go:117] "RemoveContainer" containerID="e0fecd98bfa39d9c2c53d83d664bc034530c6ba45b3a9a1361aa2d2fc97a532a" Mar 13 12:09:31 crc kubenswrapper[4632]: E0313 12:09:31.437876 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0fecd98bfa39d9c2c53d83d664bc034530c6ba45b3a9a1361aa2d2fc97a532a\": container with ID starting with e0fecd98bfa39d9c2c53d83d664bc034530c6ba45b3a9a1361aa2d2fc97a532a not found: ID does not exist" containerID="e0fecd98bfa39d9c2c53d83d664bc034530c6ba45b3a9a1361aa2d2fc97a532a" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.437963 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0fecd98bfa39d9c2c53d83d664bc034530c6ba45b3a9a1361aa2d2fc97a532a"} err="failed to get container status \"e0fecd98bfa39d9c2c53d83d664bc034530c6ba45b3a9a1361aa2d2fc97a532a\": rpc error: code = NotFound desc = could not find container \"e0fecd98bfa39d9c2c53d83d664bc034530c6ba45b3a9a1361aa2d2fc97a532a\": container with ID starting with e0fecd98bfa39d9c2c53d83d664bc034530c6ba45b3a9a1361aa2d2fc97a532a not found: ID does not exist" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.437979 4632 scope.go:117] "RemoveContainer" containerID="fef21df0ca08e5753cfe1eb56acc3f885fe3e8067984bbd0d17ccfaaae953502" Mar 13 12:09:31 crc kubenswrapper[4632]: E0313 12:09:31.438235 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fef21df0ca08e5753cfe1eb56acc3f885fe3e8067984bbd0d17ccfaaae953502\": container with ID starting with fef21df0ca08e5753cfe1eb56acc3f885fe3e8067984bbd0d17ccfaaae953502 not found: ID does not exist" containerID="fef21df0ca08e5753cfe1eb56acc3f885fe3e8067984bbd0d17ccfaaae953502" Mar 13 12:09:31 crc kubenswrapper[4632]: I0313 12:09:31.438260 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fef21df0ca08e5753cfe1eb56acc3f885fe3e8067984bbd0d17ccfaaae953502"} err="failed to get container status \"fef21df0ca08e5753cfe1eb56acc3f885fe3e8067984bbd0d17ccfaaae953502\": rpc error: code = NotFound desc = could not find container \"fef21df0ca08e5753cfe1eb56acc3f885fe3e8067984bbd0d17ccfaaae953502\": container with ID starting with fef21df0ca08e5753cfe1eb56acc3f885fe3e8067984bbd0d17ccfaaae953502 not found: ID does not exist" Mar 13 12:09:32 crc kubenswrapper[4632]: I0313 12:09:32.055091 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96a9f61e-825e-4999-a5e5-931111334a3c" path="/var/lib/kubelet/pods/96a9f61e-825e-4999-a5e5-931111334a3c/volumes" Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.143576 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556730-txd8w"] Mar 13 12:10:00 crc kubenswrapper[4632]: E0313 12:10:00.144423 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a9f61e-825e-4999-a5e5-931111334a3c" containerName="extract-utilities" Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.144438 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a9f61e-825e-4999-a5e5-931111334a3c" containerName="extract-utilities" Mar 13 12:10:00 crc kubenswrapper[4632]: E0313 12:10:00.144472 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a9f61e-825e-4999-a5e5-931111334a3c" containerName="registry-server" Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.144479 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a9f61e-825e-4999-a5e5-931111334a3c" containerName="registry-server" Mar 13 12:10:00 crc kubenswrapper[4632]: E0313 12:10:00.144504 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="96a9f61e-825e-4999-a5e5-931111334a3c" containerName="extract-content" Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.144511 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="96a9f61e-825e-4999-a5e5-931111334a3c" containerName="extract-content" Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.144679 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="96a9f61e-825e-4999-a5e5-931111334a3c" containerName="registry-server" Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.145384 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556730-txd8w" Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.148583 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.149200 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.149590 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.156317 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556730-txd8w"] Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.194723 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dr924\" (UniqueName: \"kubernetes.io/projected/c4d4ab40-a5ab-4b39-b19e-043766174116-kube-api-access-dr924\") pod \"auto-csr-approver-29556730-txd8w\" (UID: \"c4d4ab40-a5ab-4b39-b19e-043766174116\") " pod="openshift-infra/auto-csr-approver-29556730-txd8w" Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.297290 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dr924\" (UniqueName: \"kubernetes.io/projected/c4d4ab40-a5ab-4b39-b19e-043766174116-kube-api-access-dr924\") pod \"auto-csr-approver-29556730-txd8w\" (UID: \"c4d4ab40-a5ab-4b39-b19e-043766174116\") " pod="openshift-infra/auto-csr-approver-29556730-txd8w" Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.317838 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dr924\" (UniqueName: \"kubernetes.io/projected/c4d4ab40-a5ab-4b39-b19e-043766174116-kube-api-access-dr924\") pod \"auto-csr-approver-29556730-txd8w\" (UID: \"c4d4ab40-a5ab-4b39-b19e-043766174116\") " pod="openshift-infra/auto-csr-approver-29556730-txd8w" Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.469220 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556730-txd8w" Mar 13 12:10:00 crc kubenswrapper[4632]: I0313 12:10:00.963359 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556730-txd8w"] Mar 13 12:10:01 crc kubenswrapper[4632]: I0313 12:10:01.549171 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556730-txd8w" event={"ID":"c4d4ab40-a5ab-4b39-b19e-043766174116","Type":"ContainerStarted","Data":"3da31d4e48d9483cf964b28af6122166fc360e9fbe694880471ddbd5aaf304be"} Mar 13 12:10:03 crc kubenswrapper[4632]: I0313 12:10:03.569696 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556730-txd8w" event={"ID":"c4d4ab40-a5ab-4b39-b19e-043766174116","Type":"ContainerStarted","Data":"47f2264b87b9b8761c8013ae1aa0f697b6d23277abd018e4664e7e2eed7771a3"} Mar 13 12:10:03 crc kubenswrapper[4632]: I0313 12:10:03.589724 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556730-txd8w" podStartSLOduration=2.226333465 podStartE2EDuration="3.589704869s" podCreationTimestamp="2026-03-13 12:10:00 +0000 UTC" firstStartedPulling="2026-03-13 12:10:00.974415129 +0000 UTC m=+7574.996945252" lastFinishedPulling="2026-03-13 12:10:02.337786523 +0000 UTC m=+7576.360316656" observedRunningTime="2026-03-13 12:10:03.586795659 +0000 UTC m=+7577.609325792" watchObservedRunningTime="2026-03-13 12:10:03.589704869 +0000 UTC m=+7577.612235002" Mar 13 12:10:04 crc kubenswrapper[4632]: I0313 12:10:04.601525 4632 generic.go:334] "Generic (PLEG): container finished" podID="c4d4ab40-a5ab-4b39-b19e-043766174116" containerID="47f2264b87b9b8761c8013ae1aa0f697b6d23277abd018e4664e7e2eed7771a3" exitCode=0 Mar 13 12:10:04 crc kubenswrapper[4632]: I0313 12:10:04.601845 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556730-txd8w" event={"ID":"c4d4ab40-a5ab-4b39-b19e-043766174116","Type":"ContainerDied","Data":"47f2264b87b9b8761c8013ae1aa0f697b6d23277abd018e4664e7e2eed7771a3"} Mar 13 12:10:05 crc kubenswrapper[4632]: I0313 12:10:05.970049 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556730-txd8w" Mar 13 12:10:06 crc kubenswrapper[4632]: I0313 12:10:06.119523 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dr924\" (UniqueName: \"kubernetes.io/projected/c4d4ab40-a5ab-4b39-b19e-043766174116-kube-api-access-dr924\") pod \"c4d4ab40-a5ab-4b39-b19e-043766174116\" (UID: \"c4d4ab40-a5ab-4b39-b19e-043766174116\") " Mar 13 12:10:06 crc kubenswrapper[4632]: I0313 12:10:06.128132 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4d4ab40-a5ab-4b39-b19e-043766174116-kube-api-access-dr924" (OuterVolumeSpecName: "kube-api-access-dr924") pod "c4d4ab40-a5ab-4b39-b19e-043766174116" (UID: "c4d4ab40-a5ab-4b39-b19e-043766174116"). InnerVolumeSpecName "kube-api-access-dr924". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:10:06 crc kubenswrapper[4632]: I0313 12:10:06.223060 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dr924\" (UniqueName: \"kubernetes.io/projected/c4d4ab40-a5ab-4b39-b19e-043766174116-kube-api-access-dr924\") on node \"crc\" DevicePath \"\"" Mar 13 12:10:06 crc kubenswrapper[4632]: I0313 12:10:06.622873 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556730-txd8w" event={"ID":"c4d4ab40-a5ab-4b39-b19e-043766174116","Type":"ContainerDied","Data":"3da31d4e48d9483cf964b28af6122166fc360e9fbe694880471ddbd5aaf304be"} Mar 13 12:10:06 crc kubenswrapper[4632]: I0313 12:10:06.622918 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556730-txd8w" Mar 13 12:10:06 crc kubenswrapper[4632]: I0313 12:10:06.622920 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3da31d4e48d9483cf964b28af6122166fc360e9fbe694880471ddbd5aaf304be" Mar 13 12:10:06 crc kubenswrapper[4632]: I0313 12:10:06.693397 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556724-thbvk"] Mar 13 12:10:06 crc kubenswrapper[4632]: I0313 12:10:06.706517 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556724-thbvk"] Mar 13 12:10:08 crc kubenswrapper[4632]: I0313 12:10:08.056729 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe" path="/var/lib/kubelet/pods/af27fbb5-f222-4fcc-a4d0-c4b4673cf1fe/volumes" Mar 13 12:10:24 crc kubenswrapper[4632]: I0313 12:10:24.839864 4632 scope.go:117] "RemoveContainer" containerID="b2b91ac6b566e0c21758b3baa48d2497ca87c0016cb01ff16589e8a5fd981c2d" Mar 13 12:10:32 crc kubenswrapper[4632]: I0313 12:10:32.876555 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-26z26"] Mar 13 12:10:32 crc kubenswrapper[4632]: E0313 12:10:32.878092 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4d4ab40-a5ab-4b39-b19e-043766174116" containerName="oc" Mar 13 12:10:32 crc kubenswrapper[4632]: I0313 12:10:32.878108 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4d4ab40-a5ab-4b39-b19e-043766174116" containerName="oc" Mar 13 12:10:32 crc kubenswrapper[4632]: I0313 12:10:32.878284 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4d4ab40-a5ab-4b39-b19e-043766174116" containerName="oc" Mar 13 12:10:32 crc kubenswrapper[4632]: I0313 12:10:32.881513 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-26z26" Mar 13 12:10:32 crc kubenswrapper[4632]: I0313 12:10:32.906106 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-26z26"] Mar 13 12:10:32 crc kubenswrapper[4632]: I0313 12:10:32.956590 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-utilities\") pod \"community-operators-26z26\" (UID: \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\") " pod="openshift-marketplace/community-operators-26z26" Mar 13 12:10:32 crc kubenswrapper[4632]: I0313 12:10:32.956634 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-catalog-content\") pod \"community-operators-26z26\" (UID: \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\") " pod="openshift-marketplace/community-operators-26z26" Mar 13 12:10:32 crc kubenswrapper[4632]: I0313 12:10:32.956964 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5zh92\" (UniqueName: \"kubernetes.io/projected/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-kube-api-access-5zh92\") pod \"community-operators-26z26\" (UID: \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\") " pod="openshift-marketplace/community-operators-26z26" Mar 13 12:10:33 crc kubenswrapper[4632]: I0313 12:10:33.059082 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5zh92\" (UniqueName: \"kubernetes.io/projected/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-kube-api-access-5zh92\") pod \"community-operators-26z26\" (UID: \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\") " pod="openshift-marketplace/community-operators-26z26" Mar 13 12:10:33 crc kubenswrapper[4632]: I0313 12:10:33.059357 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-utilities\") pod \"community-operators-26z26\" (UID: \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\") " pod="openshift-marketplace/community-operators-26z26" Mar 13 12:10:33 crc kubenswrapper[4632]: I0313 12:10:33.059424 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-catalog-content\") pod \"community-operators-26z26\" (UID: \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\") " pod="openshift-marketplace/community-operators-26z26" Mar 13 12:10:33 crc kubenswrapper[4632]: I0313 12:10:33.060072 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-utilities\") pod \"community-operators-26z26\" (UID: \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\") " pod="openshift-marketplace/community-operators-26z26" Mar 13 12:10:33 crc kubenswrapper[4632]: I0313 12:10:33.060108 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-catalog-content\") pod \"community-operators-26z26\" (UID: \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\") " pod="openshift-marketplace/community-operators-26z26" Mar 13 12:10:33 crc kubenswrapper[4632]: I0313 12:10:33.081287 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5zh92\" (UniqueName: \"kubernetes.io/projected/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-kube-api-access-5zh92\") pod \"community-operators-26z26\" (UID: \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\") " pod="openshift-marketplace/community-operators-26z26" Mar 13 12:10:33 crc kubenswrapper[4632]: I0313 12:10:33.214345 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-26z26" Mar 13 12:10:33 crc kubenswrapper[4632]: I0313 12:10:33.686886 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-26z26"] Mar 13 12:10:33 crc kubenswrapper[4632]: W0313 12:10:33.696313 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7654d2d1_ef2c_4ae6_a358_d5efdfaf3c6b.slice/crio-49fadcbb7750d63803d60167fee630b03775174c0b2e88c50e45e7e7c2903030 WatchSource:0}: Error finding container 49fadcbb7750d63803d60167fee630b03775174c0b2e88c50e45e7e7c2903030: Status 404 returned error can't find the container with id 49fadcbb7750d63803d60167fee630b03775174c0b2e88c50e45e7e7c2903030 Mar 13 12:10:33 crc kubenswrapper[4632]: I0313 12:10:33.874119 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-26z26" event={"ID":"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b","Type":"ContainerStarted","Data":"49fadcbb7750d63803d60167fee630b03775174c0b2e88c50e45e7e7c2903030"} Mar 13 12:10:34 crc kubenswrapper[4632]: I0313 12:10:34.894501 4632 generic.go:334] "Generic (PLEG): container finished" podID="7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" containerID="f02456bf592705883010995ba855a910e6930582529323b185cedcde26acc636" exitCode=0 Mar 13 12:10:34 crc kubenswrapper[4632]: I0313 12:10:34.894572 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-26z26" event={"ID":"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b","Type":"ContainerDied","Data":"f02456bf592705883010995ba855a910e6930582529323b185cedcde26acc636"} Mar 13 12:10:35 crc kubenswrapper[4632]: I0313 12:10:35.908835 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-26z26" event={"ID":"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b","Type":"ContainerStarted","Data":"856643e5316cdb3679641b711adfb80511d52a0130b5a170d00c6ac71cd987d6"} Mar 13 12:10:38 crc kubenswrapper[4632]: I0313 12:10:38.944334 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-26z26" event={"ID":"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b","Type":"ContainerDied","Data":"856643e5316cdb3679641b711adfb80511d52a0130b5a170d00c6ac71cd987d6"} Mar 13 12:10:38 crc kubenswrapper[4632]: I0313 12:10:38.944337 4632 generic.go:334] "Generic (PLEG): container finished" podID="7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" containerID="856643e5316cdb3679641b711adfb80511d52a0130b5a170d00c6ac71cd987d6" exitCode=0 Mar 13 12:10:39 crc kubenswrapper[4632]: I0313 12:10:39.955308 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-26z26" event={"ID":"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b","Type":"ContainerStarted","Data":"921c63ac9b57ae71823a93ed0ae4b360f16e3e68d7c9d943c870848a775d0669"} Mar 13 12:10:39 crc kubenswrapper[4632]: I0313 12:10:39.973469 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-26z26" podStartSLOduration=3.489528888 podStartE2EDuration="7.973452295s" podCreationTimestamp="2026-03-13 12:10:32 +0000 UTC" firstStartedPulling="2026-03-13 12:10:34.896540956 +0000 UTC m=+7608.919071089" lastFinishedPulling="2026-03-13 12:10:39.380464363 +0000 UTC m=+7613.402994496" observedRunningTime="2026-03-13 12:10:39.971248451 +0000 UTC m=+7613.993778634" watchObservedRunningTime="2026-03-13 12:10:39.973452295 +0000 UTC m=+7613.995982428" Mar 13 12:10:40 crc kubenswrapper[4632]: I0313 12:10:40.461355 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:10:40 crc kubenswrapper[4632]: I0313 12:10:40.463058 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:10:43 crc kubenswrapper[4632]: I0313 12:10:43.215168 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-26z26" Mar 13 12:10:43 crc kubenswrapper[4632]: I0313 12:10:43.215502 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-26z26" Mar 13 12:10:44 crc kubenswrapper[4632]: I0313 12:10:44.265587 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-26z26" podUID="7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" containerName="registry-server" probeResult="failure" output=< Mar 13 12:10:44 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:10:44 crc kubenswrapper[4632]: > Mar 13 12:10:54 crc kubenswrapper[4632]: I0313 12:10:54.280179 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-26z26" podUID="7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" containerName="registry-server" probeResult="failure" output=< Mar 13 12:10:54 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:10:54 crc kubenswrapper[4632]: > Mar 13 12:11:03 crc kubenswrapper[4632]: I0313 12:11:03.279908 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-26z26" Mar 13 12:11:03 crc kubenswrapper[4632]: I0313 12:11:03.338726 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-26z26" Mar 13 12:11:04 crc kubenswrapper[4632]: I0313 12:11:04.075231 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-26z26"] Mar 13 12:11:05 crc kubenswrapper[4632]: I0313 12:11:05.239627 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-26z26" podUID="7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" containerName="registry-server" containerID="cri-o://921c63ac9b57ae71823a93ed0ae4b360f16e3e68d7c9d943c870848a775d0669" gracePeriod=2 Mar 13 12:11:06 crc kubenswrapper[4632]: I0313 12:11:06.264711 4632 generic.go:334] "Generic (PLEG): container finished" podID="7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" containerID="921c63ac9b57ae71823a93ed0ae4b360f16e3e68d7c9d943c870848a775d0669" exitCode=0 Mar 13 12:11:06 crc kubenswrapper[4632]: I0313 12:11:06.264903 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-26z26" event={"ID":"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b","Type":"ContainerDied","Data":"921c63ac9b57ae71823a93ed0ae4b360f16e3e68d7c9d943c870848a775d0669"} Mar 13 12:11:06 crc kubenswrapper[4632]: I0313 12:11:06.613446 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-26z26" Mar 13 12:11:06 crc kubenswrapper[4632]: I0313 12:11:06.764633 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-utilities\") pod \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\" (UID: \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\") " Mar 13 12:11:06 crc kubenswrapper[4632]: I0313 12:11:06.765081 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-catalog-content\") pod \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\" (UID: \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\") " Mar 13 12:11:06 crc kubenswrapper[4632]: I0313 12:11:06.765137 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5zh92\" (UniqueName: \"kubernetes.io/projected/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-kube-api-access-5zh92\") pod \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\" (UID: \"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b\") " Mar 13 12:11:06 crc kubenswrapper[4632]: I0313 12:11:06.765602 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-utilities" (OuterVolumeSpecName: "utilities") pod "7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" (UID: "7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:11:06 crc kubenswrapper[4632]: I0313 12:11:06.765903 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:11:06 crc kubenswrapper[4632]: I0313 12:11:06.860222 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-kube-api-access-5zh92" (OuterVolumeSpecName: "kube-api-access-5zh92") pod "7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" (UID: "7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b"). InnerVolumeSpecName "kube-api-access-5zh92". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:11:06 crc kubenswrapper[4632]: I0313 12:11:06.868052 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5zh92\" (UniqueName: \"kubernetes.io/projected/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-kube-api-access-5zh92\") on node \"crc\" DevicePath \"\"" Mar 13 12:11:07 crc kubenswrapper[4632]: I0313 12:11:07.016298 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" (UID: "7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:11:07 crc kubenswrapper[4632]: I0313 12:11:07.071793 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:11:07 crc kubenswrapper[4632]: I0313 12:11:07.280689 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-26z26" event={"ID":"7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b","Type":"ContainerDied","Data":"49fadcbb7750d63803d60167fee630b03775174c0b2e88c50e45e7e7c2903030"} Mar 13 12:11:07 crc kubenswrapper[4632]: I0313 12:11:07.280790 4632 scope.go:117] "RemoveContainer" containerID="921c63ac9b57ae71823a93ed0ae4b360f16e3e68d7c9d943c870848a775d0669" Mar 13 12:11:07 crc kubenswrapper[4632]: I0313 12:11:07.281465 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-26z26" Mar 13 12:11:07 crc kubenswrapper[4632]: I0313 12:11:07.319385 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-26z26"] Mar 13 12:11:07 crc kubenswrapper[4632]: I0313 12:11:07.321028 4632 scope.go:117] "RemoveContainer" containerID="856643e5316cdb3679641b711adfb80511d52a0130b5a170d00c6ac71cd987d6" Mar 13 12:11:07 crc kubenswrapper[4632]: I0313 12:11:07.328655 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-26z26"] Mar 13 12:11:07 crc kubenswrapper[4632]: I0313 12:11:07.355560 4632 scope.go:117] "RemoveContainer" containerID="f02456bf592705883010995ba855a910e6930582529323b185cedcde26acc636" Mar 13 12:11:08 crc kubenswrapper[4632]: I0313 12:11:08.056226 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" path="/var/lib/kubelet/pods/7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b/volumes" Mar 13 12:11:10 crc kubenswrapper[4632]: I0313 12:11:10.469631 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:11:10 crc kubenswrapper[4632]: I0313 12:11:10.470795 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:11:40 crc kubenswrapper[4632]: I0313 12:11:40.461381 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:11:40 crc kubenswrapper[4632]: I0313 12:11:40.462023 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:11:40 crc kubenswrapper[4632]: I0313 12:11:40.462074 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 12:11:40 crc kubenswrapper[4632]: I0313 12:11:40.462897 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ef6a755da94d8b26aaa61b1a356ec9030e87ec1440f6bdf1f6abec8411efbdd9"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 12:11:40 crc kubenswrapper[4632]: I0313 12:11:40.463003 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://ef6a755da94d8b26aaa61b1a356ec9030e87ec1440f6bdf1f6abec8411efbdd9" gracePeriod=600 Mar 13 12:11:40 crc kubenswrapper[4632]: I0313 12:11:40.629435 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="ef6a755da94d8b26aaa61b1a356ec9030e87ec1440f6bdf1f6abec8411efbdd9" exitCode=0 Mar 13 12:11:40 crc kubenswrapper[4632]: I0313 12:11:40.629492 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"ef6a755da94d8b26aaa61b1a356ec9030e87ec1440f6bdf1f6abec8411efbdd9"} Mar 13 12:11:40 crc kubenswrapper[4632]: I0313 12:11:40.629535 4632 scope.go:117] "RemoveContainer" containerID="acc57c5b8ad70899e139ed86509fb64c5bf067344e4285c37f35406b8db0c7a6" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.640003 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9"} Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.817903 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-695f666b49-nw48z"] Mar 13 12:11:41 crc kubenswrapper[4632]: E0313 12:11:41.818392 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" containerName="extract-utilities" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.818417 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" containerName="extract-utilities" Mar 13 12:11:41 crc kubenswrapper[4632]: E0313 12:11:41.818450 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" containerName="extract-content" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.818459 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" containerName="extract-content" Mar 13 12:11:41 crc kubenswrapper[4632]: E0313 12:11:41.818479 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" containerName="registry-server" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.818487 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" containerName="registry-server" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.818732 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="7654d2d1-ef2c-4ae6-a358-d5efdfaf3c6b" containerName="registry-server" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.820915 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.870013 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxcff\" (UniqueName: \"kubernetes.io/projected/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-kube-api-access-mxcff\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.870088 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-ovndb-tls-certs\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.870168 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-internal-tls-certs\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.870324 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-public-tls-certs\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.870366 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-combined-ca-bundle\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.870549 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-httpd-config\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.870681 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-config\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.918536 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-695f666b49-nw48z"] Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.973293 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-config\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.973373 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxcff\" (UniqueName: \"kubernetes.io/projected/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-kube-api-access-mxcff\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.973419 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-ovndb-tls-certs\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.973500 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-internal-tls-certs\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.973544 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-public-tls-certs\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.973570 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-combined-ca-bundle\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.973638 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-httpd-config\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.985046 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-config\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.985193 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-internal-tls-certs\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.985594 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-combined-ca-bundle\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.988616 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-public-tls-certs\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:41 crc kubenswrapper[4632]: I0313 12:11:41.989145 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-httpd-config\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:42 crc kubenswrapper[4632]: I0313 12:11:42.003363 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-ovndb-tls-certs\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:42 crc kubenswrapper[4632]: I0313 12:11:42.006686 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxcff\" (UniqueName: \"kubernetes.io/projected/3a5c1185-e64b-44a9-b4b8-0108d4e80f9a-kube-api-access-mxcff\") pod \"neutron-695f666b49-nw48z\" (UID: \"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a\") " pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:42 crc kubenswrapper[4632]: I0313 12:11:42.140734 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:43 crc kubenswrapper[4632]: I0313 12:11:43.575843 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-695f666b49-nw48z"] Mar 13 12:11:43 crc kubenswrapper[4632]: W0313 12:11:43.605085 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3a5c1185_e64b_44a9_b4b8_0108d4e80f9a.slice/crio-6b30682db3693448e083a584c4d9224eba418a7b11fcf1ea70e89d376cc0e7a8 WatchSource:0}: Error finding container 6b30682db3693448e083a584c4d9224eba418a7b11fcf1ea70e89d376cc0e7a8: Status 404 returned error can't find the container with id 6b30682db3693448e083a584c4d9224eba418a7b11fcf1ea70e89d376cc0e7a8 Mar 13 12:11:43 crc kubenswrapper[4632]: I0313 12:11:43.668372 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-695f666b49-nw48z" event={"ID":"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a","Type":"ContainerStarted","Data":"6b30682db3693448e083a584c4d9224eba418a7b11fcf1ea70e89d376cc0e7a8"} Mar 13 12:11:44 crc kubenswrapper[4632]: I0313 12:11:44.687000 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-695f666b49-nw48z" event={"ID":"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a","Type":"ContainerStarted","Data":"c42a0ba5071954445f9abf4a2ace1075becd42a3374534a7de730140449c8869"} Mar 13 12:11:44 crc kubenswrapper[4632]: I0313 12:11:44.687663 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:11:44 crc kubenswrapper[4632]: I0313 12:11:44.687679 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-695f666b49-nw48z" event={"ID":"3a5c1185-e64b-44a9-b4b8-0108d4e80f9a","Type":"ContainerStarted","Data":"e884f11f8966438c8d4d0b48555144b504cc8fbbd1a49c1f1457de8f39a82ad3"} Mar 13 12:12:00 crc kubenswrapper[4632]: I0313 12:12:00.166565 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-695f666b49-nw48z" podStartSLOduration=19.166541269 podStartE2EDuration="19.166541269s" podCreationTimestamp="2026-03-13 12:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 12:11:44.710797102 +0000 UTC m=+7678.733327245" watchObservedRunningTime="2026-03-13 12:12:00.166541269 +0000 UTC m=+7694.189071402" Mar 13 12:12:00 crc kubenswrapper[4632]: I0313 12:12:00.169863 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556732-zffdj"] Mar 13 12:12:00 crc kubenswrapper[4632]: I0313 12:12:00.171431 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556732-zffdj" Mar 13 12:12:00 crc kubenswrapper[4632]: I0313 12:12:00.178346 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:12:00 crc kubenswrapper[4632]: I0313 12:12:00.178574 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:12:00 crc kubenswrapper[4632]: I0313 12:12:00.180696 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556732-zffdj"] Mar 13 12:12:00 crc kubenswrapper[4632]: I0313 12:12:00.181258 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:12:00 crc kubenswrapper[4632]: I0313 12:12:00.249511 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24l62\" (UniqueName: \"kubernetes.io/projected/69547021-fae1-4ad6-8745-c327bb079dce-kube-api-access-24l62\") pod \"auto-csr-approver-29556732-zffdj\" (UID: \"69547021-fae1-4ad6-8745-c327bb079dce\") " pod="openshift-infra/auto-csr-approver-29556732-zffdj" Mar 13 12:12:00 crc kubenswrapper[4632]: I0313 12:12:00.351855 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24l62\" (UniqueName: \"kubernetes.io/projected/69547021-fae1-4ad6-8745-c327bb079dce-kube-api-access-24l62\") pod \"auto-csr-approver-29556732-zffdj\" (UID: \"69547021-fae1-4ad6-8745-c327bb079dce\") " pod="openshift-infra/auto-csr-approver-29556732-zffdj" Mar 13 12:12:00 crc kubenswrapper[4632]: I0313 12:12:00.384849 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24l62\" (UniqueName: \"kubernetes.io/projected/69547021-fae1-4ad6-8745-c327bb079dce-kube-api-access-24l62\") pod \"auto-csr-approver-29556732-zffdj\" (UID: \"69547021-fae1-4ad6-8745-c327bb079dce\") " pod="openshift-infra/auto-csr-approver-29556732-zffdj" Mar 13 12:12:00 crc kubenswrapper[4632]: I0313 12:12:00.498666 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556732-zffdj" Mar 13 12:12:01 crc kubenswrapper[4632]: I0313 12:12:01.251616 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556732-zffdj"] Mar 13 12:12:01 crc kubenswrapper[4632]: W0313 12:12:01.264171 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69547021_fae1_4ad6_8745_c327bb079dce.slice/crio-b661ff26ad7f8831aa3ef67c981c437bbf2ba8ab70664de65931cf5f2426bdcd WatchSource:0}: Error finding container b661ff26ad7f8831aa3ef67c981c437bbf2ba8ab70664de65931cf5f2426bdcd: Status 404 returned error can't find the container with id b661ff26ad7f8831aa3ef67c981c437bbf2ba8ab70664de65931cf5f2426bdcd Mar 13 12:12:01 crc kubenswrapper[4632]: I0313 12:12:01.870279 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556732-zffdj" event={"ID":"69547021-fae1-4ad6-8745-c327bb079dce","Type":"ContainerStarted","Data":"b661ff26ad7f8831aa3ef67c981c437bbf2ba8ab70664de65931cf5f2426bdcd"} Mar 13 12:12:03 crc kubenswrapper[4632]: I0313 12:12:03.893925 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556732-zffdj" event={"ID":"69547021-fae1-4ad6-8745-c327bb079dce","Type":"ContainerStarted","Data":"e6be43aa992650e79f391597a8fccb4cc829615f369f4902627c5d5e92b6ab1e"} Mar 13 12:12:03 crc kubenswrapper[4632]: I0313 12:12:03.913604 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556732-zffdj" podStartSLOduration=2.966690912 podStartE2EDuration="3.913577671s" podCreationTimestamp="2026-03-13 12:12:00 +0000 UTC" firstStartedPulling="2026-03-13 12:12:01.265992751 +0000 UTC m=+7695.288522884" lastFinishedPulling="2026-03-13 12:12:02.21287951 +0000 UTC m=+7696.235409643" observedRunningTime="2026-03-13 12:12:03.909043649 +0000 UTC m=+7697.931573802" watchObservedRunningTime="2026-03-13 12:12:03.913577671 +0000 UTC m=+7697.936107804" Mar 13 12:12:04 crc kubenswrapper[4632]: I0313 12:12:04.903540 4632 generic.go:334] "Generic (PLEG): container finished" podID="69547021-fae1-4ad6-8745-c327bb079dce" containerID="e6be43aa992650e79f391597a8fccb4cc829615f369f4902627c5d5e92b6ab1e" exitCode=0 Mar 13 12:12:04 crc kubenswrapper[4632]: I0313 12:12:04.903650 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556732-zffdj" event={"ID":"69547021-fae1-4ad6-8745-c327bb079dce","Type":"ContainerDied","Data":"e6be43aa992650e79f391597a8fccb4cc829615f369f4902627c5d5e92b6ab1e"} Mar 13 12:12:06 crc kubenswrapper[4632]: I0313 12:12:06.568314 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556732-zffdj" Mar 13 12:12:06 crc kubenswrapper[4632]: I0313 12:12:06.733013 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24l62\" (UniqueName: \"kubernetes.io/projected/69547021-fae1-4ad6-8745-c327bb079dce-kube-api-access-24l62\") pod \"69547021-fae1-4ad6-8745-c327bb079dce\" (UID: \"69547021-fae1-4ad6-8745-c327bb079dce\") " Mar 13 12:12:06 crc kubenswrapper[4632]: I0313 12:12:06.742140 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69547021-fae1-4ad6-8745-c327bb079dce-kube-api-access-24l62" (OuterVolumeSpecName: "kube-api-access-24l62") pod "69547021-fae1-4ad6-8745-c327bb079dce" (UID: "69547021-fae1-4ad6-8745-c327bb079dce"). InnerVolumeSpecName "kube-api-access-24l62". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:12:06 crc kubenswrapper[4632]: I0313 12:12:06.835534 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24l62\" (UniqueName: \"kubernetes.io/projected/69547021-fae1-4ad6-8745-c327bb079dce-kube-api-access-24l62\") on node \"crc\" DevicePath \"\"" Mar 13 12:12:06 crc kubenswrapper[4632]: I0313 12:12:06.925460 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556732-zffdj" event={"ID":"69547021-fae1-4ad6-8745-c327bb079dce","Type":"ContainerDied","Data":"b661ff26ad7f8831aa3ef67c981c437bbf2ba8ab70664de65931cf5f2426bdcd"} Mar 13 12:12:06 crc kubenswrapper[4632]: I0313 12:12:06.925522 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b661ff26ad7f8831aa3ef67c981c437bbf2ba8ab70664de65931cf5f2426bdcd" Mar 13 12:12:06 crc kubenswrapper[4632]: I0313 12:12:06.925652 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556732-zffdj" Mar 13 12:12:07 crc kubenswrapper[4632]: I0313 12:12:07.008688 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556726-pjf8m"] Mar 13 12:12:07 crc kubenswrapper[4632]: I0313 12:12:07.018930 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556726-pjf8m"] Mar 13 12:12:08 crc kubenswrapper[4632]: I0313 12:12:08.058229 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2bb028cb-1d8f-4f09-8c73-135a6ce2bb46" path="/var/lib/kubelet/pods/2bb028cb-1d8f-4f09-8c73-135a6ce2bb46/volumes" Mar 13 12:12:12 crc kubenswrapper[4632]: I0313 12:12:12.159090 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-695f666b49-nw48z" Mar 13 12:12:12 crc kubenswrapper[4632]: I0313 12:12:12.271682 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-59586ff4c9-s4xn7"] Mar 13 12:12:12 crc kubenswrapper[4632]: I0313 12:12:12.271930 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-59586ff4c9-s4xn7" podUID="8b9495c7-c9ae-4a07-b216-a250d4cd274e" containerName="neutron-api" containerID="cri-o://94fc75b5bf96292690ce359a5d4ce65dd30bc2b06b1aeb4d309bd6e1dcd7e70c" gracePeriod=30 Mar 13 12:12:12 crc kubenswrapper[4632]: I0313 12:12:12.272130 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-59586ff4c9-s4xn7" podUID="8b9495c7-c9ae-4a07-b216-a250d4cd274e" containerName="neutron-httpd" containerID="cri-o://82e1ea3147a5e24713a581b7fd1d1be6dc38543edaf91f1fa20ce5282f06b072" gracePeriod=30 Mar 13 12:12:12 crc kubenswrapper[4632]: I0313 12:12:12.981274 4632 generic.go:334] "Generic (PLEG): container finished" podID="8b9495c7-c9ae-4a07-b216-a250d4cd274e" containerID="82e1ea3147a5e24713a581b7fd1d1be6dc38543edaf91f1fa20ce5282f06b072" exitCode=0 Mar 13 12:12:12 crc kubenswrapper[4632]: I0313 12:12:12.981352 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59586ff4c9-s4xn7" event={"ID":"8b9495c7-c9ae-4a07-b216-a250d4cd274e","Type":"ContainerDied","Data":"82e1ea3147a5e24713a581b7fd1d1be6dc38543edaf91f1fa20ce5282f06b072"} Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.014218 4632 generic.go:334] "Generic (PLEG): container finished" podID="8b9495c7-c9ae-4a07-b216-a250d4cd274e" containerID="94fc75b5bf96292690ce359a5d4ce65dd30bc2b06b1aeb4d309bd6e1dcd7e70c" exitCode=0 Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.014341 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59586ff4c9-s4xn7" event={"ID":"8b9495c7-c9ae-4a07-b216-a250d4cd274e","Type":"ContainerDied","Data":"94fc75b5bf96292690ce359a5d4ce65dd30bc2b06b1aeb4d309bd6e1dcd7e70c"} Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.115408 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.226974 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-ovndb-tls-certs\") pod \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.227211 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-combined-ca-bundle\") pod \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.227263 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llcp8\" (UniqueName: \"kubernetes.io/projected/8b9495c7-c9ae-4a07-b216-a250d4cd274e-kube-api-access-llcp8\") pod \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.227296 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-public-tls-certs\") pod \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.227341 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-httpd-config\") pod \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.227366 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-internal-tls-certs\") pod \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.227421 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-config\") pod \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\" (UID: \"8b9495c7-c9ae-4a07-b216-a250d4cd274e\") " Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.259810 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b9495c7-c9ae-4a07-b216-a250d4cd274e-kube-api-access-llcp8" (OuterVolumeSpecName: "kube-api-access-llcp8") pod "8b9495c7-c9ae-4a07-b216-a250d4cd274e" (UID: "8b9495c7-c9ae-4a07-b216-a250d4cd274e"). InnerVolumeSpecName "kube-api-access-llcp8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.271280 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "8b9495c7-c9ae-4a07-b216-a250d4cd274e" (UID: "8b9495c7-c9ae-4a07-b216-a250d4cd274e"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.309901 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "8b9495c7-c9ae-4a07-b216-a250d4cd274e" (UID: "8b9495c7-c9ae-4a07-b216-a250d4cd274e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.318535 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "8b9495c7-c9ae-4a07-b216-a250d4cd274e" (UID: "8b9495c7-c9ae-4a07-b216-a250d4cd274e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.333408 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llcp8\" (UniqueName: \"kubernetes.io/projected/8b9495c7-c9ae-4a07-b216-a250d4cd274e-kube-api-access-llcp8\") on node \"crc\" DevicePath \"\"" Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.334006 4632 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.334181 4632 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.334209 4632 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.343899 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8b9495c7-c9ae-4a07-b216-a250d4cd274e" (UID: "8b9495c7-c9ae-4a07-b216-a250d4cd274e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.348728 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "8b9495c7-c9ae-4a07-b216-a250d4cd274e" (UID: "8b9495c7-c9ae-4a07-b216-a250d4cd274e"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.361061 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-config" (OuterVolumeSpecName: "config") pod "8b9495c7-c9ae-4a07-b216-a250d4cd274e" (UID: "8b9495c7-c9ae-4a07-b216-a250d4cd274e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.435715 4632 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.435759 4632 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-config\") on node \"crc\" DevicePath \"\"" Mar 13 12:12:16 crc kubenswrapper[4632]: I0313 12:12:16.435773 4632 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/8b9495c7-c9ae-4a07-b216-a250d4cd274e-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 12:12:17 crc kubenswrapper[4632]: I0313 12:12:17.030562 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-59586ff4c9-s4xn7" event={"ID":"8b9495c7-c9ae-4a07-b216-a250d4cd274e","Type":"ContainerDied","Data":"baa77c1d37fb9c8cc82676bdaaab769c07869647e902c21581eef67e591e5d68"} Mar 13 12:12:17 crc kubenswrapper[4632]: I0313 12:12:17.030783 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-59586ff4c9-s4xn7" Mar 13 12:12:17 crc kubenswrapper[4632]: I0313 12:12:17.031860 4632 scope.go:117] "RemoveContainer" containerID="82e1ea3147a5e24713a581b7fd1d1be6dc38543edaf91f1fa20ce5282f06b072" Mar 13 12:12:17 crc kubenswrapper[4632]: I0313 12:12:17.078909 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-59586ff4c9-s4xn7"] Mar 13 12:12:17 crc kubenswrapper[4632]: I0313 12:12:17.083067 4632 scope.go:117] "RemoveContainer" containerID="94fc75b5bf96292690ce359a5d4ce65dd30bc2b06b1aeb4d309bd6e1dcd7e70c" Mar 13 12:12:17 crc kubenswrapper[4632]: I0313 12:12:17.087588 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-59586ff4c9-s4xn7"] Mar 13 12:12:18 crc kubenswrapper[4632]: I0313 12:12:18.060885 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b9495c7-c9ae-4a07-b216-a250d4cd274e" path="/var/lib/kubelet/pods/8b9495c7-c9ae-4a07-b216-a250d4cd274e/volumes" Mar 13 12:12:23 crc kubenswrapper[4632]: I0313 12:12:23.251067 4632 trace.go:236] Trace[334643179]: "Calculate volume metrics of multus-daemon-config for pod openshift-multus/multus-gqf22" (13-Mar-2026 12:12:21.664) (total time: 1581ms): Mar 13 12:12:23 crc kubenswrapper[4632]: Trace[334643179]: [1.581560521s] [1.581560521s] END Mar 13 12:12:25 crc kubenswrapper[4632]: I0313 12:12:25.035018 4632 scope.go:117] "RemoveContainer" containerID="d0bcc7d380da406a96a84a48eb87ce7d69302220aa1982823e25de22c01f77e0" Mar 13 12:12:43 crc kubenswrapper[4632]: I0313 12:12:43.784134 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8xrj9"] Mar 13 12:12:43 crc kubenswrapper[4632]: E0313 12:12:43.785145 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69547021-fae1-4ad6-8745-c327bb079dce" containerName="oc" Mar 13 12:12:43 crc kubenswrapper[4632]: I0313 12:12:43.785165 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="69547021-fae1-4ad6-8745-c327bb079dce" containerName="oc" Mar 13 12:12:43 crc kubenswrapper[4632]: E0313 12:12:43.785182 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b9495c7-c9ae-4a07-b216-a250d4cd274e" containerName="neutron-httpd" Mar 13 12:12:43 crc kubenswrapper[4632]: I0313 12:12:43.785190 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b9495c7-c9ae-4a07-b216-a250d4cd274e" containerName="neutron-httpd" Mar 13 12:12:43 crc kubenswrapper[4632]: E0313 12:12:43.785242 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b9495c7-c9ae-4a07-b216-a250d4cd274e" containerName="neutron-api" Mar 13 12:12:43 crc kubenswrapper[4632]: I0313 12:12:43.785252 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b9495c7-c9ae-4a07-b216-a250d4cd274e" containerName="neutron-api" Mar 13 12:12:43 crc kubenswrapper[4632]: I0313 12:12:43.785478 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b9495c7-c9ae-4a07-b216-a250d4cd274e" containerName="neutron-api" Mar 13 12:12:43 crc kubenswrapper[4632]: I0313 12:12:43.785504 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b9495c7-c9ae-4a07-b216-a250d4cd274e" containerName="neutron-httpd" Mar 13 12:12:43 crc kubenswrapper[4632]: I0313 12:12:43.785524 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="69547021-fae1-4ad6-8745-c327bb079dce" containerName="oc" Mar 13 12:12:43 crc kubenswrapper[4632]: I0313 12:12:43.797121 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8xrj9"] Mar 13 12:12:43 crc kubenswrapper[4632]: I0313 12:12:43.797267 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:12:43 crc kubenswrapper[4632]: I0313 12:12:43.903463 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58986885-b2ff-450a-b232-a26163de811a-utilities\") pod \"certified-operators-8xrj9\" (UID: \"58986885-b2ff-450a-b232-a26163de811a\") " pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:12:43 crc kubenswrapper[4632]: I0313 12:12:43.903800 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24rgb\" (UniqueName: \"kubernetes.io/projected/58986885-b2ff-450a-b232-a26163de811a-kube-api-access-24rgb\") pod \"certified-operators-8xrj9\" (UID: \"58986885-b2ff-450a-b232-a26163de811a\") " pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:12:43 crc kubenswrapper[4632]: I0313 12:12:43.903931 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58986885-b2ff-450a-b232-a26163de811a-catalog-content\") pod \"certified-operators-8xrj9\" (UID: \"58986885-b2ff-450a-b232-a26163de811a\") " pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:12:44 crc kubenswrapper[4632]: I0313 12:12:44.006087 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58986885-b2ff-450a-b232-a26163de811a-utilities\") pod \"certified-operators-8xrj9\" (UID: \"58986885-b2ff-450a-b232-a26163de811a\") " pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:12:44 crc kubenswrapper[4632]: I0313 12:12:44.006145 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24rgb\" (UniqueName: \"kubernetes.io/projected/58986885-b2ff-450a-b232-a26163de811a-kube-api-access-24rgb\") pod \"certified-operators-8xrj9\" (UID: \"58986885-b2ff-450a-b232-a26163de811a\") " pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:12:44 crc kubenswrapper[4632]: I0313 12:12:44.006195 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58986885-b2ff-450a-b232-a26163de811a-catalog-content\") pod \"certified-operators-8xrj9\" (UID: \"58986885-b2ff-450a-b232-a26163de811a\") " pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:12:44 crc kubenswrapper[4632]: I0313 12:12:44.006769 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58986885-b2ff-450a-b232-a26163de811a-catalog-content\") pod \"certified-operators-8xrj9\" (UID: \"58986885-b2ff-450a-b232-a26163de811a\") " pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:12:44 crc kubenswrapper[4632]: I0313 12:12:44.006921 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58986885-b2ff-450a-b232-a26163de811a-utilities\") pod \"certified-operators-8xrj9\" (UID: \"58986885-b2ff-450a-b232-a26163de811a\") " pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:12:44 crc kubenswrapper[4632]: I0313 12:12:44.039603 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24rgb\" (UniqueName: \"kubernetes.io/projected/58986885-b2ff-450a-b232-a26163de811a-kube-api-access-24rgb\") pod \"certified-operators-8xrj9\" (UID: \"58986885-b2ff-450a-b232-a26163de811a\") " pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:12:44 crc kubenswrapper[4632]: I0313 12:12:44.129993 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:12:44 crc kubenswrapper[4632]: I0313 12:12:44.716512 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8xrj9"] Mar 13 12:12:45 crc kubenswrapper[4632]: I0313 12:12:45.514497 4632 generic.go:334] "Generic (PLEG): container finished" podID="58986885-b2ff-450a-b232-a26163de811a" containerID="b27af3f83423049c58aa5949bc99cebcf9fba37e5504bb6ed378f00293a5e3ff" exitCode=0 Mar 13 12:12:45 crc kubenswrapper[4632]: I0313 12:12:45.514597 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8xrj9" event={"ID":"58986885-b2ff-450a-b232-a26163de811a","Type":"ContainerDied","Data":"b27af3f83423049c58aa5949bc99cebcf9fba37e5504bb6ed378f00293a5e3ff"} Mar 13 12:12:45 crc kubenswrapper[4632]: I0313 12:12:45.514872 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8xrj9" event={"ID":"58986885-b2ff-450a-b232-a26163de811a","Type":"ContainerStarted","Data":"e7db2df784e22387bb91352a6bc88e575028a70869b8cf3ee15bff73c9b7bfb4"} Mar 13 12:12:46 crc kubenswrapper[4632]: I0313 12:12:46.529894 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8xrj9" event={"ID":"58986885-b2ff-450a-b232-a26163de811a","Type":"ContainerStarted","Data":"2d9f6c616f332d122081782eff3ae0b45da85be1a8961803eef5bdd4c8ed6d19"} Mar 13 12:12:48 crc kubenswrapper[4632]: I0313 12:12:48.562916 4632 generic.go:334] "Generic (PLEG): container finished" podID="58986885-b2ff-450a-b232-a26163de811a" containerID="2d9f6c616f332d122081782eff3ae0b45da85be1a8961803eef5bdd4c8ed6d19" exitCode=0 Mar 13 12:12:48 crc kubenswrapper[4632]: I0313 12:12:48.563246 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8xrj9" event={"ID":"58986885-b2ff-450a-b232-a26163de811a","Type":"ContainerDied","Data":"2d9f6c616f332d122081782eff3ae0b45da85be1a8961803eef5bdd4c8ed6d19"} Mar 13 12:12:49 crc kubenswrapper[4632]: I0313 12:12:49.574807 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8xrj9" event={"ID":"58986885-b2ff-450a-b232-a26163de811a","Type":"ContainerStarted","Data":"2af6cb2b9380482b4ea542096a63d2330cf2514fce0a1df0deaa98cec712ba02"} Mar 13 12:12:49 crc kubenswrapper[4632]: I0313 12:12:49.612754 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8xrj9" podStartSLOduration=3.06037933 podStartE2EDuration="6.612728643s" podCreationTimestamp="2026-03-13 12:12:43 +0000 UTC" firstStartedPulling="2026-03-13 12:12:45.51641246 +0000 UTC m=+7739.538942593" lastFinishedPulling="2026-03-13 12:12:49.068761783 +0000 UTC m=+7743.091291906" observedRunningTime="2026-03-13 12:12:49.602282166 +0000 UTC m=+7743.624812319" watchObservedRunningTime="2026-03-13 12:12:49.612728643 +0000 UTC m=+7743.635258796" Mar 13 12:12:54 crc kubenswrapper[4632]: I0313 12:12:54.130324 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:12:54 crc kubenswrapper[4632]: I0313 12:12:54.130799 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:12:55 crc kubenswrapper[4632]: I0313 12:12:55.176537 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-8xrj9" podUID="58986885-b2ff-450a-b232-a26163de811a" containerName="registry-server" probeResult="failure" output=< Mar 13 12:12:55 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:12:55 crc kubenswrapper[4632]: > Mar 13 12:13:05 crc kubenswrapper[4632]: I0313 12:13:05.184095 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-8xrj9" podUID="58986885-b2ff-450a-b232-a26163de811a" containerName="registry-server" probeResult="failure" output=< Mar 13 12:13:05 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:13:05 crc kubenswrapper[4632]: > Mar 13 12:13:14 crc kubenswrapper[4632]: I0313 12:13:14.306913 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:13:14 crc kubenswrapper[4632]: I0313 12:13:14.361432 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:13:15 crc kubenswrapper[4632]: I0313 12:13:15.027922 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8xrj9"] Mar 13 12:13:15 crc kubenswrapper[4632]: I0313 12:13:15.904738 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8xrj9" podUID="58986885-b2ff-450a-b232-a26163de811a" containerName="registry-server" containerID="cri-o://2af6cb2b9380482b4ea542096a63d2330cf2514fce0a1df0deaa98cec712ba02" gracePeriod=2 Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.478324 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.594770 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24rgb\" (UniqueName: \"kubernetes.io/projected/58986885-b2ff-450a-b232-a26163de811a-kube-api-access-24rgb\") pod \"58986885-b2ff-450a-b232-a26163de811a\" (UID: \"58986885-b2ff-450a-b232-a26163de811a\") " Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.594973 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58986885-b2ff-450a-b232-a26163de811a-utilities\") pod \"58986885-b2ff-450a-b232-a26163de811a\" (UID: \"58986885-b2ff-450a-b232-a26163de811a\") " Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.595134 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58986885-b2ff-450a-b232-a26163de811a-catalog-content\") pod \"58986885-b2ff-450a-b232-a26163de811a\" (UID: \"58986885-b2ff-450a-b232-a26163de811a\") " Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.595536 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58986885-b2ff-450a-b232-a26163de811a-utilities" (OuterVolumeSpecName: "utilities") pod "58986885-b2ff-450a-b232-a26163de811a" (UID: "58986885-b2ff-450a-b232-a26163de811a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.595906 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/58986885-b2ff-450a-b232-a26163de811a-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.621152 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58986885-b2ff-450a-b232-a26163de811a-kube-api-access-24rgb" (OuterVolumeSpecName: "kube-api-access-24rgb") pod "58986885-b2ff-450a-b232-a26163de811a" (UID: "58986885-b2ff-450a-b232-a26163de811a"). InnerVolumeSpecName "kube-api-access-24rgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.669722 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58986885-b2ff-450a-b232-a26163de811a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "58986885-b2ff-450a-b232-a26163de811a" (UID: "58986885-b2ff-450a-b232-a26163de811a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.697249 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/58986885-b2ff-450a-b232-a26163de811a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.697279 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24rgb\" (UniqueName: \"kubernetes.io/projected/58986885-b2ff-450a-b232-a26163de811a-kube-api-access-24rgb\") on node \"crc\" DevicePath \"\"" Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.918548 4632 generic.go:334] "Generic (PLEG): container finished" podID="58986885-b2ff-450a-b232-a26163de811a" containerID="2af6cb2b9380482b4ea542096a63d2330cf2514fce0a1df0deaa98cec712ba02" exitCode=0 Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.918728 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8xrj9" event={"ID":"58986885-b2ff-450a-b232-a26163de811a","Type":"ContainerDied","Data":"2af6cb2b9380482b4ea542096a63d2330cf2514fce0a1df0deaa98cec712ba02"} Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.918793 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8xrj9" Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.918868 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8xrj9" event={"ID":"58986885-b2ff-450a-b232-a26163de811a","Type":"ContainerDied","Data":"e7db2df784e22387bb91352a6bc88e575028a70869b8cf3ee15bff73c9b7bfb4"} Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.918897 4632 scope.go:117] "RemoveContainer" containerID="2af6cb2b9380482b4ea542096a63d2330cf2514fce0a1df0deaa98cec712ba02" Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.945776 4632 scope.go:117] "RemoveContainer" containerID="2d9f6c616f332d122081782eff3ae0b45da85be1a8961803eef5bdd4c8ed6d19" Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.965628 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8xrj9"] Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.977778 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8xrj9"] Mar 13 12:13:16 crc kubenswrapper[4632]: I0313 12:13:16.991387 4632 scope.go:117] "RemoveContainer" containerID="b27af3f83423049c58aa5949bc99cebcf9fba37e5504bb6ed378f00293a5e3ff" Mar 13 12:13:17 crc kubenswrapper[4632]: I0313 12:13:17.033179 4632 scope.go:117] "RemoveContainer" containerID="2af6cb2b9380482b4ea542096a63d2330cf2514fce0a1df0deaa98cec712ba02" Mar 13 12:13:17 crc kubenswrapper[4632]: E0313 12:13:17.038180 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2af6cb2b9380482b4ea542096a63d2330cf2514fce0a1df0deaa98cec712ba02\": container with ID starting with 2af6cb2b9380482b4ea542096a63d2330cf2514fce0a1df0deaa98cec712ba02 not found: ID does not exist" containerID="2af6cb2b9380482b4ea542096a63d2330cf2514fce0a1df0deaa98cec712ba02" Mar 13 12:13:17 crc kubenswrapper[4632]: I0313 12:13:17.038237 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2af6cb2b9380482b4ea542096a63d2330cf2514fce0a1df0deaa98cec712ba02"} err="failed to get container status \"2af6cb2b9380482b4ea542096a63d2330cf2514fce0a1df0deaa98cec712ba02\": rpc error: code = NotFound desc = could not find container \"2af6cb2b9380482b4ea542096a63d2330cf2514fce0a1df0deaa98cec712ba02\": container with ID starting with 2af6cb2b9380482b4ea542096a63d2330cf2514fce0a1df0deaa98cec712ba02 not found: ID does not exist" Mar 13 12:13:17 crc kubenswrapper[4632]: I0313 12:13:17.038410 4632 scope.go:117] "RemoveContainer" containerID="2d9f6c616f332d122081782eff3ae0b45da85be1a8961803eef5bdd4c8ed6d19" Mar 13 12:13:17 crc kubenswrapper[4632]: E0313 12:13:17.038869 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d9f6c616f332d122081782eff3ae0b45da85be1a8961803eef5bdd4c8ed6d19\": container with ID starting with 2d9f6c616f332d122081782eff3ae0b45da85be1a8961803eef5bdd4c8ed6d19 not found: ID does not exist" containerID="2d9f6c616f332d122081782eff3ae0b45da85be1a8961803eef5bdd4c8ed6d19" Mar 13 12:13:17 crc kubenswrapper[4632]: I0313 12:13:17.038907 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d9f6c616f332d122081782eff3ae0b45da85be1a8961803eef5bdd4c8ed6d19"} err="failed to get container status \"2d9f6c616f332d122081782eff3ae0b45da85be1a8961803eef5bdd4c8ed6d19\": rpc error: code = NotFound desc = could not find container \"2d9f6c616f332d122081782eff3ae0b45da85be1a8961803eef5bdd4c8ed6d19\": container with ID starting with 2d9f6c616f332d122081782eff3ae0b45da85be1a8961803eef5bdd4c8ed6d19 not found: ID does not exist" Mar 13 12:13:17 crc kubenswrapper[4632]: I0313 12:13:17.038928 4632 scope.go:117] "RemoveContainer" containerID="b27af3f83423049c58aa5949bc99cebcf9fba37e5504bb6ed378f00293a5e3ff" Mar 13 12:13:17 crc kubenswrapper[4632]: E0313 12:13:17.039243 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b27af3f83423049c58aa5949bc99cebcf9fba37e5504bb6ed378f00293a5e3ff\": container with ID starting with b27af3f83423049c58aa5949bc99cebcf9fba37e5504bb6ed378f00293a5e3ff not found: ID does not exist" containerID="b27af3f83423049c58aa5949bc99cebcf9fba37e5504bb6ed378f00293a5e3ff" Mar 13 12:13:17 crc kubenswrapper[4632]: I0313 12:13:17.039273 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b27af3f83423049c58aa5949bc99cebcf9fba37e5504bb6ed378f00293a5e3ff"} err="failed to get container status \"b27af3f83423049c58aa5949bc99cebcf9fba37e5504bb6ed378f00293a5e3ff\": rpc error: code = NotFound desc = could not find container \"b27af3f83423049c58aa5949bc99cebcf9fba37e5504bb6ed378f00293a5e3ff\": container with ID starting with b27af3f83423049c58aa5949bc99cebcf9fba37e5504bb6ed378f00293a5e3ff not found: ID does not exist" Mar 13 12:13:18 crc kubenswrapper[4632]: I0313 12:13:18.057904 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58986885-b2ff-450a-b232-a26163de811a" path="/var/lib/kubelet/pods/58986885-b2ff-450a-b232-a26163de811a/volumes" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.170811 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-82xv7"] Mar 13 12:13:26 crc kubenswrapper[4632]: E0313 12:13:26.172265 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58986885-b2ff-450a-b232-a26163de811a" containerName="registry-server" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.172283 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="58986885-b2ff-450a-b232-a26163de811a" containerName="registry-server" Mar 13 12:13:26 crc kubenswrapper[4632]: E0313 12:13:26.172314 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58986885-b2ff-450a-b232-a26163de811a" containerName="extract-utilities" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.172321 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="58986885-b2ff-450a-b232-a26163de811a" containerName="extract-utilities" Mar 13 12:13:26 crc kubenswrapper[4632]: E0313 12:13:26.172330 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58986885-b2ff-450a-b232-a26163de811a" containerName="extract-content" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.172338 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="58986885-b2ff-450a-b232-a26163de811a" containerName="extract-content" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.172597 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="58986885-b2ff-450a-b232-a26163de811a" containerName="registry-server" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.174486 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.186107 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-82xv7"] Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.305830 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/356af638-6aee-4a8f-996e-04eda41f3c75-utilities\") pod \"redhat-operators-82xv7\" (UID: \"356af638-6aee-4a8f-996e-04eda41f3c75\") " pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.305872 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/356af638-6aee-4a8f-996e-04eda41f3c75-catalog-content\") pod \"redhat-operators-82xv7\" (UID: \"356af638-6aee-4a8f-996e-04eda41f3c75\") " pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.306055 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f497x\" (UniqueName: \"kubernetes.io/projected/356af638-6aee-4a8f-996e-04eda41f3c75-kube-api-access-f497x\") pod \"redhat-operators-82xv7\" (UID: \"356af638-6aee-4a8f-996e-04eda41f3c75\") " pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.407501 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f497x\" (UniqueName: \"kubernetes.io/projected/356af638-6aee-4a8f-996e-04eda41f3c75-kube-api-access-f497x\") pod \"redhat-operators-82xv7\" (UID: \"356af638-6aee-4a8f-996e-04eda41f3c75\") " pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.407586 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/356af638-6aee-4a8f-996e-04eda41f3c75-utilities\") pod \"redhat-operators-82xv7\" (UID: \"356af638-6aee-4a8f-996e-04eda41f3c75\") " pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.407614 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/356af638-6aee-4a8f-996e-04eda41f3c75-catalog-content\") pod \"redhat-operators-82xv7\" (UID: \"356af638-6aee-4a8f-996e-04eda41f3c75\") " pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.408735 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/356af638-6aee-4a8f-996e-04eda41f3c75-catalog-content\") pod \"redhat-operators-82xv7\" (UID: \"356af638-6aee-4a8f-996e-04eda41f3c75\") " pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.409063 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/356af638-6aee-4a8f-996e-04eda41f3c75-utilities\") pod \"redhat-operators-82xv7\" (UID: \"356af638-6aee-4a8f-996e-04eda41f3c75\") " pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.434674 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f497x\" (UniqueName: \"kubernetes.io/projected/356af638-6aee-4a8f-996e-04eda41f3c75-kube-api-access-f497x\") pod \"redhat-operators-82xv7\" (UID: \"356af638-6aee-4a8f-996e-04eda41f3c75\") " pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.494376 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:13:26 crc kubenswrapper[4632]: I0313 12:13:26.989880 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-82xv7"] Mar 13 12:13:27 crc kubenswrapper[4632]: I0313 12:13:27.027454 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82xv7" event={"ID":"356af638-6aee-4a8f-996e-04eda41f3c75","Type":"ContainerStarted","Data":"b62d309f4dfbdb74ea9d5a23db29bddea476b47f846e66dc66702ff94689f734"} Mar 13 12:13:28 crc kubenswrapper[4632]: I0313 12:13:28.048591 4632 generic.go:334] "Generic (PLEG): container finished" podID="356af638-6aee-4a8f-996e-04eda41f3c75" containerID="37e01b8e98a8f4df777413f141a70a28ede73377dfa9804760ceab29634fccc3" exitCode=0 Mar 13 12:13:28 crc kubenswrapper[4632]: I0313 12:13:28.060296 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82xv7" event={"ID":"356af638-6aee-4a8f-996e-04eda41f3c75","Type":"ContainerDied","Data":"37e01b8e98a8f4df777413f141a70a28ede73377dfa9804760ceab29634fccc3"} Mar 13 12:13:28 crc kubenswrapper[4632]: I0313 12:13:28.064435 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 12:13:30 crc kubenswrapper[4632]: I0313 12:13:30.106027 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82xv7" event={"ID":"356af638-6aee-4a8f-996e-04eda41f3c75","Type":"ContainerStarted","Data":"2c91311b1f55ad95b395eb30b7c8a1dcb54c8e11934e8e37b24fd616a95caf8f"} Mar 13 12:13:36 crc kubenswrapper[4632]: I0313 12:13:36.169813 4632 generic.go:334] "Generic (PLEG): container finished" podID="356af638-6aee-4a8f-996e-04eda41f3c75" containerID="2c91311b1f55ad95b395eb30b7c8a1dcb54c8e11934e8e37b24fd616a95caf8f" exitCode=0 Mar 13 12:13:36 crc kubenswrapper[4632]: I0313 12:13:36.170036 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82xv7" event={"ID":"356af638-6aee-4a8f-996e-04eda41f3c75","Type":"ContainerDied","Data":"2c91311b1f55ad95b395eb30b7c8a1dcb54c8e11934e8e37b24fd616a95caf8f"} Mar 13 12:13:37 crc kubenswrapper[4632]: I0313 12:13:37.182256 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82xv7" event={"ID":"356af638-6aee-4a8f-996e-04eda41f3c75","Type":"ContainerStarted","Data":"a4ec22df9491ee29893ecc4c1c955cb46cd5e215eb52c9438822b29458df34d6"} Mar 13 12:13:37 crc kubenswrapper[4632]: I0313 12:13:37.206024 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-82xv7" podStartSLOduration=2.576909333 podStartE2EDuration="11.205997093s" podCreationTimestamp="2026-03-13 12:13:26 +0000 UTC" firstStartedPulling="2026-03-13 12:13:28.062133412 +0000 UTC m=+7782.084663545" lastFinishedPulling="2026-03-13 12:13:36.691221172 +0000 UTC m=+7790.713751305" observedRunningTime="2026-03-13 12:13:37.201262906 +0000 UTC m=+7791.223793059" watchObservedRunningTime="2026-03-13 12:13:37.205997093 +0000 UTC m=+7791.228527216" Mar 13 12:13:40 crc kubenswrapper[4632]: I0313 12:13:40.461689 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:13:40 crc kubenswrapper[4632]: I0313 12:13:40.463888 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:13:46 crc kubenswrapper[4632]: I0313 12:13:46.495476 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:13:46 crc kubenswrapper[4632]: I0313 12:13:46.496019 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:13:47 crc kubenswrapper[4632]: I0313 12:13:47.543225 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-82xv7" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" containerName="registry-server" probeResult="failure" output=< Mar 13 12:13:47 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:13:47 crc kubenswrapper[4632]: > Mar 13 12:13:57 crc kubenswrapper[4632]: I0313 12:13:57.554909 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-82xv7" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" containerName="registry-server" probeResult="failure" output=< Mar 13 12:13:57 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:13:57 crc kubenswrapper[4632]: > Mar 13 12:14:00 crc kubenswrapper[4632]: I0313 12:14:00.265969 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556734-bwv75"] Mar 13 12:14:00 crc kubenswrapper[4632]: I0313 12:14:00.268399 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556734-bwv75" Mar 13 12:14:00 crc kubenswrapper[4632]: I0313 12:14:00.292973 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:14:00 crc kubenswrapper[4632]: I0313 12:14:00.293009 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:14:00 crc kubenswrapper[4632]: I0313 12:14:00.294861 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:14:00 crc kubenswrapper[4632]: I0313 12:14:00.301531 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556734-bwv75"] Mar 13 12:14:00 crc kubenswrapper[4632]: I0313 12:14:00.332547 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxgj5\" (UniqueName: \"kubernetes.io/projected/5d0a5571-d345-44cf-ba1a-46b3ef68b1ae-kube-api-access-lxgj5\") pod \"auto-csr-approver-29556734-bwv75\" (UID: \"5d0a5571-d345-44cf-ba1a-46b3ef68b1ae\") " pod="openshift-infra/auto-csr-approver-29556734-bwv75" Mar 13 12:14:00 crc kubenswrapper[4632]: I0313 12:14:00.437097 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxgj5\" (UniqueName: \"kubernetes.io/projected/5d0a5571-d345-44cf-ba1a-46b3ef68b1ae-kube-api-access-lxgj5\") pod \"auto-csr-approver-29556734-bwv75\" (UID: \"5d0a5571-d345-44cf-ba1a-46b3ef68b1ae\") " pod="openshift-infra/auto-csr-approver-29556734-bwv75" Mar 13 12:14:00 crc kubenswrapper[4632]: I0313 12:14:00.490401 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxgj5\" (UniqueName: \"kubernetes.io/projected/5d0a5571-d345-44cf-ba1a-46b3ef68b1ae-kube-api-access-lxgj5\") pod \"auto-csr-approver-29556734-bwv75\" (UID: \"5d0a5571-d345-44cf-ba1a-46b3ef68b1ae\") " pod="openshift-infra/auto-csr-approver-29556734-bwv75" Mar 13 12:14:00 crc kubenswrapper[4632]: I0313 12:14:00.600353 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556734-bwv75" Mar 13 12:14:01 crc kubenswrapper[4632]: I0313 12:14:01.798041 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556734-bwv75"] Mar 13 12:14:01 crc kubenswrapper[4632]: W0313 12:14:01.816436 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d0a5571_d345_44cf_ba1a_46b3ef68b1ae.slice/crio-b1227e617d866c1a39d8664c23193a023ef1eeed2c0352c7b8799a2a86a8c01c WatchSource:0}: Error finding container b1227e617d866c1a39d8664c23193a023ef1eeed2c0352c7b8799a2a86a8c01c: Status 404 returned error can't find the container with id b1227e617d866c1a39d8664c23193a023ef1eeed2c0352c7b8799a2a86a8c01c Mar 13 12:14:02 crc kubenswrapper[4632]: I0313 12:14:02.436786 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556734-bwv75" event={"ID":"5d0a5571-d345-44cf-ba1a-46b3ef68b1ae","Type":"ContainerStarted","Data":"b1227e617d866c1a39d8664c23193a023ef1eeed2c0352c7b8799a2a86a8c01c"} Mar 13 12:14:04 crc kubenswrapper[4632]: I0313 12:14:04.472607 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556734-bwv75" event={"ID":"5d0a5571-d345-44cf-ba1a-46b3ef68b1ae","Type":"ContainerStarted","Data":"6faf20a0273a34b58a9864d2f460cdaebfb4cad46108ca19c20cf494270e4fe6"} Mar 13 12:14:04 crc kubenswrapper[4632]: I0313 12:14:04.488358 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556734-bwv75" podStartSLOduration=3.436713212 podStartE2EDuration="4.488338958s" podCreationTimestamp="2026-03-13 12:14:00 +0000 UTC" firstStartedPulling="2026-03-13 12:14:01.82010712 +0000 UTC m=+7815.842637253" lastFinishedPulling="2026-03-13 12:14:02.871732866 +0000 UTC m=+7816.894262999" observedRunningTime="2026-03-13 12:14:04.485446037 +0000 UTC m=+7818.507976170" watchObservedRunningTime="2026-03-13 12:14:04.488338958 +0000 UTC m=+7818.510869091" Mar 13 12:14:06 crc kubenswrapper[4632]: I0313 12:14:06.492849 4632 generic.go:334] "Generic (PLEG): container finished" podID="5d0a5571-d345-44cf-ba1a-46b3ef68b1ae" containerID="6faf20a0273a34b58a9864d2f460cdaebfb4cad46108ca19c20cf494270e4fe6" exitCode=0 Mar 13 12:14:06 crc kubenswrapper[4632]: I0313 12:14:06.492916 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556734-bwv75" event={"ID":"5d0a5571-d345-44cf-ba1a-46b3ef68b1ae","Type":"ContainerDied","Data":"6faf20a0273a34b58a9864d2f460cdaebfb4cad46108ca19c20cf494270e4fe6"} Mar 13 12:14:07 crc kubenswrapper[4632]: I0313 12:14:07.565263 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-82xv7" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" containerName="registry-server" probeResult="failure" output=< Mar 13 12:14:07 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:14:07 crc kubenswrapper[4632]: > Mar 13 12:14:08 crc kubenswrapper[4632]: I0313 12:14:08.067390 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556734-bwv75" Mar 13 12:14:08 crc kubenswrapper[4632]: I0313 12:14:08.197841 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lxgj5\" (UniqueName: \"kubernetes.io/projected/5d0a5571-d345-44cf-ba1a-46b3ef68b1ae-kube-api-access-lxgj5\") pod \"5d0a5571-d345-44cf-ba1a-46b3ef68b1ae\" (UID: \"5d0a5571-d345-44cf-ba1a-46b3ef68b1ae\") " Mar 13 12:14:08 crc kubenswrapper[4632]: I0313 12:14:08.224071 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d0a5571-d345-44cf-ba1a-46b3ef68b1ae-kube-api-access-lxgj5" (OuterVolumeSpecName: "kube-api-access-lxgj5") pod "5d0a5571-d345-44cf-ba1a-46b3ef68b1ae" (UID: "5d0a5571-d345-44cf-ba1a-46b3ef68b1ae"). InnerVolumeSpecName "kube-api-access-lxgj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:14:08 crc kubenswrapper[4632]: I0313 12:14:08.300876 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lxgj5\" (UniqueName: \"kubernetes.io/projected/5d0a5571-d345-44cf-ba1a-46b3ef68b1ae-kube-api-access-lxgj5\") on node \"crc\" DevicePath \"\"" Mar 13 12:14:08 crc kubenswrapper[4632]: I0313 12:14:08.514183 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556734-bwv75" event={"ID":"5d0a5571-d345-44cf-ba1a-46b3ef68b1ae","Type":"ContainerDied","Data":"b1227e617d866c1a39d8664c23193a023ef1eeed2c0352c7b8799a2a86a8c01c"} Mar 13 12:14:08 crc kubenswrapper[4632]: I0313 12:14:08.514250 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556734-bwv75" Mar 13 12:14:08 crc kubenswrapper[4632]: I0313 12:14:08.515488 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b1227e617d866c1a39d8664c23193a023ef1eeed2c0352c7b8799a2a86a8c01c" Mar 13 12:14:08 crc kubenswrapper[4632]: I0313 12:14:08.621318 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556728-lnbz4"] Mar 13 12:14:08 crc kubenswrapper[4632]: I0313 12:14:08.629555 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556728-lnbz4"] Mar 13 12:14:10 crc kubenswrapper[4632]: I0313 12:14:10.056630 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4980ff16-68a2-4b11-83d2-9d8ad1fa105c" path="/var/lib/kubelet/pods/4980ff16-68a2-4b11-83d2-9d8ad1fa105c/volumes" Mar 13 12:14:10 crc kubenswrapper[4632]: I0313 12:14:10.460985 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:14:10 crc kubenswrapper[4632]: I0313 12:14:10.461051 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:14:17 crc kubenswrapper[4632]: I0313 12:14:17.563795 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-82xv7" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" containerName="registry-server" probeResult="failure" output=< Mar 13 12:14:17 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:14:17 crc kubenswrapper[4632]: > Mar 13 12:14:25 crc kubenswrapper[4632]: I0313 12:14:25.233540 4632 scope.go:117] "RemoveContainer" containerID="f67ad12914918a2b5742053c25b75fbf60ba190fadc812d9c25ad10140a8556c" Mar 13 12:14:27 crc kubenswrapper[4632]: I0313 12:14:27.543410 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-82xv7" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" containerName="registry-server" probeResult="failure" output=< Mar 13 12:14:27 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:14:27 crc kubenswrapper[4632]: > Mar 13 12:14:37 crc kubenswrapper[4632]: I0313 12:14:37.552437 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-82xv7" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" containerName="registry-server" probeResult="failure" output=< Mar 13 12:14:37 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:14:37 crc kubenswrapper[4632]: > Mar 13 12:14:40 crc kubenswrapper[4632]: I0313 12:14:40.460885 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:14:40 crc kubenswrapper[4632]: I0313 12:14:40.462193 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:14:40 crc kubenswrapper[4632]: I0313 12:14:40.464180 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 12:14:40 crc kubenswrapper[4632]: I0313 12:14:40.468280 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 12:14:40 crc kubenswrapper[4632]: I0313 12:14:40.469902 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" gracePeriod=600 Mar 13 12:14:40 crc kubenswrapper[4632]: E0313 12:14:40.603731 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:14:40 crc kubenswrapper[4632]: I0313 12:14:40.837020 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" exitCode=0 Mar 13 12:14:40 crc kubenswrapper[4632]: I0313 12:14:40.837113 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9"} Mar 13 12:14:40 crc kubenswrapper[4632]: I0313 12:14:40.838633 4632 scope.go:117] "RemoveContainer" containerID="ef6a755da94d8b26aaa61b1a356ec9030e87ec1440f6bdf1f6abec8411efbdd9" Mar 13 12:14:40 crc kubenswrapper[4632]: I0313 12:14:40.840745 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:14:40 crc kubenswrapper[4632]: E0313 12:14:40.841355 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:14:46 crc kubenswrapper[4632]: I0313 12:14:46.567378 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:14:46 crc kubenswrapper[4632]: I0313 12:14:46.622735 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:14:46 crc kubenswrapper[4632]: I0313 12:14:46.815886 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-82xv7"] Mar 13 12:14:47 crc kubenswrapper[4632]: I0313 12:14:47.924359 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-82xv7" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" containerName="registry-server" containerID="cri-o://a4ec22df9491ee29893ecc4c1c955cb46cd5e215eb52c9438822b29458df34d6" gracePeriod=2 Mar 13 12:14:48 crc kubenswrapper[4632]: I0313 12:14:48.926315 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:14:48 crc kubenswrapper[4632]: I0313 12:14:48.934029 4632 generic.go:334] "Generic (PLEG): container finished" podID="356af638-6aee-4a8f-996e-04eda41f3c75" containerID="a4ec22df9491ee29893ecc4c1c955cb46cd5e215eb52c9438822b29458df34d6" exitCode=0 Mar 13 12:14:48 crc kubenswrapper[4632]: I0313 12:14:48.934067 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-82xv7" Mar 13 12:14:48 crc kubenswrapper[4632]: I0313 12:14:48.934091 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82xv7" event={"ID":"356af638-6aee-4a8f-996e-04eda41f3c75","Type":"ContainerDied","Data":"a4ec22df9491ee29893ecc4c1c955cb46cd5e215eb52c9438822b29458df34d6"} Mar 13 12:14:48 crc kubenswrapper[4632]: I0313 12:14:48.934151 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-82xv7" event={"ID":"356af638-6aee-4a8f-996e-04eda41f3c75","Type":"ContainerDied","Data":"b62d309f4dfbdb74ea9d5a23db29bddea476b47f846e66dc66702ff94689f734"} Mar 13 12:14:48 crc kubenswrapper[4632]: I0313 12:14:48.934171 4632 scope.go:117] "RemoveContainer" containerID="a4ec22df9491ee29893ecc4c1c955cb46cd5e215eb52c9438822b29458df34d6" Mar 13 12:14:48 crc kubenswrapper[4632]: I0313 12:14:48.962554 4632 scope.go:117] "RemoveContainer" containerID="2c91311b1f55ad95b395eb30b7c8a1dcb54c8e11934e8e37b24fd616a95caf8f" Mar 13 12:14:48 crc kubenswrapper[4632]: I0313 12:14:48.997828 4632 scope.go:117] "RemoveContainer" containerID="37e01b8e98a8f4df777413f141a70a28ede73377dfa9804760ceab29634fccc3" Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.047222 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/356af638-6aee-4a8f-996e-04eda41f3c75-catalog-content\") pod \"356af638-6aee-4a8f-996e-04eda41f3c75\" (UID: \"356af638-6aee-4a8f-996e-04eda41f3c75\") " Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.047375 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/356af638-6aee-4a8f-996e-04eda41f3c75-utilities\") pod \"356af638-6aee-4a8f-996e-04eda41f3c75\" (UID: \"356af638-6aee-4a8f-996e-04eda41f3c75\") " Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.047464 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f497x\" (UniqueName: \"kubernetes.io/projected/356af638-6aee-4a8f-996e-04eda41f3c75-kube-api-access-f497x\") pod \"356af638-6aee-4a8f-996e-04eda41f3c75\" (UID: \"356af638-6aee-4a8f-996e-04eda41f3c75\") " Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.054324 4632 scope.go:117] "RemoveContainer" containerID="a4ec22df9491ee29893ecc4c1c955cb46cd5e215eb52c9438822b29458df34d6" Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.057725 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/356af638-6aee-4a8f-996e-04eda41f3c75-utilities" (OuterVolumeSpecName: "utilities") pod "356af638-6aee-4a8f-996e-04eda41f3c75" (UID: "356af638-6aee-4a8f-996e-04eda41f3c75"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:14:49 crc kubenswrapper[4632]: E0313 12:14:49.060025 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4ec22df9491ee29893ecc4c1c955cb46cd5e215eb52c9438822b29458df34d6\": container with ID starting with a4ec22df9491ee29893ecc4c1c955cb46cd5e215eb52c9438822b29458df34d6 not found: ID does not exist" containerID="a4ec22df9491ee29893ecc4c1c955cb46cd5e215eb52c9438822b29458df34d6" Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.060093 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4ec22df9491ee29893ecc4c1c955cb46cd5e215eb52c9438822b29458df34d6"} err="failed to get container status \"a4ec22df9491ee29893ecc4c1c955cb46cd5e215eb52c9438822b29458df34d6\": rpc error: code = NotFound desc = could not find container \"a4ec22df9491ee29893ecc4c1c955cb46cd5e215eb52c9438822b29458df34d6\": container with ID starting with a4ec22df9491ee29893ecc4c1c955cb46cd5e215eb52c9438822b29458df34d6 not found: ID does not exist" Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.060119 4632 scope.go:117] "RemoveContainer" containerID="2c91311b1f55ad95b395eb30b7c8a1dcb54c8e11934e8e37b24fd616a95caf8f" Mar 13 12:14:49 crc kubenswrapper[4632]: E0313 12:14:49.060739 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2c91311b1f55ad95b395eb30b7c8a1dcb54c8e11934e8e37b24fd616a95caf8f\": container with ID starting with 2c91311b1f55ad95b395eb30b7c8a1dcb54c8e11934e8e37b24fd616a95caf8f not found: ID does not exist" containerID="2c91311b1f55ad95b395eb30b7c8a1dcb54c8e11934e8e37b24fd616a95caf8f" Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.060777 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2c91311b1f55ad95b395eb30b7c8a1dcb54c8e11934e8e37b24fd616a95caf8f"} err="failed to get container status \"2c91311b1f55ad95b395eb30b7c8a1dcb54c8e11934e8e37b24fd616a95caf8f\": rpc error: code = NotFound desc = could not find container \"2c91311b1f55ad95b395eb30b7c8a1dcb54c8e11934e8e37b24fd616a95caf8f\": container with ID starting with 2c91311b1f55ad95b395eb30b7c8a1dcb54c8e11934e8e37b24fd616a95caf8f not found: ID does not exist" Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.060798 4632 scope.go:117] "RemoveContainer" containerID="37e01b8e98a8f4df777413f141a70a28ede73377dfa9804760ceab29634fccc3" Mar 13 12:14:49 crc kubenswrapper[4632]: E0313 12:14:49.061203 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37e01b8e98a8f4df777413f141a70a28ede73377dfa9804760ceab29634fccc3\": container with ID starting with 37e01b8e98a8f4df777413f141a70a28ede73377dfa9804760ceab29634fccc3 not found: ID does not exist" containerID="37e01b8e98a8f4df777413f141a70a28ede73377dfa9804760ceab29634fccc3" Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.061236 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37e01b8e98a8f4df777413f141a70a28ede73377dfa9804760ceab29634fccc3"} err="failed to get container status \"37e01b8e98a8f4df777413f141a70a28ede73377dfa9804760ceab29634fccc3\": rpc error: code = NotFound desc = could not find container \"37e01b8e98a8f4df777413f141a70a28ede73377dfa9804760ceab29634fccc3\": container with ID starting with 37e01b8e98a8f4df777413f141a70a28ede73377dfa9804760ceab29634fccc3 not found: ID does not exist" Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.078804 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/356af638-6aee-4a8f-996e-04eda41f3c75-kube-api-access-f497x" (OuterVolumeSpecName: "kube-api-access-f497x") pod "356af638-6aee-4a8f-996e-04eda41f3c75" (UID: "356af638-6aee-4a8f-996e-04eda41f3c75"). InnerVolumeSpecName "kube-api-access-f497x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.152494 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/356af638-6aee-4a8f-996e-04eda41f3c75-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.152597 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f497x\" (UniqueName: \"kubernetes.io/projected/356af638-6aee-4a8f-996e-04eda41f3c75-kube-api-access-f497x\") on node \"crc\" DevicePath \"\"" Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.238152 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/356af638-6aee-4a8f-996e-04eda41f3c75-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "356af638-6aee-4a8f-996e-04eda41f3c75" (UID: "356af638-6aee-4a8f-996e-04eda41f3c75"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.254526 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/356af638-6aee-4a8f-996e-04eda41f3c75-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.620001 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-82xv7"] Mar 13 12:14:49 crc kubenswrapper[4632]: I0313 12:14:49.665744 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-82xv7"] Mar 13 12:14:50 crc kubenswrapper[4632]: I0313 12:14:50.057089 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" path="/var/lib/kubelet/pods/356af638-6aee-4a8f-996e-04eda41f3c75/volumes" Mar 13 12:14:53 crc kubenswrapper[4632]: E0313 12:14:53.109727 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/system.slice/rpm-ostreed.service\": RecentStats: unable to find data in memory cache]" Mar 13 12:14:54 crc kubenswrapper[4632]: I0313 12:14:54.044221 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:14:54 crc kubenswrapper[4632]: E0313 12:14:54.044741 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.179644 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj"] Mar 13 12:15:00 crc kubenswrapper[4632]: E0313 12:15:00.183057 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" containerName="extract-content" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.183087 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" containerName="extract-content" Mar 13 12:15:00 crc kubenswrapper[4632]: E0313 12:15:00.183151 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d0a5571-d345-44cf-ba1a-46b3ef68b1ae" containerName="oc" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.183159 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d0a5571-d345-44cf-ba1a-46b3ef68b1ae" containerName="oc" Mar 13 12:15:00 crc kubenswrapper[4632]: E0313 12:15:00.183184 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" containerName="extract-utilities" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.183192 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" containerName="extract-utilities" Mar 13 12:15:00 crc kubenswrapper[4632]: E0313 12:15:00.183202 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" containerName="registry-server" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.183208 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" containerName="registry-server" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.184727 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d0a5571-d345-44cf-ba1a-46b3ef68b1ae" containerName="oc" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.184758 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="356af638-6aee-4a8f-996e-04eda41f3c75" containerName="registry-server" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.192193 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.207033 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.207039 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.214622 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj"] Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.235529 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-secret-volume\") pod \"collect-profiles-29556735-8mfsj\" (UID: \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.235693 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxpjx\" (UniqueName: \"kubernetes.io/projected/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-kube-api-access-wxpjx\") pod \"collect-profiles-29556735-8mfsj\" (UID: \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.235791 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-config-volume\") pod \"collect-profiles-29556735-8mfsj\" (UID: \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.338183 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxpjx\" (UniqueName: \"kubernetes.io/projected/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-kube-api-access-wxpjx\") pod \"collect-profiles-29556735-8mfsj\" (UID: \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.338281 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-config-volume\") pod \"collect-profiles-29556735-8mfsj\" (UID: \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.338323 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-secret-volume\") pod \"collect-profiles-29556735-8mfsj\" (UID: \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.339262 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-config-volume\") pod \"collect-profiles-29556735-8mfsj\" (UID: \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.350485 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-secret-volume\") pod \"collect-profiles-29556735-8mfsj\" (UID: \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.371262 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxpjx\" (UniqueName: \"kubernetes.io/projected/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-kube-api-access-wxpjx\") pod \"collect-profiles-29556735-8mfsj\" (UID: \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" Mar 13 12:15:00 crc kubenswrapper[4632]: I0313 12:15:00.534344 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" Mar 13 12:15:01 crc kubenswrapper[4632]: I0313 12:15:01.164931 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj"] Mar 13 12:15:02 crc kubenswrapper[4632]: I0313 12:15:02.075023 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" event={"ID":"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3","Type":"ContainerStarted","Data":"9b0bafda64d1039901c8fa27c0404ac16cc9d6c17ec7e7294c6e46712544cdb4"} Mar 13 12:15:02 crc kubenswrapper[4632]: I0313 12:15:02.076218 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" event={"ID":"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3","Type":"ContainerStarted","Data":"fb8a8ee86021160eb679abf8d44c17e744720b3d880df97becd5875a7084ef14"} Mar 13 12:15:03 crc kubenswrapper[4632]: I0313 12:15:03.085459 4632 generic.go:334] "Generic (PLEG): container finished" podID="ee86cbb5-0041-46eb-8f35-c159f9fbc3b3" containerID="9b0bafda64d1039901c8fa27c0404ac16cc9d6c17ec7e7294c6e46712544cdb4" exitCode=0 Mar 13 12:15:03 crc kubenswrapper[4632]: I0313 12:15:03.085530 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" event={"ID":"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3","Type":"ContainerDied","Data":"9b0bafda64d1039901c8fa27c0404ac16cc9d6c17ec7e7294c6e46712544cdb4"} Mar 13 12:15:04 crc kubenswrapper[4632]: I0313 12:15:04.583602 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" Mar 13 12:15:04 crc kubenswrapper[4632]: I0313 12:15:04.638138 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-config-volume\") pod \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\" (UID: \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\") " Mar 13 12:15:04 crc kubenswrapper[4632]: I0313 12:15:04.638181 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxpjx\" (UniqueName: \"kubernetes.io/projected/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-kube-api-access-wxpjx\") pod \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\" (UID: \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\") " Mar 13 12:15:04 crc kubenswrapper[4632]: I0313 12:15:04.638308 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-secret-volume\") pod \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\" (UID: \"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3\") " Mar 13 12:15:04 crc kubenswrapper[4632]: I0313 12:15:04.638920 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-config-volume" (OuterVolumeSpecName: "config-volume") pod "ee86cbb5-0041-46eb-8f35-c159f9fbc3b3" (UID: "ee86cbb5-0041-46eb-8f35-c159f9fbc3b3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 12:15:04 crc kubenswrapper[4632]: I0313 12:15:04.644441 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ee86cbb5-0041-46eb-8f35-c159f9fbc3b3" (UID: "ee86cbb5-0041-46eb-8f35-c159f9fbc3b3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:15:04 crc kubenswrapper[4632]: I0313 12:15:04.649227 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-kube-api-access-wxpjx" (OuterVolumeSpecName: "kube-api-access-wxpjx") pod "ee86cbb5-0041-46eb-8f35-c159f9fbc3b3" (UID: "ee86cbb5-0041-46eb-8f35-c159f9fbc3b3"). InnerVolumeSpecName "kube-api-access-wxpjx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:15:04 crc kubenswrapper[4632]: I0313 12:15:04.740183 4632 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 12:15:04 crc kubenswrapper[4632]: I0313 12:15:04.740224 4632 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 12:15:04 crc kubenswrapper[4632]: I0313 12:15:04.740234 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxpjx\" (UniqueName: \"kubernetes.io/projected/ee86cbb5-0041-46eb-8f35-c159f9fbc3b3-kube-api-access-wxpjx\") on node \"crc\" DevicePath \"\"" Mar 13 12:15:05 crc kubenswrapper[4632]: I0313 12:15:05.104359 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" event={"ID":"ee86cbb5-0041-46eb-8f35-c159f9fbc3b3","Type":"ContainerDied","Data":"fb8a8ee86021160eb679abf8d44c17e744720b3d880df97becd5875a7084ef14"} Mar 13 12:15:05 crc kubenswrapper[4632]: I0313 12:15:05.104391 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556735-8mfsj" Mar 13 12:15:05 crc kubenswrapper[4632]: I0313 12:15:05.104403 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fb8a8ee86021160eb679abf8d44c17e744720b3d880df97becd5875a7084ef14" Mar 13 12:15:05 crc kubenswrapper[4632]: I0313 12:15:05.678268 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh"] Mar 13 12:15:05 crc kubenswrapper[4632]: I0313 12:15:05.684744 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556690-p82jh"] Mar 13 12:15:06 crc kubenswrapper[4632]: I0313 12:15:06.055898 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9dc72a85-cdb5-4b11-9e0a-158d269edf96" path="/var/lib/kubelet/pods/9dc72a85-cdb5-4b11-9e0a-158d269edf96/volumes" Mar 13 12:15:09 crc kubenswrapper[4632]: I0313 12:15:09.045077 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:15:09 crc kubenswrapper[4632]: E0313 12:15:09.045614 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:15:21 crc kubenswrapper[4632]: I0313 12:15:21.050590 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:15:21 crc kubenswrapper[4632]: E0313 12:15:21.051366 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:15:25 crc kubenswrapper[4632]: I0313 12:15:25.484341 4632 scope.go:117] "RemoveContainer" containerID="526c0b7d143109242f29250c0cffd4a40f383eaf78da9d0786f09bf0aa0eccb3" Mar 13 12:15:33 crc kubenswrapper[4632]: I0313 12:15:33.044066 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:15:33 crc kubenswrapper[4632]: E0313 12:15:33.044789 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:15:45 crc kubenswrapper[4632]: I0313 12:15:45.044212 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:15:45 crc kubenswrapper[4632]: E0313 12:15:45.045031 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:15:58 crc kubenswrapper[4632]: I0313 12:15:58.056041 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:15:58 crc kubenswrapper[4632]: E0313 12:15:58.057752 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:16:00 crc kubenswrapper[4632]: I0313 12:16:00.153342 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556736-8cfm9"] Mar 13 12:16:00 crc kubenswrapper[4632]: E0313 12:16:00.154127 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee86cbb5-0041-46eb-8f35-c159f9fbc3b3" containerName="collect-profiles" Mar 13 12:16:00 crc kubenswrapper[4632]: I0313 12:16:00.154141 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee86cbb5-0041-46eb-8f35-c159f9fbc3b3" containerName="collect-profiles" Mar 13 12:16:00 crc kubenswrapper[4632]: I0313 12:16:00.154361 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee86cbb5-0041-46eb-8f35-c159f9fbc3b3" containerName="collect-profiles" Mar 13 12:16:00 crc kubenswrapper[4632]: I0313 12:16:00.154989 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556736-8cfm9" Mar 13 12:16:00 crc kubenswrapper[4632]: I0313 12:16:00.163147 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556736-8cfm9"] Mar 13 12:16:00 crc kubenswrapper[4632]: I0313 12:16:00.165497 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:16:00 crc kubenswrapper[4632]: I0313 12:16:00.166844 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:16:00 crc kubenswrapper[4632]: I0313 12:16:00.167320 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:16:00 crc kubenswrapper[4632]: I0313 12:16:00.258877 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zfh5\" (UniqueName: \"kubernetes.io/projected/d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e-kube-api-access-9zfh5\") pod \"auto-csr-approver-29556736-8cfm9\" (UID: \"d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e\") " pod="openshift-infra/auto-csr-approver-29556736-8cfm9" Mar 13 12:16:00 crc kubenswrapper[4632]: I0313 12:16:00.362249 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zfh5\" (UniqueName: \"kubernetes.io/projected/d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e-kube-api-access-9zfh5\") pod \"auto-csr-approver-29556736-8cfm9\" (UID: \"d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e\") " pod="openshift-infra/auto-csr-approver-29556736-8cfm9" Mar 13 12:16:00 crc kubenswrapper[4632]: I0313 12:16:00.391130 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zfh5\" (UniqueName: \"kubernetes.io/projected/d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e-kube-api-access-9zfh5\") pod \"auto-csr-approver-29556736-8cfm9\" (UID: \"d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e\") " pod="openshift-infra/auto-csr-approver-29556736-8cfm9" Mar 13 12:16:00 crc kubenswrapper[4632]: I0313 12:16:00.478598 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556736-8cfm9" Mar 13 12:16:01 crc kubenswrapper[4632]: I0313 12:16:01.007199 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556736-8cfm9"] Mar 13 12:16:01 crc kubenswrapper[4632]: I0313 12:16:01.664357 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556736-8cfm9" event={"ID":"d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e","Type":"ContainerStarted","Data":"3e1d631a8dd69de33bc2cd49fed751c78109691b52c716f91c7edaac078f7b75"} Mar 13 12:16:03 crc kubenswrapper[4632]: I0313 12:16:03.686136 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556736-8cfm9" event={"ID":"d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e","Type":"ContainerStarted","Data":"d94c02e7a4464d40b765382224dccb8b8b7d7ff087751a6cc11225fceac593a0"} Mar 13 12:16:03 crc kubenswrapper[4632]: I0313 12:16:03.715756 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556736-8cfm9" podStartSLOduration=2.767809073 podStartE2EDuration="3.714873247s" podCreationTimestamp="2026-03-13 12:16:00 +0000 UTC" firstStartedPulling="2026-03-13 12:16:01.018915338 +0000 UTC m=+7935.041445471" lastFinishedPulling="2026-03-13 12:16:01.965979512 +0000 UTC m=+7935.988509645" observedRunningTime="2026-03-13 12:16:03.702473782 +0000 UTC m=+7937.725003915" watchObservedRunningTime="2026-03-13 12:16:03.714873247 +0000 UTC m=+7937.737403390" Mar 13 12:16:05 crc kubenswrapper[4632]: I0313 12:16:05.703373 4632 generic.go:334] "Generic (PLEG): container finished" podID="d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e" containerID="d94c02e7a4464d40b765382224dccb8b8b7d7ff087751a6cc11225fceac593a0" exitCode=0 Mar 13 12:16:05 crc kubenswrapper[4632]: I0313 12:16:05.703452 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556736-8cfm9" event={"ID":"d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e","Type":"ContainerDied","Data":"d94c02e7a4464d40b765382224dccb8b8b7d7ff087751a6cc11225fceac593a0"} Mar 13 12:16:07 crc kubenswrapper[4632]: I0313 12:16:07.575525 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556736-8cfm9" Mar 13 12:16:07 crc kubenswrapper[4632]: I0313 12:16:07.714753 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zfh5\" (UniqueName: \"kubernetes.io/projected/d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e-kube-api-access-9zfh5\") pod \"d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e\" (UID: \"d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e\") " Mar 13 12:16:07 crc kubenswrapper[4632]: I0313 12:16:07.721187 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e-kube-api-access-9zfh5" (OuterVolumeSpecName: "kube-api-access-9zfh5") pod "d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e" (UID: "d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e"). InnerVolumeSpecName "kube-api-access-9zfh5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:16:07 crc kubenswrapper[4632]: I0313 12:16:07.728838 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556736-8cfm9" event={"ID":"d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e","Type":"ContainerDied","Data":"3e1d631a8dd69de33bc2cd49fed751c78109691b52c716f91c7edaac078f7b75"} Mar 13 12:16:07 crc kubenswrapper[4632]: I0313 12:16:07.728879 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e1d631a8dd69de33bc2cd49fed751c78109691b52c716f91c7edaac078f7b75" Mar 13 12:16:07 crc kubenswrapper[4632]: I0313 12:16:07.728965 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556736-8cfm9" Mar 13 12:16:07 crc kubenswrapper[4632]: I0313 12:16:07.818039 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zfh5\" (UniqueName: \"kubernetes.io/projected/d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e-kube-api-access-9zfh5\") on node \"crc\" DevicePath \"\"" Mar 13 12:16:08 crc kubenswrapper[4632]: I0313 12:16:08.662293 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556730-txd8w"] Mar 13 12:16:08 crc kubenswrapper[4632]: I0313 12:16:08.681599 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556730-txd8w"] Mar 13 12:16:10 crc kubenswrapper[4632]: I0313 12:16:10.060804 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4d4ab40-a5ab-4b39-b19e-043766174116" path="/var/lib/kubelet/pods/c4d4ab40-a5ab-4b39-b19e-043766174116/volumes" Mar 13 12:16:13 crc kubenswrapper[4632]: I0313 12:16:13.044604 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:16:13 crc kubenswrapper[4632]: E0313 12:16:13.045480 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:16:25 crc kubenswrapper[4632]: I0313 12:16:25.636015 4632 scope.go:117] "RemoveContainer" containerID="47f2264b87b9b8761c8013ae1aa0f697b6d23277abd018e4664e7e2eed7771a3" Mar 13 12:16:28 crc kubenswrapper[4632]: I0313 12:16:28.054340 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:16:28 crc kubenswrapper[4632]: E0313 12:16:28.055121 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:16:43 crc kubenswrapper[4632]: I0313 12:16:43.044568 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:16:43 crc kubenswrapper[4632]: E0313 12:16:43.045345 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:16:55 crc kubenswrapper[4632]: I0313 12:16:55.044286 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:16:55 crc kubenswrapper[4632]: E0313 12:16:55.044930 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:17:06 crc kubenswrapper[4632]: I0313 12:17:06.045456 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:17:06 crc kubenswrapper[4632]: E0313 12:17:06.046725 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:17:19 crc kubenswrapper[4632]: I0313 12:17:19.044470 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:17:19 crc kubenswrapper[4632]: E0313 12:17:19.045214 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:17:32 crc kubenswrapper[4632]: I0313 12:17:32.044866 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:17:32 crc kubenswrapper[4632]: E0313 12:17:32.045662 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:17:44 crc kubenswrapper[4632]: I0313 12:17:44.044695 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:17:44 crc kubenswrapper[4632]: E0313 12:17:44.045480 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:17:56 crc kubenswrapper[4632]: I0313 12:17:56.044609 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:17:56 crc kubenswrapper[4632]: E0313 12:17:56.045457 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:18:00 crc kubenswrapper[4632]: I0313 12:18:00.152247 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556738-ktc6z"] Mar 13 12:18:00 crc kubenswrapper[4632]: E0313 12:18:00.153419 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e" containerName="oc" Mar 13 12:18:00 crc kubenswrapper[4632]: I0313 12:18:00.153438 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e" containerName="oc" Mar 13 12:18:00 crc kubenswrapper[4632]: I0313 12:18:00.153663 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e" containerName="oc" Mar 13 12:18:00 crc kubenswrapper[4632]: I0313 12:18:00.154533 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556738-ktc6z" Mar 13 12:18:00 crc kubenswrapper[4632]: I0313 12:18:00.161797 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:18:00 crc kubenswrapper[4632]: I0313 12:18:00.161855 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:18:00 crc kubenswrapper[4632]: I0313 12:18:00.161984 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:18:00 crc kubenswrapper[4632]: I0313 12:18:00.163431 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556738-ktc6z"] Mar 13 12:18:00 crc kubenswrapper[4632]: I0313 12:18:00.275285 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kmzqs\" (UniqueName: \"kubernetes.io/projected/fd947fd4-4e97-4720-98a3-d345ae5dd3fc-kube-api-access-kmzqs\") pod \"auto-csr-approver-29556738-ktc6z\" (UID: \"fd947fd4-4e97-4720-98a3-d345ae5dd3fc\") " pod="openshift-infra/auto-csr-approver-29556738-ktc6z" Mar 13 12:18:00 crc kubenswrapper[4632]: I0313 12:18:00.377472 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kmzqs\" (UniqueName: \"kubernetes.io/projected/fd947fd4-4e97-4720-98a3-d345ae5dd3fc-kube-api-access-kmzqs\") pod \"auto-csr-approver-29556738-ktc6z\" (UID: \"fd947fd4-4e97-4720-98a3-d345ae5dd3fc\") " pod="openshift-infra/auto-csr-approver-29556738-ktc6z" Mar 13 12:18:00 crc kubenswrapper[4632]: I0313 12:18:00.409351 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kmzqs\" (UniqueName: \"kubernetes.io/projected/fd947fd4-4e97-4720-98a3-d345ae5dd3fc-kube-api-access-kmzqs\") pod \"auto-csr-approver-29556738-ktc6z\" (UID: \"fd947fd4-4e97-4720-98a3-d345ae5dd3fc\") " pod="openshift-infra/auto-csr-approver-29556738-ktc6z" Mar 13 12:18:00 crc kubenswrapper[4632]: I0313 12:18:00.489327 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556738-ktc6z" Mar 13 12:18:00 crc kubenswrapper[4632]: I0313 12:18:00.632867 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/swift-proxy-7dbf8b9ddc-6p5vh" podUID="03ca050c-63a7-4b37-91fe-fe5c322cca78" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Mar 13 12:18:00 crc kubenswrapper[4632]: I0313 12:18:00.974908 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556738-ktc6z"] Mar 13 12:18:01 crc kubenswrapper[4632]: I0313 12:18:01.243502 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556738-ktc6z" event={"ID":"fd947fd4-4e97-4720-98a3-d345ae5dd3fc","Type":"ContainerStarted","Data":"9498ecf660d2d20f8dd70c874c835638519a77ddd689a248744483cd6ffdf3e6"} Mar 13 12:18:02 crc kubenswrapper[4632]: I0313 12:18:02.261175 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556738-ktc6z" event={"ID":"fd947fd4-4e97-4720-98a3-d345ae5dd3fc","Type":"ContainerStarted","Data":"fbfc844073b7954c305603f6ba9bca1ebae6e886287d4969b865a335340183e5"} Mar 13 12:18:02 crc kubenswrapper[4632]: I0313 12:18:02.290455 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556738-ktc6z" podStartSLOduration=1.352389181 podStartE2EDuration="2.290421807s" podCreationTimestamp="2026-03-13 12:18:00 +0000 UTC" firstStartedPulling="2026-03-13 12:18:00.98414603 +0000 UTC m=+8055.006676173" lastFinishedPulling="2026-03-13 12:18:01.922178656 +0000 UTC m=+8055.944708799" observedRunningTime="2026-03-13 12:18:02.277545811 +0000 UTC m=+8056.300075944" watchObservedRunningTime="2026-03-13 12:18:02.290421807 +0000 UTC m=+8056.312951950" Mar 13 12:18:03 crc kubenswrapper[4632]: I0313 12:18:03.272798 4632 generic.go:334] "Generic (PLEG): container finished" podID="fd947fd4-4e97-4720-98a3-d345ae5dd3fc" containerID="fbfc844073b7954c305603f6ba9bca1ebae6e886287d4969b865a335340183e5" exitCode=0 Mar 13 12:18:03 crc kubenswrapper[4632]: I0313 12:18:03.272860 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556738-ktc6z" event={"ID":"fd947fd4-4e97-4720-98a3-d345ae5dd3fc","Type":"ContainerDied","Data":"fbfc844073b7954c305603f6ba9bca1ebae6e886287d4969b865a335340183e5"} Mar 13 12:18:04 crc kubenswrapper[4632]: I0313 12:18:04.669628 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556738-ktc6z" Mar 13 12:18:04 crc kubenswrapper[4632]: I0313 12:18:04.763178 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kmzqs\" (UniqueName: \"kubernetes.io/projected/fd947fd4-4e97-4720-98a3-d345ae5dd3fc-kube-api-access-kmzqs\") pod \"fd947fd4-4e97-4720-98a3-d345ae5dd3fc\" (UID: \"fd947fd4-4e97-4720-98a3-d345ae5dd3fc\") " Mar 13 12:18:04 crc kubenswrapper[4632]: I0313 12:18:04.768763 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fd947fd4-4e97-4720-98a3-d345ae5dd3fc-kube-api-access-kmzqs" (OuterVolumeSpecName: "kube-api-access-kmzqs") pod "fd947fd4-4e97-4720-98a3-d345ae5dd3fc" (UID: "fd947fd4-4e97-4720-98a3-d345ae5dd3fc"). InnerVolumeSpecName "kube-api-access-kmzqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:18:04 crc kubenswrapper[4632]: I0313 12:18:04.865474 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kmzqs\" (UniqueName: \"kubernetes.io/projected/fd947fd4-4e97-4720-98a3-d345ae5dd3fc-kube-api-access-kmzqs\") on node \"crc\" DevicePath \"\"" Mar 13 12:18:05 crc kubenswrapper[4632]: I0313 12:18:05.293603 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556738-ktc6z" event={"ID":"fd947fd4-4e97-4720-98a3-d345ae5dd3fc","Type":"ContainerDied","Data":"9498ecf660d2d20f8dd70c874c835638519a77ddd689a248744483cd6ffdf3e6"} Mar 13 12:18:05 crc kubenswrapper[4632]: I0313 12:18:05.293643 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9498ecf660d2d20f8dd70c874c835638519a77ddd689a248744483cd6ffdf3e6" Mar 13 12:18:05 crc kubenswrapper[4632]: I0313 12:18:05.293926 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556738-ktc6z" Mar 13 12:18:05 crc kubenswrapper[4632]: I0313 12:18:05.387162 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556732-zffdj"] Mar 13 12:18:05 crc kubenswrapper[4632]: I0313 12:18:05.396214 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556732-zffdj"] Mar 13 12:18:06 crc kubenswrapper[4632]: I0313 12:18:06.057085 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69547021-fae1-4ad6-8745-c327bb079dce" path="/var/lib/kubelet/pods/69547021-fae1-4ad6-8745-c327bb079dce/volumes" Mar 13 12:18:09 crc kubenswrapper[4632]: I0313 12:18:09.044538 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:18:09 crc kubenswrapper[4632]: E0313 12:18:09.045059 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:18:24 crc kubenswrapper[4632]: I0313 12:18:24.045086 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:18:24 crc kubenswrapper[4632]: E0313 12:18:24.045805 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:18:25 crc kubenswrapper[4632]: I0313 12:18:25.855996 4632 scope.go:117] "RemoveContainer" containerID="e6be43aa992650e79f391597a8fccb4cc829615f369f4902627c5d5e92b6ab1e" Mar 13 12:18:35 crc kubenswrapper[4632]: I0313 12:18:35.044816 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:18:35 crc kubenswrapper[4632]: E0313 12:18:35.045639 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:18:50 crc kubenswrapper[4632]: I0313 12:18:50.044701 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:18:50 crc kubenswrapper[4632]: E0313 12:18:50.045400 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:19:05 crc kubenswrapper[4632]: I0313 12:19:05.044858 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:19:05 crc kubenswrapper[4632]: E0313 12:19:05.046312 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:19:17 crc kubenswrapper[4632]: I0313 12:19:17.044530 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:19:17 crc kubenswrapper[4632]: E0313 12:19:17.045348 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:19:29 crc kubenswrapper[4632]: I0313 12:19:29.045287 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:19:29 crc kubenswrapper[4632]: E0313 12:19:29.046055 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:19:44 crc kubenswrapper[4632]: I0313 12:19:44.044659 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:19:45 crc kubenswrapper[4632]: I0313 12:19:45.308788 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"d2f7d92ea8336c364393ccfd7369387047df3a4555b1b7f7be871c5ae3268440"} Mar 13 12:20:00 crc kubenswrapper[4632]: I0313 12:20:00.223233 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556740-fmh5b"] Mar 13 12:20:00 crc kubenswrapper[4632]: E0313 12:20:00.233008 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fd947fd4-4e97-4720-98a3-d345ae5dd3fc" containerName="oc" Mar 13 12:20:00 crc kubenswrapper[4632]: I0313 12:20:00.233058 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="fd947fd4-4e97-4720-98a3-d345ae5dd3fc" containerName="oc" Mar 13 12:20:00 crc kubenswrapper[4632]: I0313 12:20:00.235001 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="fd947fd4-4e97-4720-98a3-d345ae5dd3fc" containerName="oc" Mar 13 12:20:00 crc kubenswrapper[4632]: I0313 12:20:00.244521 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556740-fmh5b" Mar 13 12:20:00 crc kubenswrapper[4632]: I0313 12:20:00.256111 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:20:00 crc kubenswrapper[4632]: I0313 12:20:00.256114 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:20:00 crc kubenswrapper[4632]: I0313 12:20:00.256124 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:20:00 crc kubenswrapper[4632]: I0313 12:20:00.265595 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh5z2\" (UniqueName: \"kubernetes.io/projected/0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c-kube-api-access-sh5z2\") pod \"auto-csr-approver-29556740-fmh5b\" (UID: \"0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c\") " pod="openshift-infra/auto-csr-approver-29556740-fmh5b" Mar 13 12:20:00 crc kubenswrapper[4632]: I0313 12:20:00.267890 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556740-fmh5b"] Mar 13 12:20:00 crc kubenswrapper[4632]: I0313 12:20:00.367615 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh5z2\" (UniqueName: \"kubernetes.io/projected/0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c-kube-api-access-sh5z2\") pod \"auto-csr-approver-29556740-fmh5b\" (UID: \"0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c\") " pod="openshift-infra/auto-csr-approver-29556740-fmh5b" Mar 13 12:20:00 crc kubenswrapper[4632]: I0313 12:20:00.397061 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh5z2\" (UniqueName: \"kubernetes.io/projected/0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c-kube-api-access-sh5z2\") pod \"auto-csr-approver-29556740-fmh5b\" (UID: \"0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c\") " pod="openshift-infra/auto-csr-approver-29556740-fmh5b" Mar 13 12:20:00 crc kubenswrapper[4632]: I0313 12:20:00.573466 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556740-fmh5b" Mar 13 12:20:01 crc kubenswrapper[4632]: I0313 12:20:01.684793 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556740-fmh5b"] Mar 13 12:20:01 crc kubenswrapper[4632]: W0313 12:20:01.716753 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f86aa5e_9cfc_458f_ae11_71e5e4dcfe9c.slice/crio-59c3307f4abaa31382716abc4886c1fdbb076e0a57ed93e0d5dc37cfa2b758c4 WatchSource:0}: Error finding container 59c3307f4abaa31382716abc4886c1fdbb076e0a57ed93e0d5dc37cfa2b758c4: Status 404 returned error can't find the container with id 59c3307f4abaa31382716abc4886c1fdbb076e0a57ed93e0d5dc37cfa2b758c4 Mar 13 12:20:01 crc kubenswrapper[4632]: I0313 12:20:01.727128 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 12:20:02 crc kubenswrapper[4632]: I0313 12:20:02.498967 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556740-fmh5b" event={"ID":"0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c","Type":"ContainerStarted","Data":"59c3307f4abaa31382716abc4886c1fdbb076e0a57ed93e0d5dc37cfa2b758c4"} Mar 13 12:20:05 crc kubenswrapper[4632]: I0313 12:20:05.539986 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556740-fmh5b" event={"ID":"0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c","Type":"ContainerStarted","Data":"e0e701c935a2c4084fd4e093f0c21450f3afd1589228584f67fcd3cbe4d41395"} Mar 13 12:20:05 crc kubenswrapper[4632]: I0313 12:20:05.574368 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556740-fmh5b" podStartSLOduration=3.447666901 podStartE2EDuration="5.573240782s" podCreationTimestamp="2026-03-13 12:20:00 +0000 UTC" firstStartedPulling="2026-03-13 12:20:01.723118144 +0000 UTC m=+8175.745648277" lastFinishedPulling="2026-03-13 12:20:03.848692025 +0000 UTC m=+8177.871222158" observedRunningTime="2026-03-13 12:20:05.557351271 +0000 UTC m=+8179.579881404" watchObservedRunningTime="2026-03-13 12:20:05.573240782 +0000 UTC m=+8179.595770925" Mar 13 12:20:06 crc kubenswrapper[4632]: I0313 12:20:06.552303 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556740-fmh5b" event={"ID":"0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c","Type":"ContainerDied","Data":"e0e701c935a2c4084fd4e093f0c21450f3afd1589228584f67fcd3cbe4d41395"} Mar 13 12:20:06 crc kubenswrapper[4632]: I0313 12:20:06.554383 4632 generic.go:334] "Generic (PLEG): container finished" podID="0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c" containerID="e0e701c935a2c4084fd4e093f0c21450f3afd1589228584f67fcd3cbe4d41395" exitCode=0 Mar 13 12:20:08 crc kubenswrapper[4632]: I0313 12:20:08.038508 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556740-fmh5b" Mar 13 12:20:08 crc kubenswrapper[4632]: I0313 12:20:08.227075 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh5z2\" (UniqueName: \"kubernetes.io/projected/0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c-kube-api-access-sh5z2\") pod \"0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c\" (UID: \"0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c\") " Mar 13 12:20:08 crc kubenswrapper[4632]: I0313 12:20:08.235698 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c-kube-api-access-sh5z2" (OuterVolumeSpecName: "kube-api-access-sh5z2") pod "0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c" (UID: "0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c"). InnerVolumeSpecName "kube-api-access-sh5z2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:20:08 crc kubenswrapper[4632]: I0313 12:20:08.329609 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh5z2\" (UniqueName: \"kubernetes.io/projected/0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c-kube-api-access-sh5z2\") on node \"crc\" DevicePath \"\"" Mar 13 12:20:08 crc kubenswrapper[4632]: I0313 12:20:08.580299 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556740-fmh5b" event={"ID":"0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c","Type":"ContainerDied","Data":"59c3307f4abaa31382716abc4886c1fdbb076e0a57ed93e0d5dc37cfa2b758c4"} Mar 13 12:20:08 crc kubenswrapper[4632]: I0313 12:20:08.580375 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556740-fmh5b" Mar 13 12:20:08 crc kubenswrapper[4632]: I0313 12:20:08.584290 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59c3307f4abaa31382716abc4886c1fdbb076e0a57ed93e0d5dc37cfa2b758c4" Mar 13 12:20:08 crc kubenswrapper[4632]: I0313 12:20:08.787317 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556734-bwv75"] Mar 13 12:20:08 crc kubenswrapper[4632]: I0313 12:20:08.796745 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556734-bwv75"] Mar 13 12:20:10 crc kubenswrapper[4632]: I0313 12:20:10.056459 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d0a5571-d345-44cf-ba1a-46b3ef68b1ae" path="/var/lib/kubelet/pods/5d0a5571-d345-44cf-ba1a-46b3ef68b1ae/volumes" Mar 13 12:20:25 crc kubenswrapper[4632]: I0313 12:20:25.976970 4632 scope.go:117] "RemoveContainer" containerID="6faf20a0273a34b58a9864d2f460cdaebfb4cad46108ca19c20cf494270e4fe6" Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.553535 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-n44pj"] Mar 13 12:21:08 crc kubenswrapper[4632]: E0313 12:21:08.554467 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c" containerName="oc" Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.554482 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c" containerName="oc" Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.554696 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c" containerName="oc" Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.558088 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.593718 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n44pj"] Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.636218 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7n2rg\" (UniqueName: \"kubernetes.io/projected/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-kube-api-access-7n2rg\") pod \"community-operators-n44pj\" (UID: \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\") " pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.636318 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-catalog-content\") pod \"community-operators-n44pj\" (UID: \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\") " pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.636514 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-utilities\") pod \"community-operators-n44pj\" (UID: \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\") " pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.739066 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7n2rg\" (UniqueName: \"kubernetes.io/projected/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-kube-api-access-7n2rg\") pod \"community-operators-n44pj\" (UID: \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\") " pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.739136 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-catalog-content\") pod \"community-operators-n44pj\" (UID: \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\") " pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.739184 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-utilities\") pod \"community-operators-n44pj\" (UID: \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\") " pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.740195 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-utilities\") pod \"community-operators-n44pj\" (UID: \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\") " pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.741068 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-catalog-content\") pod \"community-operators-n44pj\" (UID: \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\") " pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.770875 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7n2rg\" (UniqueName: \"kubernetes.io/projected/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-kube-api-access-7n2rg\") pod \"community-operators-n44pj\" (UID: \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\") " pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:08 crc kubenswrapper[4632]: I0313 12:21:08.881461 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:09 crc kubenswrapper[4632]: I0313 12:21:09.752992 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-n44pj"] Mar 13 12:21:09 crc kubenswrapper[4632]: W0313 12:21:09.768794 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddaf8e5c5_e40a_40d4_b4ea_90347a55cf3c.slice/crio-3c2fbf96cda473030c977a4f1c925da83dc9d613ad78a8d681a07cc04ca99524 WatchSource:0}: Error finding container 3c2fbf96cda473030c977a4f1c925da83dc9d613ad78a8d681a07cc04ca99524: Status 404 returned error can't find the container with id 3c2fbf96cda473030c977a4f1c925da83dc9d613ad78a8d681a07cc04ca99524 Mar 13 12:21:10 crc kubenswrapper[4632]: I0313 12:21:10.257024 4632 generic.go:334] "Generic (PLEG): container finished" podID="daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" containerID="de03b8a6c2df4b8371e57e5a125b76cbf3601ae9167dab0f2c3ad95c58607eae" exitCode=0 Mar 13 12:21:10 crc kubenswrapper[4632]: I0313 12:21:10.257076 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n44pj" event={"ID":"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c","Type":"ContainerDied","Data":"de03b8a6c2df4b8371e57e5a125b76cbf3601ae9167dab0f2c3ad95c58607eae"} Mar 13 12:21:10 crc kubenswrapper[4632]: I0313 12:21:10.257111 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n44pj" event={"ID":"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c","Type":"ContainerStarted","Data":"3c2fbf96cda473030c977a4f1c925da83dc9d613ad78a8d681a07cc04ca99524"} Mar 13 12:21:11 crc kubenswrapper[4632]: I0313 12:21:11.269351 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n44pj" event={"ID":"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c","Type":"ContainerStarted","Data":"81e63ca60a0e0dd00edeac06af45faf13f8dc28e5ffab65ff127a569ef805ba8"} Mar 13 12:21:14 crc kubenswrapper[4632]: I0313 12:21:14.297712 4632 generic.go:334] "Generic (PLEG): container finished" podID="daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" containerID="81e63ca60a0e0dd00edeac06af45faf13f8dc28e5ffab65ff127a569ef805ba8" exitCode=0 Mar 13 12:21:14 crc kubenswrapper[4632]: I0313 12:21:14.297785 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n44pj" event={"ID":"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c","Type":"ContainerDied","Data":"81e63ca60a0e0dd00edeac06af45faf13f8dc28e5ffab65ff127a569ef805ba8"} Mar 13 12:21:15 crc kubenswrapper[4632]: I0313 12:21:15.310919 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n44pj" event={"ID":"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c","Type":"ContainerStarted","Data":"0d61eedbe365424460c70706e237ac566dd5083940c2f53811f5b3ff865d54b0"} Mar 13 12:21:15 crc kubenswrapper[4632]: I0313 12:21:15.348637 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-n44pj" podStartSLOduration=2.862916245 podStartE2EDuration="7.34861085s" podCreationTimestamp="2026-03-13 12:21:08 +0000 UTC" firstStartedPulling="2026-03-13 12:21:10.259348322 +0000 UTC m=+8244.281878455" lastFinishedPulling="2026-03-13 12:21:14.745042927 +0000 UTC m=+8248.767573060" observedRunningTime="2026-03-13 12:21:15.338923122 +0000 UTC m=+8249.361453285" watchObservedRunningTime="2026-03-13 12:21:15.34861085 +0000 UTC m=+8249.371141013" Mar 13 12:21:18 crc kubenswrapper[4632]: I0313 12:21:18.881730 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:18 crc kubenswrapper[4632]: I0313 12:21:18.882564 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:19 crc kubenswrapper[4632]: I0313 12:21:19.932200 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-n44pj" podUID="daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" containerName="registry-server" probeResult="failure" output=< Mar 13 12:21:19 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:21:19 crc kubenswrapper[4632]: > Mar 13 12:21:24 crc kubenswrapper[4632]: I0313 12:21:24.661222 4632 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.013797589s: [/var/lib/containers/storage/overlay/529e5121e17145dccbfe8df9c66facc7a0f7ba465e25d548cd766cbc085df8a5/diff /var/log/pods/openstack_cinder-scheduler-0_d2c1c19b-95a5-4db1-8e54-36fe83704b25/cinder-scheduler/1.log]; will not log again for this container unless duration exceeds 2s Mar 13 12:21:24 crc kubenswrapper[4632]: I0313 12:21:24.678684 4632 trace.go:236] Trace[240102934]: "Calculate volume metrics of run-httpd for pod openstack/ceilometer-0" (13-Mar-2026 12:21:22.650) (total time: 2006ms): Mar 13 12:21:24 crc kubenswrapper[4632]: Trace[240102934]: [2.006686126s] [2.006686126s] END Mar 13 12:21:24 crc kubenswrapper[4632]: I0313 12:21:24.685187 4632 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" probeResult="failure" output="command timed out" Mar 13 12:21:24 crc kubenswrapper[4632]: I0313 12:21:24.718745 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="2cb2f546-c8c5-4ec9-aba8-d3782431de10" containerName="galera" probeResult="failure" output="command timed out" Mar 13 12:21:29 crc kubenswrapper[4632]: I0313 12:21:29.941333 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-n44pj" podUID="daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" containerName="registry-server" probeResult="failure" output=< Mar 13 12:21:29 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:21:29 crc kubenswrapper[4632]: > Mar 13 12:21:38 crc kubenswrapper[4632]: I0313 12:21:38.952099 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:39 crc kubenswrapper[4632]: I0313 12:21:39.013268 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:39 crc kubenswrapper[4632]: I0313 12:21:39.778254 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n44pj"] Mar 13 12:21:40 crc kubenswrapper[4632]: I0313 12:21:40.549473 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-n44pj" podUID="daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" containerName="registry-server" containerID="cri-o://0d61eedbe365424460c70706e237ac566dd5083940c2f53811f5b3ff865d54b0" gracePeriod=2 Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.319837 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.437138 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-utilities\") pod \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\" (UID: \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\") " Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.437220 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7n2rg\" (UniqueName: \"kubernetes.io/projected/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-kube-api-access-7n2rg\") pod \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\" (UID: \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\") " Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.437590 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-catalog-content\") pod \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\" (UID: \"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c\") " Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.437871 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-utilities" (OuterVolumeSpecName: "utilities") pod "daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" (UID: "daf8e5c5-e40a-40d4-b4ea-90347a55cf3c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.438436 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.457236 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-kube-api-access-7n2rg" (OuterVolumeSpecName: "kube-api-access-7n2rg") pod "daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" (UID: "daf8e5c5-e40a-40d4-b4ea-90347a55cf3c"). InnerVolumeSpecName "kube-api-access-7n2rg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.499985 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" (UID: "daf8e5c5-e40a-40d4-b4ea-90347a55cf3c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.540562 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7n2rg\" (UniqueName: \"kubernetes.io/projected/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-kube-api-access-7n2rg\") on node \"crc\" DevicePath \"\"" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.540622 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.562993 4632 generic.go:334] "Generic (PLEG): container finished" podID="daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" containerID="0d61eedbe365424460c70706e237ac566dd5083940c2f53811f5b3ff865d54b0" exitCode=0 Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.563056 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-n44pj" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.563065 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n44pj" event={"ID":"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c","Type":"ContainerDied","Data":"0d61eedbe365424460c70706e237ac566dd5083940c2f53811f5b3ff865d54b0"} Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.563107 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-n44pj" event={"ID":"daf8e5c5-e40a-40d4-b4ea-90347a55cf3c","Type":"ContainerDied","Data":"3c2fbf96cda473030c977a4f1c925da83dc9d613ad78a8d681a07cc04ca99524"} Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.563162 4632 scope.go:117] "RemoveContainer" containerID="0d61eedbe365424460c70706e237ac566dd5083940c2f53811f5b3ff865d54b0" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.609618 4632 scope.go:117] "RemoveContainer" containerID="81e63ca60a0e0dd00edeac06af45faf13f8dc28e5ffab65ff127a569ef805ba8" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.614788 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-n44pj"] Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.625391 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-n44pj"] Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.640480 4632 scope.go:117] "RemoveContainer" containerID="de03b8a6c2df4b8371e57e5a125b76cbf3601ae9167dab0f2c3ad95c58607eae" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.677718 4632 scope.go:117] "RemoveContainer" containerID="0d61eedbe365424460c70706e237ac566dd5083940c2f53811f5b3ff865d54b0" Mar 13 12:21:41 crc kubenswrapper[4632]: E0313 12:21:41.684024 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d61eedbe365424460c70706e237ac566dd5083940c2f53811f5b3ff865d54b0\": container with ID starting with 0d61eedbe365424460c70706e237ac566dd5083940c2f53811f5b3ff865d54b0 not found: ID does not exist" containerID="0d61eedbe365424460c70706e237ac566dd5083940c2f53811f5b3ff865d54b0" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.684078 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d61eedbe365424460c70706e237ac566dd5083940c2f53811f5b3ff865d54b0"} err="failed to get container status \"0d61eedbe365424460c70706e237ac566dd5083940c2f53811f5b3ff865d54b0\": rpc error: code = NotFound desc = could not find container \"0d61eedbe365424460c70706e237ac566dd5083940c2f53811f5b3ff865d54b0\": container with ID starting with 0d61eedbe365424460c70706e237ac566dd5083940c2f53811f5b3ff865d54b0 not found: ID does not exist" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.684111 4632 scope.go:117] "RemoveContainer" containerID="81e63ca60a0e0dd00edeac06af45faf13f8dc28e5ffab65ff127a569ef805ba8" Mar 13 12:21:41 crc kubenswrapper[4632]: E0313 12:21:41.684910 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81e63ca60a0e0dd00edeac06af45faf13f8dc28e5ffab65ff127a569ef805ba8\": container with ID starting with 81e63ca60a0e0dd00edeac06af45faf13f8dc28e5ffab65ff127a569ef805ba8 not found: ID does not exist" containerID="81e63ca60a0e0dd00edeac06af45faf13f8dc28e5ffab65ff127a569ef805ba8" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.684978 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81e63ca60a0e0dd00edeac06af45faf13f8dc28e5ffab65ff127a569ef805ba8"} err="failed to get container status \"81e63ca60a0e0dd00edeac06af45faf13f8dc28e5ffab65ff127a569ef805ba8\": rpc error: code = NotFound desc = could not find container \"81e63ca60a0e0dd00edeac06af45faf13f8dc28e5ffab65ff127a569ef805ba8\": container with ID starting with 81e63ca60a0e0dd00edeac06af45faf13f8dc28e5ffab65ff127a569ef805ba8 not found: ID does not exist" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.685008 4632 scope.go:117] "RemoveContainer" containerID="de03b8a6c2df4b8371e57e5a125b76cbf3601ae9167dab0f2c3ad95c58607eae" Mar 13 12:21:41 crc kubenswrapper[4632]: E0313 12:21:41.685436 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"de03b8a6c2df4b8371e57e5a125b76cbf3601ae9167dab0f2c3ad95c58607eae\": container with ID starting with de03b8a6c2df4b8371e57e5a125b76cbf3601ae9167dab0f2c3ad95c58607eae not found: ID does not exist" containerID="de03b8a6c2df4b8371e57e5a125b76cbf3601ae9167dab0f2c3ad95c58607eae" Mar 13 12:21:41 crc kubenswrapper[4632]: I0313 12:21:41.685457 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"de03b8a6c2df4b8371e57e5a125b76cbf3601ae9167dab0f2c3ad95c58607eae"} err="failed to get container status \"de03b8a6c2df4b8371e57e5a125b76cbf3601ae9167dab0f2c3ad95c58607eae\": rpc error: code = NotFound desc = could not find container \"de03b8a6c2df4b8371e57e5a125b76cbf3601ae9167dab0f2c3ad95c58607eae\": container with ID starting with de03b8a6c2df4b8371e57e5a125b76cbf3601ae9167dab0f2c3ad95c58607eae not found: ID does not exist" Mar 13 12:21:42 crc kubenswrapper[4632]: I0313 12:21:42.069005 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" path="/var/lib/kubelet/pods/daf8e5c5-e40a-40d4-b4ea-90347a55cf3c/volumes" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.584214 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-hnpsr"] Mar 13 12:21:47 crc kubenswrapper[4632]: E0313 12:21:47.585634 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" containerName="extract-utilities" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.585650 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" containerName="extract-utilities" Mar 13 12:21:47 crc kubenswrapper[4632]: E0313 12:21:47.585665 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" containerName="extract-content" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.585672 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" containerName="extract-content" Mar 13 12:21:47 crc kubenswrapper[4632]: E0313 12:21:47.585684 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" containerName="registry-server" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.585691 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" containerName="registry-server" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.585920 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="daf8e5c5-e40a-40d4-b4ea-90347a55cf3c" containerName="registry-server" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.587270 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.591976 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xdn5\" (UniqueName: \"kubernetes.io/projected/e37141ae-e543-40fa-876c-e7b7d9e1598f-kube-api-access-6xdn5\") pod \"redhat-marketplace-hnpsr\" (UID: \"e37141ae-e543-40fa-876c-e7b7d9e1598f\") " pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.592035 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37141ae-e543-40fa-876c-e7b7d9e1598f-utilities\") pod \"redhat-marketplace-hnpsr\" (UID: \"e37141ae-e543-40fa-876c-e7b7d9e1598f\") " pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.592537 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37141ae-e543-40fa-876c-e7b7d9e1598f-catalog-content\") pod \"redhat-marketplace-hnpsr\" (UID: \"e37141ae-e543-40fa-876c-e7b7d9e1598f\") " pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.601624 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnpsr"] Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.704838 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xdn5\" (UniqueName: \"kubernetes.io/projected/e37141ae-e543-40fa-876c-e7b7d9e1598f-kube-api-access-6xdn5\") pod \"redhat-marketplace-hnpsr\" (UID: \"e37141ae-e543-40fa-876c-e7b7d9e1598f\") " pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.704930 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37141ae-e543-40fa-876c-e7b7d9e1598f-utilities\") pod \"redhat-marketplace-hnpsr\" (UID: \"e37141ae-e543-40fa-876c-e7b7d9e1598f\") " pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.705117 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37141ae-e543-40fa-876c-e7b7d9e1598f-catalog-content\") pod \"redhat-marketplace-hnpsr\" (UID: \"e37141ae-e543-40fa-876c-e7b7d9e1598f\") " pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.706010 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37141ae-e543-40fa-876c-e7b7d9e1598f-utilities\") pod \"redhat-marketplace-hnpsr\" (UID: \"e37141ae-e543-40fa-876c-e7b7d9e1598f\") " pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.706093 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37141ae-e543-40fa-876c-e7b7d9e1598f-catalog-content\") pod \"redhat-marketplace-hnpsr\" (UID: \"e37141ae-e543-40fa-876c-e7b7d9e1598f\") " pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.726121 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xdn5\" (UniqueName: \"kubernetes.io/projected/e37141ae-e543-40fa-876c-e7b7d9e1598f-kube-api-access-6xdn5\") pod \"redhat-marketplace-hnpsr\" (UID: \"e37141ae-e543-40fa-876c-e7b7d9e1598f\") " pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:21:47 crc kubenswrapper[4632]: I0313 12:21:47.946798 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:21:48 crc kubenswrapper[4632]: I0313 12:21:48.656785 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnpsr"] Mar 13 12:21:49 crc kubenswrapper[4632]: I0313 12:21:49.663768 4632 generic.go:334] "Generic (PLEG): container finished" podID="e37141ae-e543-40fa-876c-e7b7d9e1598f" containerID="030c12dcf041fb69ab0b030b3ab62bdd3a22b3a40f28da17a2039141caaeedee" exitCode=0 Mar 13 12:21:49 crc kubenswrapper[4632]: I0313 12:21:49.664004 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnpsr" event={"ID":"e37141ae-e543-40fa-876c-e7b7d9e1598f","Type":"ContainerDied","Data":"030c12dcf041fb69ab0b030b3ab62bdd3a22b3a40f28da17a2039141caaeedee"} Mar 13 12:21:49 crc kubenswrapper[4632]: I0313 12:21:49.665180 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnpsr" event={"ID":"e37141ae-e543-40fa-876c-e7b7d9e1598f","Type":"ContainerStarted","Data":"9caf9004314593e78add6272bfed81d0b4f7fa3929f6184e4cb0ffa2bbc615fc"} Mar 13 12:21:50 crc kubenswrapper[4632]: I0313 12:21:50.675348 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnpsr" event={"ID":"e37141ae-e543-40fa-876c-e7b7d9e1598f","Type":"ContainerStarted","Data":"ead807d406032ccc6de750261a0d5e2b71a645cfaa7f853ed7fc97cd308b03bb"} Mar 13 12:21:52 crc kubenswrapper[4632]: I0313 12:21:52.695849 4632 generic.go:334] "Generic (PLEG): container finished" podID="e37141ae-e543-40fa-876c-e7b7d9e1598f" containerID="ead807d406032ccc6de750261a0d5e2b71a645cfaa7f853ed7fc97cd308b03bb" exitCode=0 Mar 13 12:21:52 crc kubenswrapper[4632]: I0313 12:21:52.696205 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnpsr" event={"ID":"e37141ae-e543-40fa-876c-e7b7d9e1598f","Type":"ContainerDied","Data":"ead807d406032ccc6de750261a0d5e2b71a645cfaa7f853ed7fc97cd308b03bb"} Mar 13 12:21:54 crc kubenswrapper[4632]: I0313 12:21:54.714417 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnpsr" event={"ID":"e37141ae-e543-40fa-876c-e7b7d9e1598f","Type":"ContainerStarted","Data":"59dd5b6704b2e6f99f2da9cb87a940f54fb4b4c36738389e043006b38d198693"} Mar 13 12:21:54 crc kubenswrapper[4632]: I0313 12:21:54.736328 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-hnpsr" podStartSLOduration=3.85097033 podStartE2EDuration="7.73630813s" podCreationTimestamp="2026-03-13 12:21:47 +0000 UTC" firstStartedPulling="2026-03-13 12:21:49.666367416 +0000 UTC m=+8283.688897549" lastFinishedPulling="2026-03-13 12:21:53.551705216 +0000 UTC m=+8287.574235349" observedRunningTime="2026-03-13 12:21:54.733597304 +0000 UTC m=+8288.756127457" watchObservedRunningTime="2026-03-13 12:21:54.73630813 +0000 UTC m=+8288.758838263" Mar 13 12:21:57 crc kubenswrapper[4632]: I0313 12:21:57.947728 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:21:57 crc kubenswrapper[4632]: I0313 12:21:57.948117 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:21:58 crc kubenswrapper[4632]: I0313 12:21:58.995773 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-hnpsr" podUID="e37141ae-e543-40fa-876c-e7b7d9e1598f" containerName="registry-server" probeResult="failure" output=< Mar 13 12:21:58 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:21:58 crc kubenswrapper[4632]: > Mar 13 12:22:00 crc kubenswrapper[4632]: I0313 12:22:00.243355 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556742-2lkrj"] Mar 13 12:22:00 crc kubenswrapper[4632]: I0313 12:22:00.245010 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556742-2lkrj" Mar 13 12:22:00 crc kubenswrapper[4632]: I0313 12:22:00.253392 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfptl\" (UniqueName: \"kubernetes.io/projected/a5121453-468a-432e-b110-fd0cd60ed92b-kube-api-access-vfptl\") pod \"auto-csr-approver-29556742-2lkrj\" (UID: \"a5121453-468a-432e-b110-fd0cd60ed92b\") " pod="openshift-infra/auto-csr-approver-29556742-2lkrj" Mar 13 12:22:00 crc kubenswrapper[4632]: I0313 12:22:00.256051 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556742-2lkrj"] Mar 13 12:22:00 crc kubenswrapper[4632]: I0313 12:22:00.265190 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:22:00 crc kubenswrapper[4632]: I0313 12:22:00.265288 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:22:00 crc kubenswrapper[4632]: I0313 12:22:00.267128 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:22:00 crc kubenswrapper[4632]: I0313 12:22:00.355246 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfptl\" (UniqueName: \"kubernetes.io/projected/a5121453-468a-432e-b110-fd0cd60ed92b-kube-api-access-vfptl\") pod \"auto-csr-approver-29556742-2lkrj\" (UID: \"a5121453-468a-432e-b110-fd0cd60ed92b\") " pod="openshift-infra/auto-csr-approver-29556742-2lkrj" Mar 13 12:22:00 crc kubenswrapper[4632]: I0313 12:22:00.386729 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfptl\" (UniqueName: \"kubernetes.io/projected/a5121453-468a-432e-b110-fd0cd60ed92b-kube-api-access-vfptl\") pod \"auto-csr-approver-29556742-2lkrj\" (UID: \"a5121453-468a-432e-b110-fd0cd60ed92b\") " pod="openshift-infra/auto-csr-approver-29556742-2lkrj" Mar 13 12:22:00 crc kubenswrapper[4632]: I0313 12:22:00.583433 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556742-2lkrj" Mar 13 12:22:01 crc kubenswrapper[4632]: I0313 12:22:01.175093 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556742-2lkrj"] Mar 13 12:22:01 crc kubenswrapper[4632]: I0313 12:22:01.774263 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556742-2lkrj" event={"ID":"a5121453-468a-432e-b110-fd0cd60ed92b","Type":"ContainerStarted","Data":"041768bd180226590294d48b8e490a5a4f923f76232d5da42f6d9ec2ac219f63"} Mar 13 12:22:03 crc kubenswrapper[4632]: I0313 12:22:03.798411 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556742-2lkrj" event={"ID":"a5121453-468a-432e-b110-fd0cd60ed92b","Type":"ContainerStarted","Data":"77b11f376c487493e748aed75424d32e4d98e9395efe94071abc3a7b13ebc06d"} Mar 13 12:22:03 crc kubenswrapper[4632]: I0313 12:22:03.824553 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556742-2lkrj" podStartSLOduration=2.764748743 podStartE2EDuration="3.824530915s" podCreationTimestamp="2026-03-13 12:22:00 +0000 UTC" firstStartedPulling="2026-03-13 12:22:01.182132759 +0000 UTC m=+8295.204662892" lastFinishedPulling="2026-03-13 12:22:02.241914931 +0000 UTC m=+8296.264445064" observedRunningTime="2026-03-13 12:22:03.813909195 +0000 UTC m=+8297.836439328" watchObservedRunningTime="2026-03-13 12:22:03.824530915 +0000 UTC m=+8297.847061058" Mar 13 12:22:04 crc kubenswrapper[4632]: I0313 12:22:04.808922 4632 generic.go:334] "Generic (PLEG): container finished" podID="a5121453-468a-432e-b110-fd0cd60ed92b" containerID="77b11f376c487493e748aed75424d32e4d98e9395efe94071abc3a7b13ebc06d" exitCode=0 Mar 13 12:22:04 crc kubenswrapper[4632]: I0313 12:22:04.808990 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556742-2lkrj" event={"ID":"a5121453-468a-432e-b110-fd0cd60ed92b","Type":"ContainerDied","Data":"77b11f376c487493e748aed75424d32e4d98e9395efe94071abc3a7b13ebc06d"} Mar 13 12:22:06 crc kubenswrapper[4632]: I0313 12:22:06.407921 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556742-2lkrj" Mar 13 12:22:06 crc kubenswrapper[4632]: I0313 12:22:06.576698 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfptl\" (UniqueName: \"kubernetes.io/projected/a5121453-468a-432e-b110-fd0cd60ed92b-kube-api-access-vfptl\") pod \"a5121453-468a-432e-b110-fd0cd60ed92b\" (UID: \"a5121453-468a-432e-b110-fd0cd60ed92b\") " Mar 13 12:22:06 crc kubenswrapper[4632]: I0313 12:22:06.588253 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5121453-468a-432e-b110-fd0cd60ed92b-kube-api-access-vfptl" (OuterVolumeSpecName: "kube-api-access-vfptl") pod "a5121453-468a-432e-b110-fd0cd60ed92b" (UID: "a5121453-468a-432e-b110-fd0cd60ed92b"). InnerVolumeSpecName "kube-api-access-vfptl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:22:06 crc kubenswrapper[4632]: I0313 12:22:06.679659 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfptl\" (UniqueName: \"kubernetes.io/projected/a5121453-468a-432e-b110-fd0cd60ed92b-kube-api-access-vfptl\") on node \"crc\" DevicePath \"\"" Mar 13 12:22:06 crc kubenswrapper[4632]: I0313 12:22:06.828046 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556742-2lkrj" event={"ID":"a5121453-468a-432e-b110-fd0cd60ed92b","Type":"ContainerDied","Data":"041768bd180226590294d48b8e490a5a4f923f76232d5da42f6d9ec2ac219f63"} Mar 13 12:22:06 crc kubenswrapper[4632]: I0313 12:22:06.828425 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556742-2lkrj" Mar 13 12:22:06 crc kubenswrapper[4632]: I0313 12:22:06.828095 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="041768bd180226590294d48b8e490a5a4f923f76232d5da42f6d9ec2ac219f63" Mar 13 12:22:06 crc kubenswrapper[4632]: I0313 12:22:06.946382 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556736-8cfm9"] Mar 13 12:22:06 crc kubenswrapper[4632]: I0313 12:22:06.955599 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556736-8cfm9"] Mar 13 12:22:08 crc kubenswrapper[4632]: I0313 12:22:08.015283 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:22:08 crc kubenswrapper[4632]: I0313 12:22:08.091658 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e" path="/var/lib/kubelet/pods/d3f5dea9-fbf6-48ea-89ef-cf2c15d1689e/volumes" Mar 13 12:22:08 crc kubenswrapper[4632]: I0313 12:22:08.094312 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:22:08 crc kubenswrapper[4632]: I0313 12:22:08.257017 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnpsr"] Mar 13 12:22:09 crc kubenswrapper[4632]: I0313 12:22:09.862373 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-hnpsr" podUID="e37141ae-e543-40fa-876c-e7b7d9e1598f" containerName="registry-server" containerID="cri-o://59dd5b6704b2e6f99f2da9cb87a940f54fb4b4c36738389e043006b38d198693" gracePeriod=2 Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.423465 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.458982 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xdn5\" (UniqueName: \"kubernetes.io/projected/e37141ae-e543-40fa-876c-e7b7d9e1598f-kube-api-access-6xdn5\") pod \"e37141ae-e543-40fa-876c-e7b7d9e1598f\" (UID: \"e37141ae-e543-40fa-876c-e7b7d9e1598f\") " Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.459303 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37141ae-e543-40fa-876c-e7b7d9e1598f-utilities\") pod \"e37141ae-e543-40fa-876c-e7b7d9e1598f\" (UID: \"e37141ae-e543-40fa-876c-e7b7d9e1598f\") " Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.459337 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37141ae-e543-40fa-876c-e7b7d9e1598f-catalog-content\") pod \"e37141ae-e543-40fa-876c-e7b7d9e1598f\" (UID: \"e37141ae-e543-40fa-876c-e7b7d9e1598f\") " Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.460106 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e37141ae-e543-40fa-876c-e7b7d9e1598f-utilities" (OuterVolumeSpecName: "utilities") pod "e37141ae-e543-40fa-876c-e7b7d9e1598f" (UID: "e37141ae-e543-40fa-876c-e7b7d9e1598f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.460545 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.464152 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.467468 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e37141ae-e543-40fa-876c-e7b7d9e1598f-kube-api-access-6xdn5" (OuterVolumeSpecName: "kube-api-access-6xdn5") pod "e37141ae-e543-40fa-876c-e7b7d9e1598f" (UID: "e37141ae-e543-40fa-876c-e7b7d9e1598f"). InnerVolumeSpecName "kube-api-access-6xdn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.502508 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e37141ae-e543-40fa-876c-e7b7d9e1598f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e37141ae-e543-40fa-876c-e7b7d9e1598f" (UID: "e37141ae-e543-40fa-876c-e7b7d9e1598f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.562329 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xdn5\" (UniqueName: \"kubernetes.io/projected/e37141ae-e543-40fa-876c-e7b7d9e1598f-kube-api-access-6xdn5\") on node \"crc\" DevicePath \"\"" Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.562384 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e37141ae-e543-40fa-876c-e7b7d9e1598f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.562394 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e37141ae-e543-40fa-876c-e7b7d9e1598f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.875474 4632 generic.go:334] "Generic (PLEG): container finished" podID="e37141ae-e543-40fa-876c-e7b7d9e1598f" containerID="59dd5b6704b2e6f99f2da9cb87a940f54fb4b4c36738389e043006b38d198693" exitCode=0 Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.875572 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-hnpsr" Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.875561 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnpsr" event={"ID":"e37141ae-e543-40fa-876c-e7b7d9e1598f","Type":"ContainerDied","Data":"59dd5b6704b2e6f99f2da9cb87a940f54fb4b4c36738389e043006b38d198693"} Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.875786 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-hnpsr" event={"ID":"e37141ae-e543-40fa-876c-e7b7d9e1598f","Type":"ContainerDied","Data":"9caf9004314593e78add6272bfed81d0b4f7fa3929f6184e4cb0ffa2bbc615fc"} Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.875819 4632 scope.go:117] "RemoveContainer" containerID="59dd5b6704b2e6f99f2da9cb87a940f54fb4b4c36738389e043006b38d198693" Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.930713 4632 scope.go:117] "RemoveContainer" containerID="ead807d406032ccc6de750261a0d5e2b71a645cfaa7f853ed7fc97cd308b03bb" Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.931779 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnpsr"] Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.959753 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-hnpsr"] Mar 13 12:22:10 crc kubenswrapper[4632]: I0313 12:22:10.977586 4632 scope.go:117] "RemoveContainer" containerID="030c12dcf041fb69ab0b030b3ab62bdd3a22b3a40f28da17a2039141caaeedee" Mar 13 12:22:11 crc kubenswrapper[4632]: I0313 12:22:11.030282 4632 scope.go:117] "RemoveContainer" containerID="59dd5b6704b2e6f99f2da9cb87a940f54fb4b4c36738389e043006b38d198693" Mar 13 12:22:11 crc kubenswrapper[4632]: E0313 12:22:11.030750 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59dd5b6704b2e6f99f2da9cb87a940f54fb4b4c36738389e043006b38d198693\": container with ID starting with 59dd5b6704b2e6f99f2da9cb87a940f54fb4b4c36738389e043006b38d198693 not found: ID does not exist" containerID="59dd5b6704b2e6f99f2da9cb87a940f54fb4b4c36738389e043006b38d198693" Mar 13 12:22:11 crc kubenswrapper[4632]: I0313 12:22:11.030789 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59dd5b6704b2e6f99f2da9cb87a940f54fb4b4c36738389e043006b38d198693"} err="failed to get container status \"59dd5b6704b2e6f99f2da9cb87a940f54fb4b4c36738389e043006b38d198693\": rpc error: code = NotFound desc = could not find container \"59dd5b6704b2e6f99f2da9cb87a940f54fb4b4c36738389e043006b38d198693\": container with ID starting with 59dd5b6704b2e6f99f2da9cb87a940f54fb4b4c36738389e043006b38d198693 not found: ID does not exist" Mar 13 12:22:11 crc kubenswrapper[4632]: I0313 12:22:11.030816 4632 scope.go:117] "RemoveContainer" containerID="ead807d406032ccc6de750261a0d5e2b71a645cfaa7f853ed7fc97cd308b03bb" Mar 13 12:22:11 crc kubenswrapper[4632]: E0313 12:22:11.031085 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ead807d406032ccc6de750261a0d5e2b71a645cfaa7f853ed7fc97cd308b03bb\": container with ID starting with ead807d406032ccc6de750261a0d5e2b71a645cfaa7f853ed7fc97cd308b03bb not found: ID does not exist" containerID="ead807d406032ccc6de750261a0d5e2b71a645cfaa7f853ed7fc97cd308b03bb" Mar 13 12:22:11 crc kubenswrapper[4632]: I0313 12:22:11.031116 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ead807d406032ccc6de750261a0d5e2b71a645cfaa7f853ed7fc97cd308b03bb"} err="failed to get container status \"ead807d406032ccc6de750261a0d5e2b71a645cfaa7f853ed7fc97cd308b03bb\": rpc error: code = NotFound desc = could not find container \"ead807d406032ccc6de750261a0d5e2b71a645cfaa7f853ed7fc97cd308b03bb\": container with ID starting with ead807d406032ccc6de750261a0d5e2b71a645cfaa7f853ed7fc97cd308b03bb not found: ID does not exist" Mar 13 12:22:11 crc kubenswrapper[4632]: I0313 12:22:11.031134 4632 scope.go:117] "RemoveContainer" containerID="030c12dcf041fb69ab0b030b3ab62bdd3a22b3a40f28da17a2039141caaeedee" Mar 13 12:22:11 crc kubenswrapper[4632]: E0313 12:22:11.031574 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"030c12dcf041fb69ab0b030b3ab62bdd3a22b3a40f28da17a2039141caaeedee\": container with ID starting with 030c12dcf041fb69ab0b030b3ab62bdd3a22b3a40f28da17a2039141caaeedee not found: ID does not exist" containerID="030c12dcf041fb69ab0b030b3ab62bdd3a22b3a40f28da17a2039141caaeedee" Mar 13 12:22:11 crc kubenswrapper[4632]: I0313 12:22:11.031604 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"030c12dcf041fb69ab0b030b3ab62bdd3a22b3a40f28da17a2039141caaeedee"} err="failed to get container status \"030c12dcf041fb69ab0b030b3ab62bdd3a22b3a40f28da17a2039141caaeedee\": rpc error: code = NotFound desc = could not find container \"030c12dcf041fb69ab0b030b3ab62bdd3a22b3a40f28da17a2039141caaeedee\": container with ID starting with 030c12dcf041fb69ab0b030b3ab62bdd3a22b3a40f28da17a2039141caaeedee not found: ID does not exist" Mar 13 12:22:12 crc kubenswrapper[4632]: I0313 12:22:12.061190 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e37141ae-e543-40fa-876c-e7b7d9e1598f" path="/var/lib/kubelet/pods/e37141ae-e543-40fa-876c-e7b7d9e1598f/volumes" Mar 13 12:22:26 crc kubenswrapper[4632]: I0313 12:22:26.349766 4632 scope.go:117] "RemoveContainer" containerID="d94c02e7a4464d40b765382224dccb8b8b7d7ff087751a6cc11225fceac593a0" Mar 13 12:22:40 crc kubenswrapper[4632]: I0313 12:22:40.461438 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:22:40 crc kubenswrapper[4632]: I0313 12:22:40.462849 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:22:46 crc kubenswrapper[4632]: I0313 12:22:46.747696 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6xp6q"] Mar 13 12:22:46 crc kubenswrapper[4632]: E0313 12:22:46.748495 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37141ae-e543-40fa-876c-e7b7d9e1598f" containerName="extract-content" Mar 13 12:22:46 crc kubenswrapper[4632]: I0313 12:22:46.748507 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37141ae-e543-40fa-876c-e7b7d9e1598f" containerName="extract-content" Mar 13 12:22:46 crc kubenswrapper[4632]: E0313 12:22:46.748535 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37141ae-e543-40fa-876c-e7b7d9e1598f" containerName="extract-utilities" Mar 13 12:22:46 crc kubenswrapper[4632]: I0313 12:22:46.748541 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37141ae-e543-40fa-876c-e7b7d9e1598f" containerName="extract-utilities" Mar 13 12:22:46 crc kubenswrapper[4632]: E0313 12:22:46.748564 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5121453-468a-432e-b110-fd0cd60ed92b" containerName="oc" Mar 13 12:22:46 crc kubenswrapper[4632]: I0313 12:22:46.748570 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5121453-468a-432e-b110-fd0cd60ed92b" containerName="oc" Mar 13 12:22:46 crc kubenswrapper[4632]: E0313 12:22:46.748584 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e37141ae-e543-40fa-876c-e7b7d9e1598f" containerName="registry-server" Mar 13 12:22:46 crc kubenswrapper[4632]: I0313 12:22:46.748590 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e37141ae-e543-40fa-876c-e7b7d9e1598f" containerName="registry-server" Mar 13 12:22:46 crc kubenswrapper[4632]: I0313 12:22:46.748752 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="e37141ae-e543-40fa-876c-e7b7d9e1598f" containerName="registry-server" Mar 13 12:22:46 crc kubenswrapper[4632]: I0313 12:22:46.748771 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5121453-468a-432e-b110-fd0cd60ed92b" containerName="oc" Mar 13 12:22:46 crc kubenswrapper[4632]: I0313 12:22:46.750099 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:22:46 crc kubenswrapper[4632]: I0313 12:22:46.766360 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6xp6q"] Mar 13 12:22:46 crc kubenswrapper[4632]: I0313 12:22:46.906838 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006513e3-67e6-4969-82fb-37e5ac8eaf4a-utilities\") pod \"certified-operators-6xp6q\" (UID: \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\") " pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:22:46 crc kubenswrapper[4632]: I0313 12:22:46.906920 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006513e3-67e6-4969-82fb-37e5ac8eaf4a-catalog-content\") pod \"certified-operators-6xp6q\" (UID: \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\") " pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:22:46 crc kubenswrapper[4632]: I0313 12:22:46.907203 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv4gx\" (UniqueName: \"kubernetes.io/projected/006513e3-67e6-4969-82fb-37e5ac8eaf4a-kube-api-access-sv4gx\") pod \"certified-operators-6xp6q\" (UID: \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\") " pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:22:47 crc kubenswrapper[4632]: I0313 12:22:47.008917 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006513e3-67e6-4969-82fb-37e5ac8eaf4a-catalog-content\") pod \"certified-operators-6xp6q\" (UID: \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\") " pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:22:47 crc kubenswrapper[4632]: I0313 12:22:47.009140 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv4gx\" (UniqueName: \"kubernetes.io/projected/006513e3-67e6-4969-82fb-37e5ac8eaf4a-kube-api-access-sv4gx\") pod \"certified-operators-6xp6q\" (UID: \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\") " pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:22:47 crc kubenswrapper[4632]: I0313 12:22:47.009187 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006513e3-67e6-4969-82fb-37e5ac8eaf4a-utilities\") pod \"certified-operators-6xp6q\" (UID: \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\") " pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:22:47 crc kubenswrapper[4632]: I0313 12:22:47.009530 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006513e3-67e6-4969-82fb-37e5ac8eaf4a-catalog-content\") pod \"certified-operators-6xp6q\" (UID: \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\") " pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:22:47 crc kubenswrapper[4632]: I0313 12:22:47.009566 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006513e3-67e6-4969-82fb-37e5ac8eaf4a-utilities\") pod \"certified-operators-6xp6q\" (UID: \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\") " pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:22:47 crc kubenswrapper[4632]: I0313 12:22:47.035345 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv4gx\" (UniqueName: \"kubernetes.io/projected/006513e3-67e6-4969-82fb-37e5ac8eaf4a-kube-api-access-sv4gx\") pod \"certified-operators-6xp6q\" (UID: \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\") " pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:22:47 crc kubenswrapper[4632]: I0313 12:22:47.079818 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:22:47 crc kubenswrapper[4632]: I0313 12:22:47.674776 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6xp6q"] Mar 13 12:22:48 crc kubenswrapper[4632]: I0313 12:22:48.295859 4632 generic.go:334] "Generic (PLEG): container finished" podID="006513e3-67e6-4969-82fb-37e5ac8eaf4a" containerID="5e9374890eb775e4c1bb97be5e3b2ff1cac17016a5c436abe4d78eea14dc7c46" exitCode=0 Mar 13 12:22:48 crc kubenswrapper[4632]: I0313 12:22:48.296128 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xp6q" event={"ID":"006513e3-67e6-4969-82fb-37e5ac8eaf4a","Type":"ContainerDied","Data":"5e9374890eb775e4c1bb97be5e3b2ff1cac17016a5c436abe4d78eea14dc7c46"} Mar 13 12:22:48 crc kubenswrapper[4632]: I0313 12:22:48.297094 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xp6q" event={"ID":"006513e3-67e6-4969-82fb-37e5ac8eaf4a","Type":"ContainerStarted","Data":"5567f226f31b002032305cc353d86ae94e1489dfc0d354345a4756c71df1ee5c"} Mar 13 12:22:50 crc kubenswrapper[4632]: I0313 12:22:50.316337 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xp6q" event={"ID":"006513e3-67e6-4969-82fb-37e5ac8eaf4a","Type":"ContainerStarted","Data":"45e1f9656cfbf9b22004031ee39b6063266926993e09e1b5fb34fc7a8ec2578a"} Mar 13 12:22:53 crc kubenswrapper[4632]: I0313 12:22:53.353533 4632 generic.go:334] "Generic (PLEG): container finished" podID="006513e3-67e6-4969-82fb-37e5ac8eaf4a" containerID="45e1f9656cfbf9b22004031ee39b6063266926993e09e1b5fb34fc7a8ec2578a" exitCode=0 Mar 13 12:22:53 crc kubenswrapper[4632]: I0313 12:22:53.353626 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xp6q" event={"ID":"006513e3-67e6-4969-82fb-37e5ac8eaf4a","Type":"ContainerDied","Data":"45e1f9656cfbf9b22004031ee39b6063266926993e09e1b5fb34fc7a8ec2578a"} Mar 13 12:22:54 crc kubenswrapper[4632]: I0313 12:22:54.375604 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xp6q" event={"ID":"006513e3-67e6-4969-82fb-37e5ac8eaf4a","Type":"ContainerStarted","Data":"99ba058ba824509ca07bb72d4c4f8e7b66dbe556f20ebb5fe9b5fbc3c4237556"} Mar 13 12:22:54 crc kubenswrapper[4632]: I0313 12:22:54.401302 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6xp6q" podStartSLOduration=2.931517953 podStartE2EDuration="8.401281921s" podCreationTimestamp="2026-03-13 12:22:46 +0000 UTC" firstStartedPulling="2026-03-13 12:22:48.298341294 +0000 UTC m=+8342.320871427" lastFinishedPulling="2026-03-13 12:22:53.768105262 +0000 UTC m=+8347.790635395" observedRunningTime="2026-03-13 12:22:54.39836616 +0000 UTC m=+8348.420896293" watchObservedRunningTime="2026-03-13 12:22:54.401281921 +0000 UTC m=+8348.423812054" Mar 13 12:22:57 crc kubenswrapper[4632]: I0313 12:22:57.080367 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:22:57 crc kubenswrapper[4632]: I0313 12:22:57.080806 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:22:58 crc kubenswrapper[4632]: I0313 12:22:58.139895 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-6xp6q" podUID="006513e3-67e6-4969-82fb-37e5ac8eaf4a" containerName="registry-server" probeResult="failure" output=< Mar 13 12:22:58 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:22:58 crc kubenswrapper[4632]: > Mar 13 12:23:07 crc kubenswrapper[4632]: I0313 12:23:07.146154 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:23:07 crc kubenswrapper[4632]: I0313 12:23:07.206355 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:23:07 crc kubenswrapper[4632]: I0313 12:23:07.386349 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6xp6q"] Mar 13 12:23:08 crc kubenswrapper[4632]: I0313 12:23:08.501274 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6xp6q" podUID="006513e3-67e6-4969-82fb-37e5ac8eaf4a" containerName="registry-server" containerID="cri-o://99ba058ba824509ca07bb72d4c4f8e7b66dbe556f20ebb5fe9b5fbc3c4237556" gracePeriod=2 Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.112212 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.267118 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv4gx\" (UniqueName: \"kubernetes.io/projected/006513e3-67e6-4969-82fb-37e5ac8eaf4a-kube-api-access-sv4gx\") pod \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\" (UID: \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\") " Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.267318 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006513e3-67e6-4969-82fb-37e5ac8eaf4a-utilities\") pod \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\" (UID: \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\") " Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.267502 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006513e3-67e6-4969-82fb-37e5ac8eaf4a-catalog-content\") pod \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\" (UID: \"006513e3-67e6-4969-82fb-37e5ac8eaf4a\") " Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.268158 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/006513e3-67e6-4969-82fb-37e5ac8eaf4a-utilities" (OuterVolumeSpecName: "utilities") pod "006513e3-67e6-4969-82fb-37e5ac8eaf4a" (UID: "006513e3-67e6-4969-82fb-37e5ac8eaf4a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.276286 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/006513e3-67e6-4969-82fb-37e5ac8eaf4a-kube-api-access-sv4gx" (OuterVolumeSpecName: "kube-api-access-sv4gx") pod "006513e3-67e6-4969-82fb-37e5ac8eaf4a" (UID: "006513e3-67e6-4969-82fb-37e5ac8eaf4a"). InnerVolumeSpecName "kube-api-access-sv4gx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.325961 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/006513e3-67e6-4969-82fb-37e5ac8eaf4a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "006513e3-67e6-4969-82fb-37e5ac8eaf4a" (UID: "006513e3-67e6-4969-82fb-37e5ac8eaf4a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.370202 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/006513e3-67e6-4969-82fb-37e5ac8eaf4a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.370241 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv4gx\" (UniqueName: \"kubernetes.io/projected/006513e3-67e6-4969-82fb-37e5ac8eaf4a-kube-api-access-sv4gx\") on node \"crc\" DevicePath \"\"" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.370253 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/006513e3-67e6-4969-82fb-37e5ac8eaf4a-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.511768 4632 generic.go:334] "Generic (PLEG): container finished" podID="006513e3-67e6-4969-82fb-37e5ac8eaf4a" containerID="99ba058ba824509ca07bb72d4c4f8e7b66dbe556f20ebb5fe9b5fbc3c4237556" exitCode=0 Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.511817 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xp6q" event={"ID":"006513e3-67e6-4969-82fb-37e5ac8eaf4a","Type":"ContainerDied","Data":"99ba058ba824509ca07bb72d4c4f8e7b66dbe556f20ebb5fe9b5fbc3c4237556"} Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.511845 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6xp6q" event={"ID":"006513e3-67e6-4969-82fb-37e5ac8eaf4a","Type":"ContainerDied","Data":"5567f226f31b002032305cc353d86ae94e1489dfc0d354345a4756c71df1ee5c"} Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.511864 4632 scope.go:117] "RemoveContainer" containerID="99ba058ba824509ca07bb72d4c4f8e7b66dbe556f20ebb5fe9b5fbc3c4237556" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.512023 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6xp6q" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.547846 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6xp6q"] Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.550698 4632 scope.go:117] "RemoveContainer" containerID="45e1f9656cfbf9b22004031ee39b6063266926993e09e1b5fb34fc7a8ec2578a" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.556164 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6xp6q"] Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.588230 4632 scope.go:117] "RemoveContainer" containerID="5e9374890eb775e4c1bb97be5e3b2ff1cac17016a5c436abe4d78eea14dc7c46" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.642370 4632 scope.go:117] "RemoveContainer" containerID="99ba058ba824509ca07bb72d4c4f8e7b66dbe556f20ebb5fe9b5fbc3c4237556" Mar 13 12:23:09 crc kubenswrapper[4632]: E0313 12:23:09.642859 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99ba058ba824509ca07bb72d4c4f8e7b66dbe556f20ebb5fe9b5fbc3c4237556\": container with ID starting with 99ba058ba824509ca07bb72d4c4f8e7b66dbe556f20ebb5fe9b5fbc3c4237556 not found: ID does not exist" containerID="99ba058ba824509ca07bb72d4c4f8e7b66dbe556f20ebb5fe9b5fbc3c4237556" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.642893 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99ba058ba824509ca07bb72d4c4f8e7b66dbe556f20ebb5fe9b5fbc3c4237556"} err="failed to get container status \"99ba058ba824509ca07bb72d4c4f8e7b66dbe556f20ebb5fe9b5fbc3c4237556\": rpc error: code = NotFound desc = could not find container \"99ba058ba824509ca07bb72d4c4f8e7b66dbe556f20ebb5fe9b5fbc3c4237556\": container with ID starting with 99ba058ba824509ca07bb72d4c4f8e7b66dbe556f20ebb5fe9b5fbc3c4237556 not found: ID does not exist" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.642923 4632 scope.go:117] "RemoveContainer" containerID="45e1f9656cfbf9b22004031ee39b6063266926993e09e1b5fb34fc7a8ec2578a" Mar 13 12:23:09 crc kubenswrapper[4632]: E0313 12:23:09.643288 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"45e1f9656cfbf9b22004031ee39b6063266926993e09e1b5fb34fc7a8ec2578a\": container with ID starting with 45e1f9656cfbf9b22004031ee39b6063266926993e09e1b5fb34fc7a8ec2578a not found: ID does not exist" containerID="45e1f9656cfbf9b22004031ee39b6063266926993e09e1b5fb34fc7a8ec2578a" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.643333 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"45e1f9656cfbf9b22004031ee39b6063266926993e09e1b5fb34fc7a8ec2578a"} err="failed to get container status \"45e1f9656cfbf9b22004031ee39b6063266926993e09e1b5fb34fc7a8ec2578a\": rpc error: code = NotFound desc = could not find container \"45e1f9656cfbf9b22004031ee39b6063266926993e09e1b5fb34fc7a8ec2578a\": container with ID starting with 45e1f9656cfbf9b22004031ee39b6063266926993e09e1b5fb34fc7a8ec2578a not found: ID does not exist" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.643366 4632 scope.go:117] "RemoveContainer" containerID="5e9374890eb775e4c1bb97be5e3b2ff1cac17016a5c436abe4d78eea14dc7c46" Mar 13 12:23:09 crc kubenswrapper[4632]: E0313 12:23:09.643788 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e9374890eb775e4c1bb97be5e3b2ff1cac17016a5c436abe4d78eea14dc7c46\": container with ID starting with 5e9374890eb775e4c1bb97be5e3b2ff1cac17016a5c436abe4d78eea14dc7c46 not found: ID does not exist" containerID="5e9374890eb775e4c1bb97be5e3b2ff1cac17016a5c436abe4d78eea14dc7c46" Mar 13 12:23:09 crc kubenswrapper[4632]: I0313 12:23:09.643875 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e9374890eb775e4c1bb97be5e3b2ff1cac17016a5c436abe4d78eea14dc7c46"} err="failed to get container status \"5e9374890eb775e4c1bb97be5e3b2ff1cac17016a5c436abe4d78eea14dc7c46\": rpc error: code = NotFound desc = could not find container \"5e9374890eb775e4c1bb97be5e3b2ff1cac17016a5c436abe4d78eea14dc7c46\": container with ID starting with 5e9374890eb775e4c1bb97be5e3b2ff1cac17016a5c436abe4d78eea14dc7c46 not found: ID does not exist" Mar 13 12:23:10 crc kubenswrapper[4632]: I0313 12:23:10.056552 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="006513e3-67e6-4969-82fb-37e5ac8eaf4a" path="/var/lib/kubelet/pods/006513e3-67e6-4969-82fb-37e5ac8eaf4a/volumes" Mar 13 12:23:10 crc kubenswrapper[4632]: I0313 12:23:10.461509 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:23:10 crc kubenswrapper[4632]: I0313 12:23:10.461882 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:23:10 crc kubenswrapper[4632]: I0313 12:23:10.462108 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 12:23:10 crc kubenswrapper[4632]: I0313 12:23:10.464780 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d2f7d92ea8336c364393ccfd7369387047df3a4555b1b7f7be871c5ae3268440"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 12:23:10 crc kubenswrapper[4632]: I0313 12:23:10.464998 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://d2f7d92ea8336c364393ccfd7369387047df3a4555b1b7f7be871c5ae3268440" gracePeriod=600 Mar 13 12:23:11 crc kubenswrapper[4632]: I0313 12:23:11.533022 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="d2f7d92ea8336c364393ccfd7369387047df3a4555b1b7f7be871c5ae3268440" exitCode=0 Mar 13 12:23:11 crc kubenswrapper[4632]: I0313 12:23:11.533081 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"d2f7d92ea8336c364393ccfd7369387047df3a4555b1b7f7be871c5ae3268440"} Mar 13 12:23:11 crc kubenswrapper[4632]: I0313 12:23:11.533567 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f"} Mar 13 12:23:11 crc kubenswrapper[4632]: I0313 12:23:11.533592 4632 scope.go:117] "RemoveContainer" containerID="5cc922706d30866ac208574ee6bcc0812dd5d20bcd356efd2bc6fcac169085a9" Mar 13 12:24:00 crc kubenswrapper[4632]: I0313 12:24:00.179553 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556744-fb92p"] Mar 13 12:24:00 crc kubenswrapper[4632]: E0313 12:24:00.180563 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="006513e3-67e6-4969-82fb-37e5ac8eaf4a" containerName="extract-utilities" Mar 13 12:24:00 crc kubenswrapper[4632]: I0313 12:24:00.180580 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="006513e3-67e6-4969-82fb-37e5ac8eaf4a" containerName="extract-utilities" Mar 13 12:24:00 crc kubenswrapper[4632]: E0313 12:24:00.180614 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="006513e3-67e6-4969-82fb-37e5ac8eaf4a" containerName="registry-server" Mar 13 12:24:00 crc kubenswrapper[4632]: I0313 12:24:00.180621 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="006513e3-67e6-4969-82fb-37e5ac8eaf4a" containerName="registry-server" Mar 13 12:24:00 crc kubenswrapper[4632]: E0313 12:24:00.180633 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="006513e3-67e6-4969-82fb-37e5ac8eaf4a" containerName="extract-content" Mar 13 12:24:00 crc kubenswrapper[4632]: I0313 12:24:00.180641 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="006513e3-67e6-4969-82fb-37e5ac8eaf4a" containerName="extract-content" Mar 13 12:24:00 crc kubenswrapper[4632]: I0313 12:24:00.180860 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="006513e3-67e6-4969-82fb-37e5ac8eaf4a" containerName="registry-server" Mar 13 12:24:00 crc kubenswrapper[4632]: I0313 12:24:00.181575 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556744-fb92p" Mar 13 12:24:00 crc kubenswrapper[4632]: I0313 12:24:00.183814 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:24:00 crc kubenswrapper[4632]: I0313 12:24:00.190867 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:24:00 crc kubenswrapper[4632]: I0313 12:24:00.191225 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:24:00 crc kubenswrapper[4632]: I0313 12:24:00.199636 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556744-fb92p"] Mar 13 12:24:00 crc kubenswrapper[4632]: I0313 12:24:00.343385 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvhdx\" (UniqueName: \"kubernetes.io/projected/e1006ed9-194b-4d1b-91cf-7722ce335023-kube-api-access-jvhdx\") pod \"auto-csr-approver-29556744-fb92p\" (UID: \"e1006ed9-194b-4d1b-91cf-7722ce335023\") " pod="openshift-infra/auto-csr-approver-29556744-fb92p" Mar 13 12:24:00 crc kubenswrapper[4632]: I0313 12:24:00.445900 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvhdx\" (UniqueName: \"kubernetes.io/projected/e1006ed9-194b-4d1b-91cf-7722ce335023-kube-api-access-jvhdx\") pod \"auto-csr-approver-29556744-fb92p\" (UID: \"e1006ed9-194b-4d1b-91cf-7722ce335023\") " pod="openshift-infra/auto-csr-approver-29556744-fb92p" Mar 13 12:24:00 crc kubenswrapper[4632]: I0313 12:24:00.468804 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvhdx\" (UniqueName: \"kubernetes.io/projected/e1006ed9-194b-4d1b-91cf-7722ce335023-kube-api-access-jvhdx\") pod \"auto-csr-approver-29556744-fb92p\" (UID: \"e1006ed9-194b-4d1b-91cf-7722ce335023\") " pod="openshift-infra/auto-csr-approver-29556744-fb92p" Mar 13 12:24:00 crc kubenswrapper[4632]: I0313 12:24:00.510737 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556744-fb92p" Mar 13 12:24:01 crc kubenswrapper[4632]: I0313 12:24:01.028145 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556744-fb92p"] Mar 13 12:24:01 crc kubenswrapper[4632]: I0313 12:24:01.886382 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556744-fb92p" event={"ID":"e1006ed9-194b-4d1b-91cf-7722ce335023","Type":"ContainerStarted","Data":"6e17f362d9dcb71b1160cdd1d9f6c0be75d7ea8d28408713bea723564ffc14ee"} Mar 13 12:24:02 crc kubenswrapper[4632]: I0313 12:24:02.895807 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556744-fb92p" event={"ID":"e1006ed9-194b-4d1b-91cf-7722ce335023","Type":"ContainerStarted","Data":"8c147fe4b276fdf885236433df734b60b43a6fbbd4e1c4d2a7bec9fd5c3cc6e2"} Mar 13 12:24:02 crc kubenswrapper[4632]: I0313 12:24:02.918270 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556744-fb92p" podStartSLOduration=1.645995664 podStartE2EDuration="2.918250546s" podCreationTimestamp="2026-03-13 12:24:00 +0000 UTC" firstStartedPulling="2026-03-13 12:24:01.035024288 +0000 UTC m=+8415.057554421" lastFinishedPulling="2026-03-13 12:24:02.30727917 +0000 UTC m=+8416.329809303" observedRunningTime="2026-03-13 12:24:02.909398798 +0000 UTC m=+8416.931928931" watchObservedRunningTime="2026-03-13 12:24:02.918250546 +0000 UTC m=+8416.940780679" Mar 13 12:24:04 crc kubenswrapper[4632]: I0313 12:24:04.915513 4632 generic.go:334] "Generic (PLEG): container finished" podID="e1006ed9-194b-4d1b-91cf-7722ce335023" containerID="8c147fe4b276fdf885236433df734b60b43a6fbbd4e1c4d2a7bec9fd5c3cc6e2" exitCode=0 Mar 13 12:24:04 crc kubenswrapper[4632]: I0313 12:24:04.915574 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556744-fb92p" event={"ID":"e1006ed9-194b-4d1b-91cf-7722ce335023","Type":"ContainerDied","Data":"8c147fe4b276fdf885236433df734b60b43a6fbbd4e1c4d2a7bec9fd5c3cc6e2"} Mar 13 12:24:06 crc kubenswrapper[4632]: I0313 12:24:06.399010 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556744-fb92p" Mar 13 12:24:06 crc kubenswrapper[4632]: I0313 12:24:06.579889 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvhdx\" (UniqueName: \"kubernetes.io/projected/e1006ed9-194b-4d1b-91cf-7722ce335023-kube-api-access-jvhdx\") pod \"e1006ed9-194b-4d1b-91cf-7722ce335023\" (UID: \"e1006ed9-194b-4d1b-91cf-7722ce335023\") " Mar 13 12:24:06 crc kubenswrapper[4632]: I0313 12:24:06.585549 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1006ed9-194b-4d1b-91cf-7722ce335023-kube-api-access-jvhdx" (OuterVolumeSpecName: "kube-api-access-jvhdx") pod "e1006ed9-194b-4d1b-91cf-7722ce335023" (UID: "e1006ed9-194b-4d1b-91cf-7722ce335023"). InnerVolumeSpecName "kube-api-access-jvhdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:24:06 crc kubenswrapper[4632]: I0313 12:24:06.682567 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvhdx\" (UniqueName: \"kubernetes.io/projected/e1006ed9-194b-4d1b-91cf-7722ce335023-kube-api-access-jvhdx\") on node \"crc\" DevicePath \"\"" Mar 13 12:24:06 crc kubenswrapper[4632]: I0313 12:24:06.936760 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556744-fb92p" event={"ID":"e1006ed9-194b-4d1b-91cf-7722ce335023","Type":"ContainerDied","Data":"6e17f362d9dcb71b1160cdd1d9f6c0be75d7ea8d28408713bea723564ffc14ee"} Mar 13 12:24:06 crc kubenswrapper[4632]: I0313 12:24:06.936818 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e17f362d9dcb71b1160cdd1d9f6c0be75d7ea8d28408713bea723564ffc14ee" Mar 13 12:24:06 crc kubenswrapper[4632]: I0313 12:24:06.936823 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556744-fb92p" Mar 13 12:24:07 crc kubenswrapper[4632]: I0313 12:24:07.012582 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556738-ktc6z"] Mar 13 12:24:07 crc kubenswrapper[4632]: I0313 12:24:07.023244 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556738-ktc6z"] Mar 13 12:24:08 crc kubenswrapper[4632]: I0313 12:24:08.056488 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fd947fd4-4e97-4720-98a3-d345ae5dd3fc" path="/var/lib/kubelet/pods/fd947fd4-4e97-4720-98a3-d345ae5dd3fc/volumes" Mar 13 12:24:26 crc kubenswrapper[4632]: I0313 12:24:26.603674 4632 scope.go:117] "RemoveContainer" containerID="fbfc844073b7954c305603f6ba9bca1ebae6e886287d4969b865a335340183e5" Mar 13 12:24:48 crc kubenswrapper[4632]: I0313 12:24:48.901067 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-5hbsx"] Mar 13 12:24:48 crc kubenswrapper[4632]: E0313 12:24:48.901831 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1006ed9-194b-4d1b-91cf-7722ce335023" containerName="oc" Mar 13 12:24:48 crc kubenswrapper[4632]: I0313 12:24:48.901844 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1006ed9-194b-4d1b-91cf-7722ce335023" containerName="oc" Mar 13 12:24:48 crc kubenswrapper[4632]: I0313 12:24:48.903320 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1006ed9-194b-4d1b-91cf-7722ce335023" containerName="oc" Mar 13 12:24:48 crc kubenswrapper[4632]: I0313 12:24:48.905493 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:24:48 crc kubenswrapper[4632]: I0313 12:24:48.924457 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5hbsx"] Mar 13 12:24:49 crc kubenswrapper[4632]: I0313 12:24:49.057451 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28d46ba0-2f3e-4780-8645-3551c59cbd90-catalog-content\") pod \"redhat-operators-5hbsx\" (UID: \"28d46ba0-2f3e-4780-8645-3551c59cbd90\") " pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:24:49 crc kubenswrapper[4632]: I0313 12:24:49.057514 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28d46ba0-2f3e-4780-8645-3551c59cbd90-utilities\") pod \"redhat-operators-5hbsx\" (UID: \"28d46ba0-2f3e-4780-8645-3551c59cbd90\") " pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:24:49 crc kubenswrapper[4632]: I0313 12:24:49.057559 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf7mj\" (UniqueName: \"kubernetes.io/projected/28d46ba0-2f3e-4780-8645-3551c59cbd90-kube-api-access-wf7mj\") pod \"redhat-operators-5hbsx\" (UID: \"28d46ba0-2f3e-4780-8645-3551c59cbd90\") " pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:24:49 crc kubenswrapper[4632]: I0313 12:24:49.159904 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28d46ba0-2f3e-4780-8645-3551c59cbd90-catalog-content\") pod \"redhat-operators-5hbsx\" (UID: \"28d46ba0-2f3e-4780-8645-3551c59cbd90\") " pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:24:49 crc kubenswrapper[4632]: I0313 12:24:49.159985 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28d46ba0-2f3e-4780-8645-3551c59cbd90-utilities\") pod \"redhat-operators-5hbsx\" (UID: \"28d46ba0-2f3e-4780-8645-3551c59cbd90\") " pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:24:49 crc kubenswrapper[4632]: I0313 12:24:49.160017 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf7mj\" (UniqueName: \"kubernetes.io/projected/28d46ba0-2f3e-4780-8645-3551c59cbd90-kube-api-access-wf7mj\") pod \"redhat-operators-5hbsx\" (UID: \"28d46ba0-2f3e-4780-8645-3551c59cbd90\") " pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:24:49 crc kubenswrapper[4632]: I0313 12:24:49.160104 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28d46ba0-2f3e-4780-8645-3551c59cbd90-catalog-content\") pod \"redhat-operators-5hbsx\" (UID: \"28d46ba0-2f3e-4780-8645-3551c59cbd90\") " pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:24:49 crc kubenswrapper[4632]: I0313 12:24:49.160275 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28d46ba0-2f3e-4780-8645-3551c59cbd90-utilities\") pod \"redhat-operators-5hbsx\" (UID: \"28d46ba0-2f3e-4780-8645-3551c59cbd90\") " pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:24:49 crc kubenswrapper[4632]: I0313 12:24:49.180028 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf7mj\" (UniqueName: \"kubernetes.io/projected/28d46ba0-2f3e-4780-8645-3551c59cbd90-kube-api-access-wf7mj\") pod \"redhat-operators-5hbsx\" (UID: \"28d46ba0-2f3e-4780-8645-3551c59cbd90\") " pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:24:49 crc kubenswrapper[4632]: I0313 12:24:49.230553 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:24:49 crc kubenswrapper[4632]: I0313 12:24:49.856099 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-5hbsx"] Mar 13 12:24:50 crc kubenswrapper[4632]: I0313 12:24:50.343454 4632 generic.go:334] "Generic (PLEG): container finished" podID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerID="7be8436687bd0d966081b362829fe60e928b205102fb92b02070bb54e66a10a9" exitCode=0 Mar 13 12:24:50 crc kubenswrapper[4632]: I0313 12:24:50.343516 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5hbsx" event={"ID":"28d46ba0-2f3e-4780-8645-3551c59cbd90","Type":"ContainerDied","Data":"7be8436687bd0d966081b362829fe60e928b205102fb92b02070bb54e66a10a9"} Mar 13 12:24:50 crc kubenswrapper[4632]: I0313 12:24:50.343750 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5hbsx" event={"ID":"28d46ba0-2f3e-4780-8645-3551c59cbd90","Type":"ContainerStarted","Data":"6694309764a236c0cf4af5875ea9754ac8fe7e1c78280a3c51d6648b38d11efe"} Mar 13 12:24:52 crc kubenswrapper[4632]: I0313 12:24:52.372329 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5hbsx" event={"ID":"28d46ba0-2f3e-4780-8645-3551c59cbd90","Type":"ContainerStarted","Data":"f2b0a8aae3cf593228d6090f4f4dcba61eee1aa4e51a3ec5b41ca13e73160ca4"} Mar 13 12:24:57 crc kubenswrapper[4632]: I0313 12:24:57.422172 4632 generic.go:334] "Generic (PLEG): container finished" podID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerID="f2b0a8aae3cf593228d6090f4f4dcba61eee1aa4e51a3ec5b41ca13e73160ca4" exitCode=0 Mar 13 12:24:57 crc kubenswrapper[4632]: I0313 12:24:57.422616 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5hbsx" event={"ID":"28d46ba0-2f3e-4780-8645-3551c59cbd90","Type":"ContainerDied","Data":"f2b0a8aae3cf593228d6090f4f4dcba61eee1aa4e51a3ec5b41ca13e73160ca4"} Mar 13 12:24:58 crc kubenswrapper[4632]: I0313 12:24:58.440918 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5hbsx" event={"ID":"28d46ba0-2f3e-4780-8645-3551c59cbd90","Type":"ContainerStarted","Data":"22b880bd9d36239a72deab6ba7c23eda91f4380144527c59c20491394e33da83"} Mar 13 12:24:58 crc kubenswrapper[4632]: I0313 12:24:58.473662 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-5hbsx" podStartSLOduration=2.9906351239999998 podStartE2EDuration="10.473634307s" podCreationTimestamp="2026-03-13 12:24:48 +0000 UTC" firstStartedPulling="2026-03-13 12:24:50.346375482 +0000 UTC m=+8464.368905616" lastFinishedPulling="2026-03-13 12:24:57.829374666 +0000 UTC m=+8471.851904799" observedRunningTime="2026-03-13 12:24:58.467667891 +0000 UTC m=+8472.490198034" watchObservedRunningTime="2026-03-13 12:24:58.473634307 +0000 UTC m=+8472.496164440" Mar 13 12:24:59 crc kubenswrapper[4632]: I0313 12:24:59.230826 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:24:59 crc kubenswrapper[4632]: I0313 12:24:59.230890 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:25:00 crc kubenswrapper[4632]: I0313 12:25:00.278956 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5hbsx" podUID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerName="registry-server" probeResult="failure" output=< Mar 13 12:25:00 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:25:00 crc kubenswrapper[4632]: > Mar 13 12:25:10 crc kubenswrapper[4632]: I0313 12:25:10.297108 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5hbsx" podUID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerName="registry-server" probeResult="failure" output=< Mar 13 12:25:10 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:25:10 crc kubenswrapper[4632]: > Mar 13 12:25:10 crc kubenswrapper[4632]: I0313 12:25:10.461272 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:25:10 crc kubenswrapper[4632]: I0313 12:25:10.461426 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:25:20 crc kubenswrapper[4632]: I0313 12:25:20.279960 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5hbsx" podUID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerName="registry-server" probeResult="failure" output=< Mar 13 12:25:20 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:25:20 crc kubenswrapper[4632]: > Mar 13 12:25:30 crc kubenswrapper[4632]: I0313 12:25:30.307710 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-5hbsx" podUID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerName="registry-server" probeResult="failure" output=< Mar 13 12:25:30 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:25:30 crc kubenswrapper[4632]: > Mar 13 12:25:39 crc kubenswrapper[4632]: I0313 12:25:39.292113 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:25:39 crc kubenswrapper[4632]: I0313 12:25:39.355158 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:25:39 crc kubenswrapper[4632]: I0313 12:25:39.537565 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5hbsx"] Mar 13 12:25:40 crc kubenswrapper[4632]: I0313 12:25:40.461509 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:25:40 crc kubenswrapper[4632]: I0313 12:25:40.461807 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:25:40 crc kubenswrapper[4632]: I0313 12:25:40.865303 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-5hbsx" podUID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerName="registry-server" containerID="cri-o://22b880bd9d36239a72deab6ba7c23eda91f4380144527c59c20491394e33da83" gracePeriod=2 Mar 13 12:25:41 crc kubenswrapper[4632]: I0313 12:25:41.840214 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:25:41 crc kubenswrapper[4632]: I0313 12:25:41.878961 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-5hbsx" Mar 13 12:25:41 crc kubenswrapper[4632]: I0313 12:25:41.879099 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5hbsx" event={"ID":"28d46ba0-2f3e-4780-8645-3551c59cbd90","Type":"ContainerDied","Data":"22b880bd9d36239a72deab6ba7c23eda91f4380144527c59c20491394e33da83"} Mar 13 12:25:41 crc kubenswrapper[4632]: I0313 12:25:41.879694 4632 generic.go:334] "Generic (PLEG): container finished" podID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerID="22b880bd9d36239a72deab6ba7c23eda91f4380144527c59c20491394e33da83" exitCode=0 Mar 13 12:25:41 crc kubenswrapper[4632]: I0313 12:25:41.879732 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-5hbsx" event={"ID":"28d46ba0-2f3e-4780-8645-3551c59cbd90","Type":"ContainerDied","Data":"6694309764a236c0cf4af5875ea9754ac8fe7e1c78280a3c51d6648b38d11efe"} Mar 13 12:25:41 crc kubenswrapper[4632]: I0313 12:25:41.882128 4632 scope.go:117] "RemoveContainer" containerID="22b880bd9d36239a72deab6ba7c23eda91f4380144527c59c20491394e33da83" Mar 13 12:25:41 crc kubenswrapper[4632]: I0313 12:25:41.927785 4632 scope.go:117] "RemoveContainer" containerID="f2b0a8aae3cf593228d6090f4f4dcba61eee1aa4e51a3ec5b41ca13e73160ca4" Mar 13 12:25:41 crc kubenswrapper[4632]: I0313 12:25:41.981313 4632 scope.go:117] "RemoveContainer" containerID="7be8436687bd0d966081b362829fe60e928b205102fb92b02070bb54e66a10a9" Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.026464 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf7mj\" (UniqueName: \"kubernetes.io/projected/28d46ba0-2f3e-4780-8645-3551c59cbd90-kube-api-access-wf7mj\") pod \"28d46ba0-2f3e-4780-8645-3551c59cbd90\" (UID: \"28d46ba0-2f3e-4780-8645-3551c59cbd90\") " Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.026630 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28d46ba0-2f3e-4780-8645-3551c59cbd90-catalog-content\") pod \"28d46ba0-2f3e-4780-8645-3551c59cbd90\" (UID: \"28d46ba0-2f3e-4780-8645-3551c59cbd90\") " Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.026667 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28d46ba0-2f3e-4780-8645-3551c59cbd90-utilities\") pod \"28d46ba0-2f3e-4780-8645-3551c59cbd90\" (UID: \"28d46ba0-2f3e-4780-8645-3551c59cbd90\") " Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.032200 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28d46ba0-2f3e-4780-8645-3551c59cbd90-utilities" (OuterVolumeSpecName: "utilities") pod "28d46ba0-2f3e-4780-8645-3551c59cbd90" (UID: "28d46ba0-2f3e-4780-8645-3551c59cbd90"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.037489 4632 scope.go:117] "RemoveContainer" containerID="22b880bd9d36239a72deab6ba7c23eda91f4380144527c59c20491394e33da83" Mar 13 12:25:42 crc kubenswrapper[4632]: E0313 12:25:42.046710 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22b880bd9d36239a72deab6ba7c23eda91f4380144527c59c20491394e33da83\": container with ID starting with 22b880bd9d36239a72deab6ba7c23eda91f4380144527c59c20491394e33da83 not found: ID does not exist" containerID="22b880bd9d36239a72deab6ba7c23eda91f4380144527c59c20491394e33da83" Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.048248 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22b880bd9d36239a72deab6ba7c23eda91f4380144527c59c20491394e33da83"} err="failed to get container status \"22b880bd9d36239a72deab6ba7c23eda91f4380144527c59c20491394e33da83\": rpc error: code = NotFound desc = could not find container \"22b880bd9d36239a72deab6ba7c23eda91f4380144527c59c20491394e33da83\": container with ID starting with 22b880bd9d36239a72deab6ba7c23eda91f4380144527c59c20491394e33da83 not found: ID does not exist" Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.048288 4632 scope.go:117] "RemoveContainer" containerID="f2b0a8aae3cf593228d6090f4f4dcba61eee1aa4e51a3ec5b41ca13e73160ca4" Mar 13 12:25:42 crc kubenswrapper[4632]: E0313 12:25:42.048883 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f2b0a8aae3cf593228d6090f4f4dcba61eee1aa4e51a3ec5b41ca13e73160ca4\": container with ID starting with f2b0a8aae3cf593228d6090f4f4dcba61eee1aa4e51a3ec5b41ca13e73160ca4 not found: ID does not exist" containerID="f2b0a8aae3cf593228d6090f4f4dcba61eee1aa4e51a3ec5b41ca13e73160ca4" Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.048922 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f2b0a8aae3cf593228d6090f4f4dcba61eee1aa4e51a3ec5b41ca13e73160ca4"} err="failed to get container status \"f2b0a8aae3cf593228d6090f4f4dcba61eee1aa4e51a3ec5b41ca13e73160ca4\": rpc error: code = NotFound desc = could not find container \"f2b0a8aae3cf593228d6090f4f4dcba61eee1aa4e51a3ec5b41ca13e73160ca4\": container with ID starting with f2b0a8aae3cf593228d6090f4f4dcba61eee1aa4e51a3ec5b41ca13e73160ca4 not found: ID does not exist" Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.049006 4632 scope.go:117] "RemoveContainer" containerID="7be8436687bd0d966081b362829fe60e928b205102fb92b02070bb54e66a10a9" Mar 13 12:25:42 crc kubenswrapper[4632]: E0313 12:25:42.049525 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7be8436687bd0d966081b362829fe60e928b205102fb92b02070bb54e66a10a9\": container with ID starting with 7be8436687bd0d966081b362829fe60e928b205102fb92b02070bb54e66a10a9 not found: ID does not exist" containerID="7be8436687bd0d966081b362829fe60e928b205102fb92b02070bb54e66a10a9" Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.049555 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7be8436687bd0d966081b362829fe60e928b205102fb92b02070bb54e66a10a9"} err="failed to get container status \"7be8436687bd0d966081b362829fe60e928b205102fb92b02070bb54e66a10a9\": rpc error: code = NotFound desc = could not find container \"7be8436687bd0d966081b362829fe60e928b205102fb92b02070bb54e66a10a9\": container with ID starting with 7be8436687bd0d966081b362829fe60e928b205102fb92b02070bb54e66a10a9 not found: ID does not exist" Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.055967 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28d46ba0-2f3e-4780-8645-3551c59cbd90-kube-api-access-wf7mj" (OuterVolumeSpecName: "kube-api-access-wf7mj") pod "28d46ba0-2f3e-4780-8645-3551c59cbd90" (UID: "28d46ba0-2f3e-4780-8645-3551c59cbd90"). InnerVolumeSpecName "kube-api-access-wf7mj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.129696 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf7mj\" (UniqueName: \"kubernetes.io/projected/28d46ba0-2f3e-4780-8645-3551c59cbd90-kube-api-access-wf7mj\") on node \"crc\" DevicePath \"\"" Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.129735 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28d46ba0-2f3e-4780-8645-3551c59cbd90-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.274778 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28d46ba0-2f3e-4780-8645-3551c59cbd90-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "28d46ba0-2f3e-4780-8645-3551c59cbd90" (UID: "28d46ba0-2f3e-4780-8645-3551c59cbd90"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.334130 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28d46ba0-2f3e-4780-8645-3551c59cbd90-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.526895 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-5hbsx"] Mar 13 12:25:42 crc kubenswrapper[4632]: I0313 12:25:42.538251 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-5hbsx"] Mar 13 12:25:44 crc kubenswrapper[4632]: I0313 12:25:44.056781 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28d46ba0-2f3e-4780-8645-3551c59cbd90" path="/var/lib/kubelet/pods/28d46ba0-2f3e-4780-8645-3551c59cbd90/volumes" Mar 13 12:26:00 crc kubenswrapper[4632]: I0313 12:26:00.169221 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556746-jtv7q"] Mar 13 12:26:00 crc kubenswrapper[4632]: E0313 12:26:00.172754 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerName="registry-server" Mar 13 12:26:00 crc kubenswrapper[4632]: I0313 12:26:00.172810 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerName="registry-server" Mar 13 12:26:00 crc kubenswrapper[4632]: E0313 12:26:00.172846 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerName="extract-content" Mar 13 12:26:00 crc kubenswrapper[4632]: I0313 12:26:00.172859 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerName="extract-content" Mar 13 12:26:00 crc kubenswrapper[4632]: E0313 12:26:00.172920 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerName="extract-utilities" Mar 13 12:26:00 crc kubenswrapper[4632]: I0313 12:26:00.172936 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerName="extract-utilities" Mar 13 12:26:00 crc kubenswrapper[4632]: I0313 12:26:00.174248 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="28d46ba0-2f3e-4780-8645-3551c59cbd90" containerName="registry-server" Mar 13 12:26:00 crc kubenswrapper[4632]: I0313 12:26:00.181348 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556746-jtv7q" Mar 13 12:26:00 crc kubenswrapper[4632]: I0313 12:26:00.190867 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556746-jtv7q"] Mar 13 12:26:00 crc kubenswrapper[4632]: I0313 12:26:00.200769 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:26:00 crc kubenswrapper[4632]: I0313 12:26:00.200796 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:26:00 crc kubenswrapper[4632]: I0313 12:26:00.200775 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:26:00 crc kubenswrapper[4632]: I0313 12:26:00.301082 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2gs2\" (UniqueName: \"kubernetes.io/projected/d88882b5-a11f-4606-8e0d-59471c7feccb-kube-api-access-p2gs2\") pod \"auto-csr-approver-29556746-jtv7q\" (UID: \"d88882b5-a11f-4606-8e0d-59471c7feccb\") " pod="openshift-infra/auto-csr-approver-29556746-jtv7q" Mar 13 12:26:00 crc kubenswrapper[4632]: I0313 12:26:00.402790 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2gs2\" (UniqueName: \"kubernetes.io/projected/d88882b5-a11f-4606-8e0d-59471c7feccb-kube-api-access-p2gs2\") pod \"auto-csr-approver-29556746-jtv7q\" (UID: \"d88882b5-a11f-4606-8e0d-59471c7feccb\") " pod="openshift-infra/auto-csr-approver-29556746-jtv7q" Mar 13 12:26:00 crc kubenswrapper[4632]: I0313 12:26:00.426591 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2gs2\" (UniqueName: \"kubernetes.io/projected/d88882b5-a11f-4606-8e0d-59471c7feccb-kube-api-access-p2gs2\") pod \"auto-csr-approver-29556746-jtv7q\" (UID: \"d88882b5-a11f-4606-8e0d-59471c7feccb\") " pod="openshift-infra/auto-csr-approver-29556746-jtv7q" Mar 13 12:26:00 crc kubenswrapper[4632]: I0313 12:26:00.512445 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556746-jtv7q" Mar 13 12:26:01 crc kubenswrapper[4632]: W0313 12:26:01.021747 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd88882b5_a11f_4606_8e0d_59471c7feccb.slice/crio-8624fa59086784980c1fd5b3b7f5d596dd40965d2322849d5725dab0f4736adc WatchSource:0}: Error finding container 8624fa59086784980c1fd5b3b7f5d596dd40965d2322849d5725dab0f4736adc: Status 404 returned error can't find the container with id 8624fa59086784980c1fd5b3b7f5d596dd40965d2322849d5725dab0f4736adc Mar 13 12:26:01 crc kubenswrapper[4632]: I0313 12:26:01.029624 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 12:26:01 crc kubenswrapper[4632]: I0313 12:26:01.029827 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556746-jtv7q"] Mar 13 12:26:01 crc kubenswrapper[4632]: I0313 12:26:01.103399 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556746-jtv7q" event={"ID":"d88882b5-a11f-4606-8e0d-59471c7feccb","Type":"ContainerStarted","Data":"8624fa59086784980c1fd5b3b7f5d596dd40965d2322849d5725dab0f4736adc"} Mar 13 12:26:03 crc kubenswrapper[4632]: I0313 12:26:03.126234 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556746-jtv7q" event={"ID":"d88882b5-a11f-4606-8e0d-59471c7feccb","Type":"ContainerStarted","Data":"35ea86f2c8d1955a868a4d03ab725c4feb194878cc720326b8eb0c50ed5ce3c5"} Mar 13 12:26:03 crc kubenswrapper[4632]: I0313 12:26:03.148159 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556746-jtv7q" podStartSLOduration=2.244168615 podStartE2EDuration="3.148140896s" podCreationTimestamp="2026-03-13 12:26:00 +0000 UTC" firstStartedPulling="2026-03-13 12:26:01.024620344 +0000 UTC m=+8535.047150477" lastFinishedPulling="2026-03-13 12:26:01.928592625 +0000 UTC m=+8535.951122758" observedRunningTime="2026-03-13 12:26:03.139024862 +0000 UTC m=+8537.161555025" watchObservedRunningTime="2026-03-13 12:26:03.148140896 +0000 UTC m=+8537.170671019" Mar 13 12:26:04 crc kubenswrapper[4632]: I0313 12:26:04.137154 4632 generic.go:334] "Generic (PLEG): container finished" podID="d88882b5-a11f-4606-8e0d-59471c7feccb" containerID="35ea86f2c8d1955a868a4d03ab725c4feb194878cc720326b8eb0c50ed5ce3c5" exitCode=0 Mar 13 12:26:04 crc kubenswrapper[4632]: I0313 12:26:04.137231 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556746-jtv7q" event={"ID":"d88882b5-a11f-4606-8e0d-59471c7feccb","Type":"ContainerDied","Data":"35ea86f2c8d1955a868a4d03ab725c4feb194878cc720326b8eb0c50ed5ce3c5"} Mar 13 12:26:05 crc kubenswrapper[4632]: I0313 12:26:05.541304 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556746-jtv7q" Mar 13 12:26:05 crc kubenswrapper[4632]: I0313 12:26:05.620142 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2gs2\" (UniqueName: \"kubernetes.io/projected/d88882b5-a11f-4606-8e0d-59471c7feccb-kube-api-access-p2gs2\") pod \"d88882b5-a11f-4606-8e0d-59471c7feccb\" (UID: \"d88882b5-a11f-4606-8e0d-59471c7feccb\") " Mar 13 12:26:05 crc kubenswrapper[4632]: I0313 12:26:05.626362 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d88882b5-a11f-4606-8e0d-59471c7feccb-kube-api-access-p2gs2" (OuterVolumeSpecName: "kube-api-access-p2gs2") pod "d88882b5-a11f-4606-8e0d-59471c7feccb" (UID: "d88882b5-a11f-4606-8e0d-59471c7feccb"). InnerVolumeSpecName "kube-api-access-p2gs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:26:05 crc kubenswrapper[4632]: I0313 12:26:05.723044 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2gs2\" (UniqueName: \"kubernetes.io/projected/d88882b5-a11f-4606-8e0d-59471c7feccb-kube-api-access-p2gs2\") on node \"crc\" DevicePath \"\"" Mar 13 12:26:06 crc kubenswrapper[4632]: I0313 12:26:06.158083 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556746-jtv7q" event={"ID":"d88882b5-a11f-4606-8e0d-59471c7feccb","Type":"ContainerDied","Data":"8624fa59086784980c1fd5b3b7f5d596dd40965d2322849d5725dab0f4736adc"} Mar 13 12:26:06 crc kubenswrapper[4632]: I0313 12:26:06.158131 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556746-jtv7q" Mar 13 12:26:06 crc kubenswrapper[4632]: I0313 12:26:06.158163 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8624fa59086784980c1fd5b3b7f5d596dd40965d2322849d5725dab0f4736adc" Mar 13 12:26:06 crc kubenswrapper[4632]: I0313 12:26:06.242382 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556740-fmh5b"] Mar 13 12:26:06 crc kubenswrapper[4632]: I0313 12:26:06.254193 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556740-fmh5b"] Mar 13 12:26:08 crc kubenswrapper[4632]: I0313 12:26:08.061703 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c" path="/var/lib/kubelet/pods/0f86aa5e-9cfc-458f-ae11-71e5e4dcfe9c/volumes" Mar 13 12:26:10 crc kubenswrapper[4632]: I0313 12:26:10.461581 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:26:10 crc kubenswrapper[4632]: I0313 12:26:10.461687 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:26:10 crc kubenswrapper[4632]: I0313 12:26:10.461753 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 12:26:10 crc kubenswrapper[4632]: I0313 12:26:10.471698 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 12:26:10 crc kubenswrapper[4632]: I0313 12:26:10.471866 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" gracePeriod=600 Mar 13 12:26:10 crc kubenswrapper[4632]: E0313 12:26:10.661535 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:26:11 crc kubenswrapper[4632]: I0313 12:26:11.217860 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" exitCode=0 Mar 13 12:26:11 crc kubenswrapper[4632]: I0313 12:26:11.217907 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f"} Mar 13 12:26:11 crc kubenswrapper[4632]: I0313 12:26:11.217974 4632 scope.go:117] "RemoveContainer" containerID="d2f7d92ea8336c364393ccfd7369387047df3a4555b1b7f7be871c5ae3268440" Mar 13 12:26:11 crc kubenswrapper[4632]: I0313 12:26:11.218601 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:26:11 crc kubenswrapper[4632]: E0313 12:26:11.218953 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:26:25 crc kubenswrapper[4632]: I0313 12:26:25.044922 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:26:25 crc kubenswrapper[4632]: E0313 12:26:25.045719 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:26:26 crc kubenswrapper[4632]: I0313 12:26:26.769682 4632 scope.go:117] "RemoveContainer" containerID="e0e701c935a2c4084fd4e093f0c21450f3afd1589228584f67fcd3cbe4d41395" Mar 13 12:26:37 crc kubenswrapper[4632]: I0313 12:26:37.044447 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:26:37 crc kubenswrapper[4632]: E0313 12:26:37.045489 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:26:50 crc kubenswrapper[4632]: I0313 12:26:50.045411 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:26:50 crc kubenswrapper[4632]: E0313 12:26:50.046156 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:27:02 crc kubenswrapper[4632]: I0313 12:27:02.044222 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:27:02 crc kubenswrapper[4632]: E0313 12:27:02.045134 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:27:13 crc kubenswrapper[4632]: I0313 12:27:13.045480 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:27:13 crc kubenswrapper[4632]: E0313 12:27:13.046411 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:27:25 crc kubenswrapper[4632]: I0313 12:27:25.044218 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:27:25 crc kubenswrapper[4632]: E0313 12:27:25.044969 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:27:37 crc kubenswrapper[4632]: I0313 12:27:37.045055 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:27:37 crc kubenswrapper[4632]: E0313 12:27:37.046382 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:27:51 crc kubenswrapper[4632]: I0313 12:27:51.044316 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:27:51 crc kubenswrapper[4632]: E0313 12:27:51.045257 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:28:00 crc kubenswrapper[4632]: I0313 12:28:00.151580 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556748-tr74t"] Mar 13 12:28:00 crc kubenswrapper[4632]: E0313 12:28:00.152465 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d88882b5-a11f-4606-8e0d-59471c7feccb" containerName="oc" Mar 13 12:28:00 crc kubenswrapper[4632]: I0313 12:28:00.152479 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="d88882b5-a11f-4606-8e0d-59471c7feccb" containerName="oc" Mar 13 12:28:00 crc kubenswrapper[4632]: I0313 12:28:00.152755 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="d88882b5-a11f-4606-8e0d-59471c7feccb" containerName="oc" Mar 13 12:28:00 crc kubenswrapper[4632]: I0313 12:28:00.153373 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556748-tr74t" Mar 13 12:28:00 crc kubenswrapper[4632]: I0313 12:28:00.157610 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:28:00 crc kubenswrapper[4632]: I0313 12:28:00.157882 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:28:00 crc kubenswrapper[4632]: I0313 12:28:00.158486 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:28:00 crc kubenswrapper[4632]: I0313 12:28:00.163234 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556748-tr74t"] Mar 13 12:28:00 crc kubenswrapper[4632]: I0313 12:28:00.300133 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjbpg\" (UniqueName: \"kubernetes.io/projected/ba59e3a2-3d83-4f9d-8633-788ba1bf518c-kube-api-access-mjbpg\") pod \"auto-csr-approver-29556748-tr74t\" (UID: \"ba59e3a2-3d83-4f9d-8633-788ba1bf518c\") " pod="openshift-infra/auto-csr-approver-29556748-tr74t" Mar 13 12:28:00 crc kubenswrapper[4632]: I0313 12:28:00.401753 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjbpg\" (UniqueName: \"kubernetes.io/projected/ba59e3a2-3d83-4f9d-8633-788ba1bf518c-kube-api-access-mjbpg\") pod \"auto-csr-approver-29556748-tr74t\" (UID: \"ba59e3a2-3d83-4f9d-8633-788ba1bf518c\") " pod="openshift-infra/auto-csr-approver-29556748-tr74t" Mar 13 12:28:00 crc kubenswrapper[4632]: I0313 12:28:00.419432 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjbpg\" (UniqueName: \"kubernetes.io/projected/ba59e3a2-3d83-4f9d-8633-788ba1bf518c-kube-api-access-mjbpg\") pod \"auto-csr-approver-29556748-tr74t\" (UID: \"ba59e3a2-3d83-4f9d-8633-788ba1bf518c\") " pod="openshift-infra/auto-csr-approver-29556748-tr74t" Mar 13 12:28:00 crc kubenswrapper[4632]: I0313 12:28:00.481315 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556748-tr74t" Mar 13 12:28:01 crc kubenswrapper[4632]: I0313 12:28:01.021175 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556748-tr74t"] Mar 13 12:28:01 crc kubenswrapper[4632]: I0313 12:28:01.289819 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556748-tr74t" event={"ID":"ba59e3a2-3d83-4f9d-8633-788ba1bf518c","Type":"ContainerStarted","Data":"156d224a7e86839fcd1fe5b72dad72deacf75274b9199281fc604f2724238ed2"} Mar 13 12:28:02 crc kubenswrapper[4632]: I0313 12:28:02.300019 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556748-tr74t" event={"ID":"ba59e3a2-3d83-4f9d-8633-788ba1bf518c","Type":"ContainerStarted","Data":"d4da70cb5943a7b88f9744961515085bb09badfa367e58e3aee43668a0864bc3"} Mar 13 12:28:02 crc kubenswrapper[4632]: I0313 12:28:02.316574 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556748-tr74t" podStartSLOduration=1.488017138 podStartE2EDuration="2.316551869s" podCreationTimestamp="2026-03-13 12:28:00 +0000 UTC" firstStartedPulling="2026-03-13 12:28:01.000335578 +0000 UTC m=+8655.022865721" lastFinishedPulling="2026-03-13 12:28:01.828870319 +0000 UTC m=+8655.851400452" observedRunningTime="2026-03-13 12:28:02.313801092 +0000 UTC m=+8656.336331225" watchObservedRunningTime="2026-03-13 12:28:02.316551869 +0000 UTC m=+8656.339082022" Mar 13 12:28:03 crc kubenswrapper[4632]: I0313 12:28:03.321596 4632 generic.go:334] "Generic (PLEG): container finished" podID="ba59e3a2-3d83-4f9d-8633-788ba1bf518c" containerID="d4da70cb5943a7b88f9744961515085bb09badfa367e58e3aee43668a0864bc3" exitCode=0 Mar 13 12:28:03 crc kubenswrapper[4632]: I0313 12:28:03.321911 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556748-tr74t" event={"ID":"ba59e3a2-3d83-4f9d-8633-788ba1bf518c","Type":"ContainerDied","Data":"d4da70cb5943a7b88f9744961515085bb09badfa367e58e3aee43668a0864bc3"} Mar 13 12:28:04 crc kubenswrapper[4632]: I0313 12:28:04.713184 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556748-tr74t" Mar 13 12:28:04 crc kubenswrapper[4632]: I0313 12:28:04.907147 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mjbpg\" (UniqueName: \"kubernetes.io/projected/ba59e3a2-3d83-4f9d-8633-788ba1bf518c-kube-api-access-mjbpg\") pod \"ba59e3a2-3d83-4f9d-8633-788ba1bf518c\" (UID: \"ba59e3a2-3d83-4f9d-8633-788ba1bf518c\") " Mar 13 12:28:04 crc kubenswrapper[4632]: I0313 12:28:04.917591 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba59e3a2-3d83-4f9d-8633-788ba1bf518c-kube-api-access-mjbpg" (OuterVolumeSpecName: "kube-api-access-mjbpg") pod "ba59e3a2-3d83-4f9d-8633-788ba1bf518c" (UID: "ba59e3a2-3d83-4f9d-8633-788ba1bf518c"). InnerVolumeSpecName "kube-api-access-mjbpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:28:05 crc kubenswrapper[4632]: I0313 12:28:05.010022 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mjbpg\" (UniqueName: \"kubernetes.io/projected/ba59e3a2-3d83-4f9d-8633-788ba1bf518c-kube-api-access-mjbpg\") on node \"crc\" DevicePath \"\"" Mar 13 12:28:05 crc kubenswrapper[4632]: I0313 12:28:05.044587 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:28:05 crc kubenswrapper[4632]: E0313 12:28:05.045102 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:28:05 crc kubenswrapper[4632]: I0313 12:28:05.343969 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556748-tr74t" event={"ID":"ba59e3a2-3d83-4f9d-8633-788ba1bf518c","Type":"ContainerDied","Data":"156d224a7e86839fcd1fe5b72dad72deacf75274b9199281fc604f2724238ed2"} Mar 13 12:28:05 crc kubenswrapper[4632]: I0313 12:28:05.344015 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="156d224a7e86839fcd1fe5b72dad72deacf75274b9199281fc604f2724238ed2" Mar 13 12:28:05 crc kubenswrapper[4632]: I0313 12:28:05.344081 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556748-tr74t" Mar 13 12:28:05 crc kubenswrapper[4632]: I0313 12:28:05.400138 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556742-2lkrj"] Mar 13 12:28:05 crc kubenswrapper[4632]: I0313 12:28:05.407806 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556742-2lkrj"] Mar 13 12:28:06 crc kubenswrapper[4632]: I0313 12:28:06.059967 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5121453-468a-432e-b110-fd0cd60ed92b" path="/var/lib/kubelet/pods/a5121453-468a-432e-b110-fd0cd60ed92b/volumes" Mar 13 12:28:20 crc kubenswrapper[4632]: I0313 12:28:20.044169 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:28:20 crc kubenswrapper[4632]: E0313 12:28:20.044841 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:28:26 crc kubenswrapper[4632]: I0313 12:28:26.919972 4632 scope.go:117] "RemoveContainer" containerID="77b11f376c487493e748aed75424d32e4d98e9395efe94071abc3a7b13ebc06d" Mar 13 12:28:34 crc kubenswrapper[4632]: I0313 12:28:34.044927 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:28:34 crc kubenswrapper[4632]: E0313 12:28:34.045630 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:28:47 crc kubenswrapper[4632]: I0313 12:28:47.044461 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:28:47 crc kubenswrapper[4632]: E0313 12:28:47.046129 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:29:00 crc kubenswrapper[4632]: I0313 12:29:00.044595 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:29:00 crc kubenswrapper[4632]: E0313 12:29:00.045512 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:29:12 crc kubenswrapper[4632]: I0313 12:29:12.045559 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:29:12 crc kubenswrapper[4632]: E0313 12:29:12.046827 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:29:27 crc kubenswrapper[4632]: I0313 12:29:27.044434 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:29:27 crc kubenswrapper[4632]: E0313 12:29:27.045418 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:29:38 crc kubenswrapper[4632]: I0313 12:29:38.059107 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:29:38 crc kubenswrapper[4632]: E0313 12:29:38.059985 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:29:51 crc kubenswrapper[4632]: I0313 12:29:51.045156 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:29:51 crc kubenswrapper[4632]: E0313 12:29:51.046091 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.163636 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556750-8fj49"] Mar 13 12:30:00 crc kubenswrapper[4632]: E0313 12:30:00.164593 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba59e3a2-3d83-4f9d-8633-788ba1bf518c" containerName="oc" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.164608 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba59e3a2-3d83-4f9d-8633-788ba1bf518c" containerName="oc" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.164804 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba59e3a2-3d83-4f9d-8633-788ba1bf518c" containerName="oc" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.165410 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556750-8fj49" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.171199 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.171293 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.171402 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.188847 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556750-8fj49"] Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.253763 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4"] Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.255884 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.258240 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.258495 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.274717 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v56sh\" (UniqueName: \"kubernetes.io/projected/8cc083b6-fb70-478e-9824-d9eb3cb1fe5b-kube-api-access-v56sh\") pod \"auto-csr-approver-29556750-8fj49\" (UID: \"8cc083b6-fb70-478e-9824-d9eb3cb1fe5b\") " pod="openshift-infra/auto-csr-approver-29556750-8fj49" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.277453 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4"] Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.376822 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7632e985-c049-4f40-b7e1-06337842cc06-secret-volume\") pod \"collect-profiles-29556750-66vw4\" (UID: \"7632e985-c049-4f40-b7e1-06337842cc06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.376894 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7632e985-c049-4f40-b7e1-06337842cc06-config-volume\") pod \"collect-profiles-29556750-66vw4\" (UID: \"7632e985-c049-4f40-b7e1-06337842cc06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.376927 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgq8z\" (UniqueName: \"kubernetes.io/projected/7632e985-c049-4f40-b7e1-06337842cc06-kube-api-access-jgq8z\") pod \"collect-profiles-29556750-66vw4\" (UID: \"7632e985-c049-4f40-b7e1-06337842cc06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.377357 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v56sh\" (UniqueName: \"kubernetes.io/projected/8cc083b6-fb70-478e-9824-d9eb3cb1fe5b-kube-api-access-v56sh\") pod \"auto-csr-approver-29556750-8fj49\" (UID: \"8cc083b6-fb70-478e-9824-d9eb3cb1fe5b\") " pod="openshift-infra/auto-csr-approver-29556750-8fj49" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.404706 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v56sh\" (UniqueName: \"kubernetes.io/projected/8cc083b6-fb70-478e-9824-d9eb3cb1fe5b-kube-api-access-v56sh\") pod \"auto-csr-approver-29556750-8fj49\" (UID: \"8cc083b6-fb70-478e-9824-d9eb3cb1fe5b\") " pod="openshift-infra/auto-csr-approver-29556750-8fj49" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.521562 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556750-8fj49" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.522481 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7632e985-c049-4f40-b7e1-06337842cc06-secret-volume\") pod \"collect-profiles-29556750-66vw4\" (UID: \"7632e985-c049-4f40-b7e1-06337842cc06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.522553 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7632e985-c049-4f40-b7e1-06337842cc06-config-volume\") pod \"collect-profiles-29556750-66vw4\" (UID: \"7632e985-c049-4f40-b7e1-06337842cc06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.522577 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgq8z\" (UniqueName: \"kubernetes.io/projected/7632e985-c049-4f40-b7e1-06337842cc06-kube-api-access-jgq8z\") pod \"collect-profiles-29556750-66vw4\" (UID: \"7632e985-c049-4f40-b7e1-06337842cc06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.523661 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7632e985-c049-4f40-b7e1-06337842cc06-config-volume\") pod \"collect-profiles-29556750-66vw4\" (UID: \"7632e985-c049-4f40-b7e1-06337842cc06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.528267 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7632e985-c049-4f40-b7e1-06337842cc06-secret-volume\") pod \"collect-profiles-29556750-66vw4\" (UID: \"7632e985-c049-4f40-b7e1-06337842cc06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.545567 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgq8z\" (UniqueName: \"kubernetes.io/projected/7632e985-c049-4f40-b7e1-06337842cc06-kube-api-access-jgq8z\") pod \"collect-profiles-29556750-66vw4\" (UID: \"7632e985-c049-4f40-b7e1-06337842cc06\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" Mar 13 12:30:00 crc kubenswrapper[4632]: I0313 12:30:00.577061 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" Mar 13 12:30:01 crc kubenswrapper[4632]: I0313 12:30:01.110047 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556750-8fj49"] Mar 13 12:30:01 crc kubenswrapper[4632]: I0313 12:30:01.205175 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4"] Mar 13 12:30:01 crc kubenswrapper[4632]: I0313 12:30:01.458175 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556750-8fj49" event={"ID":"8cc083b6-fb70-478e-9824-d9eb3cb1fe5b","Type":"ContainerStarted","Data":"fa176c143f5e8cdba415104680bdf7b7395fcbe38fc69dda69b086b4bfe88541"} Mar 13 12:30:01 crc kubenswrapper[4632]: I0313 12:30:01.460878 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" event={"ID":"7632e985-c049-4f40-b7e1-06337842cc06","Type":"ContainerStarted","Data":"67ca3bd2d41705600db916cd2979b0933bfd8e0736f1cba722fe7532e8a4c7c5"} Mar 13 12:30:01 crc kubenswrapper[4632]: I0313 12:30:01.460977 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" event={"ID":"7632e985-c049-4f40-b7e1-06337842cc06","Type":"ContainerStarted","Data":"f33f1afad4f51eb46c04f717f16a08b1bc00e6bd98ef5e322de77dafe10fd1f4"} Mar 13 12:30:02 crc kubenswrapper[4632]: I0313 12:30:02.474194 4632 generic.go:334] "Generic (PLEG): container finished" podID="7632e985-c049-4f40-b7e1-06337842cc06" containerID="67ca3bd2d41705600db916cd2979b0933bfd8e0736f1cba722fe7532e8a4c7c5" exitCode=0 Mar 13 12:30:02 crc kubenswrapper[4632]: I0313 12:30:02.474372 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" event={"ID":"7632e985-c049-4f40-b7e1-06337842cc06","Type":"ContainerDied","Data":"67ca3bd2d41705600db916cd2979b0933bfd8e0736f1cba722fe7532e8a4c7c5"} Mar 13 12:30:03 crc kubenswrapper[4632]: I0313 12:30:03.045423 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:30:03 crc kubenswrapper[4632]: E0313 12:30:03.046084 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:30:03 crc kubenswrapper[4632]: I0313 12:30:03.491661 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556750-8fj49" event={"ID":"8cc083b6-fb70-478e-9824-d9eb3cb1fe5b","Type":"ContainerStarted","Data":"50ec5eed6591caef46ce66e044fc885293f40a008476cefa9221d3ccb1262877"} Mar 13 12:30:03 crc kubenswrapper[4632]: I0313 12:30:03.520572 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556750-8fj49" podStartSLOduration=1.733201816 podStartE2EDuration="3.520549823s" podCreationTimestamp="2026-03-13 12:30:00 +0000 UTC" firstStartedPulling="2026-03-13 12:30:01.120093869 +0000 UTC m=+8775.142624002" lastFinishedPulling="2026-03-13 12:30:02.907441876 +0000 UTC m=+8776.929972009" observedRunningTime="2026-03-13 12:30:03.512779522 +0000 UTC m=+8777.535309665" watchObservedRunningTime="2026-03-13 12:30:03.520549823 +0000 UTC m=+8777.543079966" Mar 13 12:30:03 crc kubenswrapper[4632]: I0313 12:30:03.882412 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" Mar 13 12:30:04 crc kubenswrapper[4632]: I0313 12:30:04.048051 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7632e985-c049-4f40-b7e1-06337842cc06-secret-volume\") pod \"7632e985-c049-4f40-b7e1-06337842cc06\" (UID: \"7632e985-c049-4f40-b7e1-06337842cc06\") " Mar 13 12:30:04 crc kubenswrapper[4632]: I0313 12:30:04.048292 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7632e985-c049-4f40-b7e1-06337842cc06-config-volume\") pod \"7632e985-c049-4f40-b7e1-06337842cc06\" (UID: \"7632e985-c049-4f40-b7e1-06337842cc06\") " Mar 13 12:30:04 crc kubenswrapper[4632]: I0313 12:30:04.048746 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7632e985-c049-4f40-b7e1-06337842cc06-config-volume" (OuterVolumeSpecName: "config-volume") pod "7632e985-c049-4f40-b7e1-06337842cc06" (UID: "7632e985-c049-4f40-b7e1-06337842cc06"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 12:30:04 crc kubenswrapper[4632]: I0313 12:30:04.049112 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgq8z\" (UniqueName: \"kubernetes.io/projected/7632e985-c049-4f40-b7e1-06337842cc06-kube-api-access-jgq8z\") pod \"7632e985-c049-4f40-b7e1-06337842cc06\" (UID: \"7632e985-c049-4f40-b7e1-06337842cc06\") " Mar 13 12:30:04 crc kubenswrapper[4632]: I0313 12:30:04.049727 4632 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/7632e985-c049-4f40-b7e1-06337842cc06-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 12:30:04 crc kubenswrapper[4632]: I0313 12:30:04.055010 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7632e985-c049-4f40-b7e1-06337842cc06-kube-api-access-jgq8z" (OuterVolumeSpecName: "kube-api-access-jgq8z") pod "7632e985-c049-4f40-b7e1-06337842cc06" (UID: "7632e985-c049-4f40-b7e1-06337842cc06"). InnerVolumeSpecName "kube-api-access-jgq8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:30:04 crc kubenswrapper[4632]: I0313 12:30:04.061084 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7632e985-c049-4f40-b7e1-06337842cc06-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "7632e985-c049-4f40-b7e1-06337842cc06" (UID: "7632e985-c049-4f40-b7e1-06337842cc06"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:30:04 crc kubenswrapper[4632]: I0313 12:30:04.151163 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgq8z\" (UniqueName: \"kubernetes.io/projected/7632e985-c049-4f40-b7e1-06337842cc06-kube-api-access-jgq8z\") on node \"crc\" DevicePath \"\"" Mar 13 12:30:04 crc kubenswrapper[4632]: I0313 12:30:04.151206 4632 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/7632e985-c049-4f40-b7e1-06337842cc06-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 12:30:04 crc kubenswrapper[4632]: I0313 12:30:04.504519 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" Mar 13 12:30:04 crc kubenswrapper[4632]: I0313 12:30:04.505019 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556750-66vw4" event={"ID":"7632e985-c049-4f40-b7e1-06337842cc06","Type":"ContainerDied","Data":"f33f1afad4f51eb46c04f717f16a08b1bc00e6bd98ef5e322de77dafe10fd1f4"} Mar 13 12:30:04 crc kubenswrapper[4632]: I0313 12:30:04.505058 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f33f1afad4f51eb46c04f717f16a08b1bc00e6bd98ef5e322de77dafe10fd1f4" Mar 13 12:30:04 crc kubenswrapper[4632]: I0313 12:30:04.578464 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv"] Mar 13 12:30:04 crc kubenswrapper[4632]: I0313 12:30:04.592101 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556705-cdjnv"] Mar 13 12:30:05 crc kubenswrapper[4632]: I0313 12:30:05.511767 4632 generic.go:334] "Generic (PLEG): container finished" podID="8cc083b6-fb70-478e-9824-d9eb3cb1fe5b" containerID="50ec5eed6591caef46ce66e044fc885293f40a008476cefa9221d3ccb1262877" exitCode=0 Mar 13 12:30:05 crc kubenswrapper[4632]: I0313 12:30:05.511809 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556750-8fj49" event={"ID":"8cc083b6-fb70-478e-9824-d9eb3cb1fe5b","Type":"ContainerDied","Data":"50ec5eed6591caef46ce66e044fc885293f40a008476cefa9221d3ccb1262877"} Mar 13 12:30:06 crc kubenswrapper[4632]: I0313 12:30:06.070501 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7" path="/var/lib/kubelet/pods/964e7b0f-4dfa-43e3-9ed5-a9c176c8cfc7/volumes" Mar 13 12:30:06 crc kubenswrapper[4632]: I0313 12:30:06.935031 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556750-8fj49" Mar 13 12:30:07 crc kubenswrapper[4632]: I0313 12:30:07.008925 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v56sh\" (UniqueName: \"kubernetes.io/projected/8cc083b6-fb70-478e-9824-d9eb3cb1fe5b-kube-api-access-v56sh\") pod \"8cc083b6-fb70-478e-9824-d9eb3cb1fe5b\" (UID: \"8cc083b6-fb70-478e-9824-d9eb3cb1fe5b\") " Mar 13 12:30:07 crc kubenswrapper[4632]: I0313 12:30:07.015724 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cc083b6-fb70-478e-9824-d9eb3cb1fe5b-kube-api-access-v56sh" (OuterVolumeSpecName: "kube-api-access-v56sh") pod "8cc083b6-fb70-478e-9824-d9eb3cb1fe5b" (UID: "8cc083b6-fb70-478e-9824-d9eb3cb1fe5b"). InnerVolumeSpecName "kube-api-access-v56sh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:30:07 crc kubenswrapper[4632]: I0313 12:30:07.111656 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v56sh\" (UniqueName: \"kubernetes.io/projected/8cc083b6-fb70-478e-9824-d9eb3cb1fe5b-kube-api-access-v56sh\") on node \"crc\" DevicePath \"\"" Mar 13 12:30:07 crc kubenswrapper[4632]: I0313 12:30:07.533850 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556750-8fj49" event={"ID":"8cc083b6-fb70-478e-9824-d9eb3cb1fe5b","Type":"ContainerDied","Data":"fa176c143f5e8cdba415104680bdf7b7395fcbe38fc69dda69b086b4bfe88541"} Mar 13 12:30:07 crc kubenswrapper[4632]: I0313 12:30:07.534148 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556750-8fj49" Mar 13 12:30:07 crc kubenswrapper[4632]: I0313 12:30:07.534154 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa176c143f5e8cdba415104680bdf7b7395fcbe38fc69dda69b086b4bfe88541" Mar 13 12:30:07 crc kubenswrapper[4632]: I0313 12:30:07.603824 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556744-fb92p"] Mar 13 12:30:07 crc kubenswrapper[4632]: I0313 12:30:07.665281 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556744-fb92p"] Mar 13 12:30:08 crc kubenswrapper[4632]: I0313 12:30:08.058269 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1006ed9-194b-4d1b-91cf-7722ce335023" path="/var/lib/kubelet/pods/e1006ed9-194b-4d1b-91cf-7722ce335023/volumes" Mar 13 12:30:17 crc kubenswrapper[4632]: I0313 12:30:17.045163 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:30:17 crc kubenswrapper[4632]: E0313 12:30:17.046237 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:30:27 crc kubenswrapper[4632]: I0313 12:30:27.037342 4632 scope.go:117] "RemoveContainer" containerID="8c147fe4b276fdf885236433df734b60b43a6fbbd4e1c4d2a7bec9fd5c3cc6e2" Mar 13 12:30:27 crc kubenswrapper[4632]: I0313 12:30:27.162952 4632 scope.go:117] "RemoveContainer" containerID="a50d24de30277dacbb16bc71e07335e3c84d2cedb12dfb6c3d660775ff2f0c54" Mar 13 12:30:28 crc kubenswrapper[4632]: I0313 12:30:28.046997 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:30:28 crc kubenswrapper[4632]: E0313 12:30:28.047667 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:30:42 crc kubenswrapper[4632]: I0313 12:30:42.045062 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:30:42 crc kubenswrapper[4632]: E0313 12:30:42.045901 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:30:54 crc kubenswrapper[4632]: I0313 12:30:54.044335 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:30:54 crc kubenswrapper[4632]: E0313 12:30:54.046330 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:31:07 crc kubenswrapper[4632]: I0313 12:31:07.044200 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:31:07 crc kubenswrapper[4632]: E0313 12:31:07.045077 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:31:18 crc kubenswrapper[4632]: I0313 12:31:18.053672 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:31:18 crc kubenswrapper[4632]: I0313 12:31:18.396536 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"ada2bd3447f81dbcb3c7c10ab1a84d7a61b81476a09d5bccd655ef21929539af"} Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.158964 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g9xk2"] Mar 13 12:31:25 crc kubenswrapper[4632]: E0313 12:31:25.160332 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cc083b6-fb70-478e-9824-d9eb3cb1fe5b" containerName="oc" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.160352 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cc083b6-fb70-478e-9824-d9eb3cb1fe5b" containerName="oc" Mar 13 12:31:25 crc kubenswrapper[4632]: E0313 12:31:25.160366 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7632e985-c049-4f40-b7e1-06337842cc06" containerName="collect-profiles" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.160374 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7632e985-c049-4f40-b7e1-06337842cc06" containerName="collect-profiles" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.160575 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="7632e985-c049-4f40-b7e1-06337842cc06" containerName="collect-profiles" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.160599 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cc083b6-fb70-478e-9824-d9eb3cb1fe5b" containerName="oc" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.162474 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.231186 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g9xk2"] Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.250526 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/890fbbb1-da06-4cd8-80ad-3606cf60429c-utilities\") pod \"community-operators-g9xk2\" (UID: \"890fbbb1-da06-4cd8-80ad-3606cf60429c\") " pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.250625 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/890fbbb1-da06-4cd8-80ad-3606cf60429c-catalog-content\") pod \"community-operators-g9xk2\" (UID: \"890fbbb1-da06-4cd8-80ad-3606cf60429c\") " pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.250661 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6qsw\" (UniqueName: \"kubernetes.io/projected/890fbbb1-da06-4cd8-80ad-3606cf60429c-kube-api-access-b6qsw\") pod \"community-operators-g9xk2\" (UID: \"890fbbb1-da06-4cd8-80ad-3606cf60429c\") " pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.352304 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/890fbbb1-da06-4cd8-80ad-3606cf60429c-utilities\") pod \"community-operators-g9xk2\" (UID: \"890fbbb1-da06-4cd8-80ad-3606cf60429c\") " pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.352363 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/890fbbb1-da06-4cd8-80ad-3606cf60429c-catalog-content\") pod \"community-operators-g9xk2\" (UID: \"890fbbb1-da06-4cd8-80ad-3606cf60429c\") " pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.352392 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6qsw\" (UniqueName: \"kubernetes.io/projected/890fbbb1-da06-4cd8-80ad-3606cf60429c-kube-api-access-b6qsw\") pod \"community-operators-g9xk2\" (UID: \"890fbbb1-da06-4cd8-80ad-3606cf60429c\") " pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.353560 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/890fbbb1-da06-4cd8-80ad-3606cf60429c-utilities\") pod \"community-operators-g9xk2\" (UID: \"890fbbb1-da06-4cd8-80ad-3606cf60429c\") " pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.353575 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/890fbbb1-da06-4cd8-80ad-3606cf60429c-catalog-content\") pod \"community-operators-g9xk2\" (UID: \"890fbbb1-da06-4cd8-80ad-3606cf60429c\") " pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.379869 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6qsw\" (UniqueName: \"kubernetes.io/projected/890fbbb1-da06-4cd8-80ad-3606cf60429c-kube-api-access-b6qsw\") pod \"community-operators-g9xk2\" (UID: \"890fbbb1-da06-4cd8-80ad-3606cf60429c\") " pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:25 crc kubenswrapper[4632]: I0313 12:31:25.484384 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:26 crc kubenswrapper[4632]: I0313 12:31:26.615724 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g9xk2"] Mar 13 12:31:26 crc kubenswrapper[4632]: W0313 12:31:26.637725 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod890fbbb1_da06_4cd8_80ad_3606cf60429c.slice/crio-91406894900e25486440e9ccbbbcf37ffdbcd032fb6ed3c80c9a9df23adbf89d WatchSource:0}: Error finding container 91406894900e25486440e9ccbbbcf37ffdbcd032fb6ed3c80c9a9df23adbf89d: Status 404 returned error can't find the container with id 91406894900e25486440e9ccbbbcf37ffdbcd032fb6ed3c80c9a9df23adbf89d Mar 13 12:31:27 crc kubenswrapper[4632]: I0313 12:31:27.482928 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9xk2" event={"ID":"890fbbb1-da06-4cd8-80ad-3606cf60429c","Type":"ContainerDied","Data":"41ca3fc9586c0ae46fcfec562f2e7230a5fd9855fcb4904bbbb958362880638c"} Mar 13 12:31:27 crc kubenswrapper[4632]: I0313 12:31:27.483149 4632 generic.go:334] "Generic (PLEG): container finished" podID="890fbbb1-da06-4cd8-80ad-3606cf60429c" containerID="41ca3fc9586c0ae46fcfec562f2e7230a5fd9855fcb4904bbbb958362880638c" exitCode=0 Mar 13 12:31:27 crc kubenswrapper[4632]: I0313 12:31:27.483760 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9xk2" event={"ID":"890fbbb1-da06-4cd8-80ad-3606cf60429c","Type":"ContainerStarted","Data":"91406894900e25486440e9ccbbbcf37ffdbcd032fb6ed3c80c9a9df23adbf89d"} Mar 13 12:31:27 crc kubenswrapper[4632]: I0313 12:31:27.489024 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 12:31:28 crc kubenswrapper[4632]: I0313 12:31:28.494306 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9xk2" event={"ID":"890fbbb1-da06-4cd8-80ad-3606cf60429c","Type":"ContainerStarted","Data":"0aa7a5a74d17454fafe7a3ea9a1ca0a93f0db659d516166286bcd7d7d51f347f"} Mar 13 12:31:31 crc kubenswrapper[4632]: I0313 12:31:31.533263 4632 generic.go:334] "Generic (PLEG): container finished" podID="890fbbb1-da06-4cd8-80ad-3606cf60429c" containerID="0aa7a5a74d17454fafe7a3ea9a1ca0a93f0db659d516166286bcd7d7d51f347f" exitCode=0 Mar 13 12:31:31 crc kubenswrapper[4632]: I0313 12:31:31.533329 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9xk2" event={"ID":"890fbbb1-da06-4cd8-80ad-3606cf60429c","Type":"ContainerDied","Data":"0aa7a5a74d17454fafe7a3ea9a1ca0a93f0db659d516166286bcd7d7d51f347f"} Mar 13 12:31:32 crc kubenswrapper[4632]: I0313 12:31:32.546441 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9xk2" event={"ID":"890fbbb1-da06-4cd8-80ad-3606cf60429c","Type":"ContainerStarted","Data":"8f38e62b583cfd0853cbd527e55c37912c0be15194f7d16e790f2fedb8e0673e"} Mar 13 12:31:32 crc kubenswrapper[4632]: I0313 12:31:32.571908 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g9xk2" podStartSLOduration=3.08852628 podStartE2EDuration="7.568733761s" podCreationTimestamp="2026-03-13 12:31:25 +0000 UTC" firstStartedPulling="2026-03-13 12:31:27.485824049 +0000 UTC m=+8861.508354192" lastFinishedPulling="2026-03-13 12:31:31.96603154 +0000 UTC m=+8865.988561673" observedRunningTime="2026-03-13 12:31:32.566321762 +0000 UTC m=+8866.588851925" watchObservedRunningTime="2026-03-13 12:31:32.568733761 +0000 UTC m=+8866.591263914" Mar 13 12:31:35 crc kubenswrapper[4632]: I0313 12:31:35.486075 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:35 crc kubenswrapper[4632]: I0313 12:31:35.486423 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:36 crc kubenswrapper[4632]: I0313 12:31:36.542502 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-g9xk2" podUID="890fbbb1-da06-4cd8-80ad-3606cf60429c" containerName="registry-server" probeResult="failure" output=< Mar 13 12:31:36 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:31:36 crc kubenswrapper[4632]: > Mar 13 12:31:45 crc kubenswrapper[4632]: I0313 12:31:45.563825 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:45 crc kubenswrapper[4632]: I0313 12:31:45.635117 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:45 crc kubenswrapper[4632]: I0313 12:31:45.812265 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g9xk2"] Mar 13 12:31:46 crc kubenswrapper[4632]: I0313 12:31:46.686060 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g9xk2" podUID="890fbbb1-da06-4cd8-80ad-3606cf60429c" containerName="registry-server" containerID="cri-o://8f38e62b583cfd0853cbd527e55c37912c0be15194f7d16e790f2fedb8e0673e" gracePeriod=2 Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.534650 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.595468 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6qsw\" (UniqueName: \"kubernetes.io/projected/890fbbb1-da06-4cd8-80ad-3606cf60429c-kube-api-access-b6qsw\") pod \"890fbbb1-da06-4cd8-80ad-3606cf60429c\" (UID: \"890fbbb1-da06-4cd8-80ad-3606cf60429c\") " Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.595746 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/890fbbb1-da06-4cd8-80ad-3606cf60429c-utilities\") pod \"890fbbb1-da06-4cd8-80ad-3606cf60429c\" (UID: \"890fbbb1-da06-4cd8-80ad-3606cf60429c\") " Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.595796 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/890fbbb1-da06-4cd8-80ad-3606cf60429c-catalog-content\") pod \"890fbbb1-da06-4cd8-80ad-3606cf60429c\" (UID: \"890fbbb1-da06-4cd8-80ad-3606cf60429c\") " Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.597656 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/890fbbb1-da06-4cd8-80ad-3606cf60429c-utilities" (OuterVolumeSpecName: "utilities") pod "890fbbb1-da06-4cd8-80ad-3606cf60429c" (UID: "890fbbb1-da06-4cd8-80ad-3606cf60429c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.616257 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/890fbbb1-da06-4cd8-80ad-3606cf60429c-kube-api-access-b6qsw" (OuterVolumeSpecName: "kube-api-access-b6qsw") pod "890fbbb1-da06-4cd8-80ad-3606cf60429c" (UID: "890fbbb1-da06-4cd8-80ad-3606cf60429c"). InnerVolumeSpecName "kube-api-access-b6qsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.663325 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/890fbbb1-da06-4cd8-80ad-3606cf60429c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "890fbbb1-da06-4cd8-80ad-3606cf60429c" (UID: "890fbbb1-da06-4cd8-80ad-3606cf60429c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.698710 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6qsw\" (UniqueName: \"kubernetes.io/projected/890fbbb1-da06-4cd8-80ad-3606cf60429c-kube-api-access-b6qsw\") on node \"crc\" DevicePath \"\"" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.698755 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/890fbbb1-da06-4cd8-80ad-3606cf60429c-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.698768 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/890fbbb1-da06-4cd8-80ad-3606cf60429c-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.702381 4632 generic.go:334] "Generic (PLEG): container finished" podID="890fbbb1-da06-4cd8-80ad-3606cf60429c" containerID="8f38e62b583cfd0853cbd527e55c37912c0be15194f7d16e790f2fedb8e0673e" exitCode=0 Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.702430 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g9xk2" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.702436 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9xk2" event={"ID":"890fbbb1-da06-4cd8-80ad-3606cf60429c","Type":"ContainerDied","Data":"8f38e62b583cfd0853cbd527e55c37912c0be15194f7d16e790f2fedb8e0673e"} Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.702473 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g9xk2" event={"ID":"890fbbb1-da06-4cd8-80ad-3606cf60429c","Type":"ContainerDied","Data":"91406894900e25486440e9ccbbbcf37ffdbcd032fb6ed3c80c9a9df23adbf89d"} Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.702495 4632 scope.go:117] "RemoveContainer" containerID="8f38e62b583cfd0853cbd527e55c37912c0be15194f7d16e790f2fedb8e0673e" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.734885 4632 scope.go:117] "RemoveContainer" containerID="0aa7a5a74d17454fafe7a3ea9a1ca0a93f0db659d516166286bcd7d7d51f347f" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.756269 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g9xk2"] Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.759132 4632 scope.go:117] "RemoveContainer" containerID="41ca3fc9586c0ae46fcfec562f2e7230a5fd9855fcb4904bbbb958362880638c" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.765074 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g9xk2"] Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.813632 4632 scope.go:117] "RemoveContainer" containerID="8f38e62b583cfd0853cbd527e55c37912c0be15194f7d16e790f2fedb8e0673e" Mar 13 12:31:47 crc kubenswrapper[4632]: E0313 12:31:47.816920 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f38e62b583cfd0853cbd527e55c37912c0be15194f7d16e790f2fedb8e0673e\": container with ID starting with 8f38e62b583cfd0853cbd527e55c37912c0be15194f7d16e790f2fedb8e0673e not found: ID does not exist" containerID="8f38e62b583cfd0853cbd527e55c37912c0be15194f7d16e790f2fedb8e0673e" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.817968 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f38e62b583cfd0853cbd527e55c37912c0be15194f7d16e790f2fedb8e0673e"} err="failed to get container status \"8f38e62b583cfd0853cbd527e55c37912c0be15194f7d16e790f2fedb8e0673e\": rpc error: code = NotFound desc = could not find container \"8f38e62b583cfd0853cbd527e55c37912c0be15194f7d16e790f2fedb8e0673e\": container with ID starting with 8f38e62b583cfd0853cbd527e55c37912c0be15194f7d16e790f2fedb8e0673e not found: ID does not exist" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.818007 4632 scope.go:117] "RemoveContainer" containerID="0aa7a5a74d17454fafe7a3ea9a1ca0a93f0db659d516166286bcd7d7d51f347f" Mar 13 12:31:47 crc kubenswrapper[4632]: E0313 12:31:47.818467 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0aa7a5a74d17454fafe7a3ea9a1ca0a93f0db659d516166286bcd7d7d51f347f\": container with ID starting with 0aa7a5a74d17454fafe7a3ea9a1ca0a93f0db659d516166286bcd7d7d51f347f not found: ID does not exist" containerID="0aa7a5a74d17454fafe7a3ea9a1ca0a93f0db659d516166286bcd7d7d51f347f" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.818507 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0aa7a5a74d17454fafe7a3ea9a1ca0a93f0db659d516166286bcd7d7d51f347f"} err="failed to get container status \"0aa7a5a74d17454fafe7a3ea9a1ca0a93f0db659d516166286bcd7d7d51f347f\": rpc error: code = NotFound desc = could not find container \"0aa7a5a74d17454fafe7a3ea9a1ca0a93f0db659d516166286bcd7d7d51f347f\": container with ID starting with 0aa7a5a74d17454fafe7a3ea9a1ca0a93f0db659d516166286bcd7d7d51f347f not found: ID does not exist" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.819120 4632 scope.go:117] "RemoveContainer" containerID="41ca3fc9586c0ae46fcfec562f2e7230a5fd9855fcb4904bbbb958362880638c" Mar 13 12:31:47 crc kubenswrapper[4632]: E0313 12:31:47.819479 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41ca3fc9586c0ae46fcfec562f2e7230a5fd9855fcb4904bbbb958362880638c\": container with ID starting with 41ca3fc9586c0ae46fcfec562f2e7230a5fd9855fcb4904bbbb958362880638c not found: ID does not exist" containerID="41ca3fc9586c0ae46fcfec562f2e7230a5fd9855fcb4904bbbb958362880638c" Mar 13 12:31:47 crc kubenswrapper[4632]: I0313 12:31:47.819528 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41ca3fc9586c0ae46fcfec562f2e7230a5fd9855fcb4904bbbb958362880638c"} err="failed to get container status \"41ca3fc9586c0ae46fcfec562f2e7230a5fd9855fcb4904bbbb958362880638c\": rpc error: code = NotFound desc = could not find container \"41ca3fc9586c0ae46fcfec562f2e7230a5fd9855fcb4904bbbb958362880638c\": container with ID starting with 41ca3fc9586c0ae46fcfec562f2e7230a5fd9855fcb4904bbbb958362880638c not found: ID does not exist" Mar 13 12:31:48 crc kubenswrapper[4632]: I0313 12:31:48.055788 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="890fbbb1-da06-4cd8-80ad-3606cf60429c" path="/var/lib/kubelet/pods/890fbbb1-da06-4cd8-80ad-3606cf60429c/volumes" Mar 13 12:32:00 crc kubenswrapper[4632]: I0313 12:32:00.201233 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556752-48xgt"] Mar 13 12:32:00 crc kubenswrapper[4632]: E0313 12:32:00.202198 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="890fbbb1-da06-4cd8-80ad-3606cf60429c" containerName="extract-utilities" Mar 13 12:32:00 crc kubenswrapper[4632]: I0313 12:32:00.202214 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="890fbbb1-da06-4cd8-80ad-3606cf60429c" containerName="extract-utilities" Mar 13 12:32:00 crc kubenswrapper[4632]: E0313 12:32:00.202226 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="890fbbb1-da06-4cd8-80ad-3606cf60429c" containerName="registry-server" Mar 13 12:32:00 crc kubenswrapper[4632]: I0313 12:32:00.202233 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="890fbbb1-da06-4cd8-80ad-3606cf60429c" containerName="registry-server" Mar 13 12:32:00 crc kubenswrapper[4632]: E0313 12:32:00.202248 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="890fbbb1-da06-4cd8-80ad-3606cf60429c" containerName="extract-content" Mar 13 12:32:00 crc kubenswrapper[4632]: I0313 12:32:00.202254 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="890fbbb1-da06-4cd8-80ad-3606cf60429c" containerName="extract-content" Mar 13 12:32:00 crc kubenswrapper[4632]: I0313 12:32:00.202464 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="890fbbb1-da06-4cd8-80ad-3606cf60429c" containerName="registry-server" Mar 13 12:32:00 crc kubenswrapper[4632]: I0313 12:32:00.204671 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556752-48xgt" Mar 13 12:32:00 crc kubenswrapper[4632]: I0313 12:32:00.214306 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556752-48xgt"] Mar 13 12:32:00 crc kubenswrapper[4632]: I0313 12:32:00.219693 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:32:00 crc kubenswrapper[4632]: I0313 12:32:00.247462 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:32:00 crc kubenswrapper[4632]: I0313 12:32:00.248268 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:32:00 crc kubenswrapper[4632]: I0313 12:32:00.258308 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmpl8\" (UniqueName: \"kubernetes.io/projected/c311ec54-27ae-4082-bd23-4df180976b2f-kube-api-access-rmpl8\") pod \"auto-csr-approver-29556752-48xgt\" (UID: \"c311ec54-27ae-4082-bd23-4df180976b2f\") " pod="openshift-infra/auto-csr-approver-29556752-48xgt" Mar 13 12:32:00 crc kubenswrapper[4632]: I0313 12:32:00.360035 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmpl8\" (UniqueName: \"kubernetes.io/projected/c311ec54-27ae-4082-bd23-4df180976b2f-kube-api-access-rmpl8\") pod \"auto-csr-approver-29556752-48xgt\" (UID: \"c311ec54-27ae-4082-bd23-4df180976b2f\") " pod="openshift-infra/auto-csr-approver-29556752-48xgt" Mar 13 12:32:00 crc kubenswrapper[4632]: I0313 12:32:00.380594 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmpl8\" (UniqueName: \"kubernetes.io/projected/c311ec54-27ae-4082-bd23-4df180976b2f-kube-api-access-rmpl8\") pod \"auto-csr-approver-29556752-48xgt\" (UID: \"c311ec54-27ae-4082-bd23-4df180976b2f\") " pod="openshift-infra/auto-csr-approver-29556752-48xgt" Mar 13 12:32:00 crc kubenswrapper[4632]: I0313 12:32:00.558563 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556752-48xgt" Mar 13 12:32:01 crc kubenswrapper[4632]: I0313 12:32:01.154488 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556752-48xgt"] Mar 13 12:32:01 crc kubenswrapper[4632]: I0313 12:32:01.841111 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556752-48xgt" event={"ID":"c311ec54-27ae-4082-bd23-4df180976b2f","Type":"ContainerStarted","Data":"fea7a8a20327e74d886dc273e5419e151a924579f2951bf3a57f9b9c8d553f18"} Mar 13 12:32:03 crc kubenswrapper[4632]: I0313 12:32:03.861356 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556752-48xgt" event={"ID":"c311ec54-27ae-4082-bd23-4df180976b2f","Type":"ContainerStarted","Data":"997fb6aac287ee23705f38733ad6b8cf02cea468d3978ae35a30d75ea0dfec0f"} Mar 13 12:32:03 crc kubenswrapper[4632]: I0313 12:32:03.882557 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556752-48xgt" podStartSLOduration=3.057365928 podStartE2EDuration="3.882533705s" podCreationTimestamp="2026-03-13 12:32:00 +0000 UTC" firstStartedPulling="2026-03-13 12:32:01.174693814 +0000 UTC m=+8895.197223947" lastFinishedPulling="2026-03-13 12:32:01.999861591 +0000 UTC m=+8896.022391724" observedRunningTime="2026-03-13 12:32:03.877582034 +0000 UTC m=+8897.900112167" watchObservedRunningTime="2026-03-13 12:32:03.882533705 +0000 UTC m=+8897.905063858" Mar 13 12:32:04 crc kubenswrapper[4632]: I0313 12:32:04.878363 4632 generic.go:334] "Generic (PLEG): container finished" podID="c311ec54-27ae-4082-bd23-4df180976b2f" containerID="997fb6aac287ee23705f38733ad6b8cf02cea468d3978ae35a30d75ea0dfec0f" exitCode=0 Mar 13 12:32:04 crc kubenswrapper[4632]: I0313 12:32:04.878442 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556752-48xgt" event={"ID":"c311ec54-27ae-4082-bd23-4df180976b2f","Type":"ContainerDied","Data":"997fb6aac287ee23705f38733ad6b8cf02cea468d3978ae35a30d75ea0dfec0f"} Mar 13 12:32:06 crc kubenswrapper[4632]: I0313 12:32:06.303958 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556752-48xgt" Mar 13 12:32:06 crc kubenswrapper[4632]: I0313 12:32:06.383451 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmpl8\" (UniqueName: \"kubernetes.io/projected/c311ec54-27ae-4082-bd23-4df180976b2f-kube-api-access-rmpl8\") pod \"c311ec54-27ae-4082-bd23-4df180976b2f\" (UID: \"c311ec54-27ae-4082-bd23-4df180976b2f\") " Mar 13 12:32:06 crc kubenswrapper[4632]: I0313 12:32:06.399404 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c311ec54-27ae-4082-bd23-4df180976b2f-kube-api-access-rmpl8" (OuterVolumeSpecName: "kube-api-access-rmpl8") pod "c311ec54-27ae-4082-bd23-4df180976b2f" (UID: "c311ec54-27ae-4082-bd23-4df180976b2f"). InnerVolumeSpecName "kube-api-access-rmpl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:32:06 crc kubenswrapper[4632]: I0313 12:32:06.487562 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmpl8\" (UniqueName: \"kubernetes.io/projected/c311ec54-27ae-4082-bd23-4df180976b2f-kube-api-access-rmpl8\") on node \"crc\" DevicePath \"\"" Mar 13 12:32:06 crc kubenswrapper[4632]: I0313 12:32:06.899006 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556752-48xgt" event={"ID":"c311ec54-27ae-4082-bd23-4df180976b2f","Type":"ContainerDied","Data":"fea7a8a20327e74d886dc273e5419e151a924579f2951bf3a57f9b9c8d553f18"} Mar 13 12:32:06 crc kubenswrapper[4632]: I0313 12:32:06.899053 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fea7a8a20327e74d886dc273e5419e151a924579f2951bf3a57f9b9c8d553f18" Mar 13 12:32:06 crc kubenswrapper[4632]: I0313 12:32:06.899067 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556752-48xgt" Mar 13 12:32:06 crc kubenswrapper[4632]: I0313 12:32:06.959349 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556746-jtv7q"] Mar 13 12:32:06 crc kubenswrapper[4632]: I0313 12:32:06.968284 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556746-jtv7q"] Mar 13 12:32:08 crc kubenswrapper[4632]: I0313 12:32:08.066743 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d88882b5-a11f-4606-8e0d-59471c7feccb" path="/var/lib/kubelet/pods/d88882b5-a11f-4606-8e0d-59471c7feccb/volumes" Mar 13 12:32:27 crc kubenswrapper[4632]: I0313 12:32:27.348814 4632 scope.go:117] "RemoveContainer" containerID="35ea86f2c8d1955a868a4d03ab725c4feb194878cc720326b8eb0c50ed5ce3c5" Mar 13 12:33:40 crc kubenswrapper[4632]: I0313 12:33:40.461336 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:33:40 crc kubenswrapper[4632]: I0313 12:33:40.461960 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:34:00 crc kubenswrapper[4632]: I0313 12:34:00.155201 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556754-kcjmj"] Mar 13 12:34:00 crc kubenswrapper[4632]: E0313 12:34:00.156479 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c311ec54-27ae-4082-bd23-4df180976b2f" containerName="oc" Mar 13 12:34:00 crc kubenswrapper[4632]: I0313 12:34:00.156504 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="c311ec54-27ae-4082-bd23-4df180976b2f" containerName="oc" Mar 13 12:34:00 crc kubenswrapper[4632]: I0313 12:34:00.156826 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="c311ec54-27ae-4082-bd23-4df180976b2f" containerName="oc" Mar 13 12:34:00 crc kubenswrapper[4632]: I0313 12:34:00.157678 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556754-kcjmj" Mar 13 12:34:00 crc kubenswrapper[4632]: I0313 12:34:00.165600 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:34:00 crc kubenswrapper[4632]: I0313 12:34:00.165935 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:34:00 crc kubenswrapper[4632]: I0313 12:34:00.169877 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:34:00 crc kubenswrapper[4632]: I0313 12:34:00.174592 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556754-kcjmj"] Mar 13 12:34:00 crc kubenswrapper[4632]: I0313 12:34:00.262632 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2v6b\" (UniqueName: \"kubernetes.io/projected/ccc70d73-b58b-4a2c-9bce-dc27405c5710-kube-api-access-r2v6b\") pod \"auto-csr-approver-29556754-kcjmj\" (UID: \"ccc70d73-b58b-4a2c-9bce-dc27405c5710\") " pod="openshift-infra/auto-csr-approver-29556754-kcjmj" Mar 13 12:34:00 crc kubenswrapper[4632]: I0313 12:34:00.365201 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2v6b\" (UniqueName: \"kubernetes.io/projected/ccc70d73-b58b-4a2c-9bce-dc27405c5710-kube-api-access-r2v6b\") pod \"auto-csr-approver-29556754-kcjmj\" (UID: \"ccc70d73-b58b-4a2c-9bce-dc27405c5710\") " pod="openshift-infra/auto-csr-approver-29556754-kcjmj" Mar 13 12:34:00 crc kubenswrapper[4632]: I0313 12:34:00.394548 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2v6b\" (UniqueName: \"kubernetes.io/projected/ccc70d73-b58b-4a2c-9bce-dc27405c5710-kube-api-access-r2v6b\") pod \"auto-csr-approver-29556754-kcjmj\" (UID: \"ccc70d73-b58b-4a2c-9bce-dc27405c5710\") " pod="openshift-infra/auto-csr-approver-29556754-kcjmj" Mar 13 12:34:00 crc kubenswrapper[4632]: I0313 12:34:00.481015 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556754-kcjmj" Mar 13 12:34:01 crc kubenswrapper[4632]: I0313 12:34:01.165454 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556754-kcjmj"] Mar 13 12:34:01 crc kubenswrapper[4632]: I0313 12:34:01.442726 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556754-kcjmj" event={"ID":"ccc70d73-b58b-4a2c-9bce-dc27405c5710","Type":"ContainerStarted","Data":"8c18300927c4a79068fc1d8cdd84afa4a73b1ab68a0bfed9583785fb44288336"} Mar 13 12:34:03 crc kubenswrapper[4632]: I0313 12:34:03.463087 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556754-kcjmj" event={"ID":"ccc70d73-b58b-4a2c-9bce-dc27405c5710","Type":"ContainerStarted","Data":"33c9be3390a29151e585ddbf79d6ef390b1a094d7878d4d4c96b9b0bb39d369c"} Mar 13 12:34:03 crc kubenswrapper[4632]: I0313 12:34:03.486256 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556754-kcjmj" podStartSLOduration=2.397711432 podStartE2EDuration="3.486232128s" podCreationTimestamp="2026-03-13 12:34:00 +0000 UTC" firstStartedPulling="2026-03-13 12:34:01.178238624 +0000 UTC m=+9015.200768747" lastFinishedPulling="2026-03-13 12:34:02.26675931 +0000 UTC m=+9016.289289443" observedRunningTime="2026-03-13 12:34:03.484200878 +0000 UTC m=+9017.506731021" watchObservedRunningTime="2026-03-13 12:34:03.486232128 +0000 UTC m=+9017.508762261" Mar 13 12:34:04 crc kubenswrapper[4632]: I0313 12:34:04.472643 4632 generic.go:334] "Generic (PLEG): container finished" podID="ccc70d73-b58b-4a2c-9bce-dc27405c5710" containerID="33c9be3390a29151e585ddbf79d6ef390b1a094d7878d4d4c96b9b0bb39d369c" exitCode=0 Mar 13 12:34:04 crc kubenswrapper[4632]: I0313 12:34:04.472845 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556754-kcjmj" event={"ID":"ccc70d73-b58b-4a2c-9bce-dc27405c5710","Type":"ContainerDied","Data":"33c9be3390a29151e585ddbf79d6ef390b1a094d7878d4d4c96b9b0bb39d369c"} Mar 13 12:34:06 crc kubenswrapper[4632]: I0313 12:34:06.005352 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556754-kcjmj" Mar 13 12:34:06 crc kubenswrapper[4632]: I0313 12:34:06.173431 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r2v6b\" (UniqueName: \"kubernetes.io/projected/ccc70d73-b58b-4a2c-9bce-dc27405c5710-kube-api-access-r2v6b\") pod \"ccc70d73-b58b-4a2c-9bce-dc27405c5710\" (UID: \"ccc70d73-b58b-4a2c-9bce-dc27405c5710\") " Mar 13 12:34:06 crc kubenswrapper[4632]: I0313 12:34:06.193445 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccc70d73-b58b-4a2c-9bce-dc27405c5710-kube-api-access-r2v6b" (OuterVolumeSpecName: "kube-api-access-r2v6b") pod "ccc70d73-b58b-4a2c-9bce-dc27405c5710" (UID: "ccc70d73-b58b-4a2c-9bce-dc27405c5710"). InnerVolumeSpecName "kube-api-access-r2v6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:34:06 crc kubenswrapper[4632]: I0313 12:34:06.276493 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r2v6b\" (UniqueName: \"kubernetes.io/projected/ccc70d73-b58b-4a2c-9bce-dc27405c5710-kube-api-access-r2v6b\") on node \"crc\" DevicePath \"\"" Mar 13 12:34:06 crc kubenswrapper[4632]: I0313 12:34:06.500899 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556754-kcjmj" event={"ID":"ccc70d73-b58b-4a2c-9bce-dc27405c5710","Type":"ContainerDied","Data":"8c18300927c4a79068fc1d8cdd84afa4a73b1ab68a0bfed9583785fb44288336"} Mar 13 12:34:06 crc kubenswrapper[4632]: I0313 12:34:06.502179 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c18300927c4a79068fc1d8cdd84afa4a73b1ab68a0bfed9583785fb44288336" Mar 13 12:34:06 crc kubenswrapper[4632]: I0313 12:34:06.501170 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556754-kcjmj" Mar 13 12:34:06 crc kubenswrapper[4632]: I0313 12:34:06.571493 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556748-tr74t"] Mar 13 12:34:06 crc kubenswrapper[4632]: I0313 12:34:06.579468 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556748-tr74t"] Mar 13 12:34:08 crc kubenswrapper[4632]: I0313 12:34:08.063395 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba59e3a2-3d83-4f9d-8633-788ba1bf518c" path="/var/lib/kubelet/pods/ba59e3a2-3d83-4f9d-8633-788ba1bf518c/volumes" Mar 13 12:34:10 crc kubenswrapper[4632]: I0313 12:34:10.461218 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:34:10 crc kubenswrapper[4632]: I0313 12:34:10.461618 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.312454 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tvzvt"] Mar 13 12:34:21 crc kubenswrapper[4632]: E0313 12:34:21.313567 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccc70d73-b58b-4a2c-9bce-dc27405c5710" containerName="oc" Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.313586 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccc70d73-b58b-4a2c-9bce-dc27405c5710" containerName="oc" Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.313858 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccc70d73-b58b-4a2c-9bce-dc27405c5710" containerName="oc" Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.317834 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.338930 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvzvt"] Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.503532 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tch8\" (UniqueName: \"kubernetes.io/projected/aefd0984-736f-4724-96e9-bba46baff210-kube-api-access-8tch8\") pod \"redhat-marketplace-tvzvt\" (UID: \"aefd0984-736f-4724-96e9-bba46baff210\") " pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.503639 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aefd0984-736f-4724-96e9-bba46baff210-utilities\") pod \"redhat-marketplace-tvzvt\" (UID: \"aefd0984-736f-4724-96e9-bba46baff210\") " pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.503840 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aefd0984-736f-4724-96e9-bba46baff210-catalog-content\") pod \"redhat-marketplace-tvzvt\" (UID: \"aefd0984-736f-4724-96e9-bba46baff210\") " pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.605800 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aefd0984-736f-4724-96e9-bba46baff210-catalog-content\") pod \"redhat-marketplace-tvzvt\" (UID: \"aefd0984-736f-4724-96e9-bba46baff210\") " pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.606001 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tch8\" (UniqueName: \"kubernetes.io/projected/aefd0984-736f-4724-96e9-bba46baff210-kube-api-access-8tch8\") pod \"redhat-marketplace-tvzvt\" (UID: \"aefd0984-736f-4724-96e9-bba46baff210\") " pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.606025 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aefd0984-736f-4724-96e9-bba46baff210-utilities\") pod \"redhat-marketplace-tvzvt\" (UID: \"aefd0984-736f-4724-96e9-bba46baff210\") " pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.606290 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aefd0984-736f-4724-96e9-bba46baff210-catalog-content\") pod \"redhat-marketplace-tvzvt\" (UID: \"aefd0984-736f-4724-96e9-bba46baff210\") " pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.606322 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aefd0984-736f-4724-96e9-bba46baff210-utilities\") pod \"redhat-marketplace-tvzvt\" (UID: \"aefd0984-736f-4724-96e9-bba46baff210\") " pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.629689 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tch8\" (UniqueName: \"kubernetes.io/projected/aefd0984-736f-4724-96e9-bba46baff210-kube-api-access-8tch8\") pod \"redhat-marketplace-tvzvt\" (UID: \"aefd0984-736f-4724-96e9-bba46baff210\") " pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:21 crc kubenswrapper[4632]: I0313 12:34:21.642381 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:22 crc kubenswrapper[4632]: I0313 12:34:22.384773 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvzvt"] Mar 13 12:34:22 crc kubenswrapper[4632]: I0313 12:34:22.756643 4632 generic.go:334] "Generic (PLEG): container finished" podID="aefd0984-736f-4724-96e9-bba46baff210" containerID="6459338d22358b0519f807dc4afc129c5e2ca57f7e8e5369a57abd6209dfd36f" exitCode=0 Mar 13 12:34:22 crc kubenswrapper[4632]: I0313 12:34:22.756714 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvzvt" event={"ID":"aefd0984-736f-4724-96e9-bba46baff210","Type":"ContainerDied","Data":"6459338d22358b0519f807dc4afc129c5e2ca57f7e8e5369a57abd6209dfd36f"} Mar 13 12:34:22 crc kubenswrapper[4632]: I0313 12:34:22.757007 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvzvt" event={"ID":"aefd0984-736f-4724-96e9-bba46baff210","Type":"ContainerStarted","Data":"3864ef386dc747b0066132dd4d6ee7a1758a98ab4c866a3de7ae196d02e95eaa"} Mar 13 12:34:23 crc kubenswrapper[4632]: I0313 12:34:23.781625 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvzvt" event={"ID":"aefd0984-736f-4724-96e9-bba46baff210","Type":"ContainerStarted","Data":"5100412017a2a84219c5731b435e390960254cc640066455806e36494b500748"} Mar 13 12:34:24 crc kubenswrapper[4632]: I0313 12:34:24.518099 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4crmg"] Mar 13 12:34:24 crc kubenswrapper[4632]: I0313 12:34:24.520605 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:24 crc kubenswrapper[4632]: I0313 12:34:24.556559 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4crmg"] Mar 13 12:34:24 crc kubenswrapper[4632]: I0313 12:34:24.717006 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxvbg\" (UniqueName: \"kubernetes.io/projected/99ae46aa-bead-481b-9416-e5a1a8be8196-kube-api-access-bxvbg\") pod \"certified-operators-4crmg\" (UID: \"99ae46aa-bead-481b-9416-e5a1a8be8196\") " pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:24 crc kubenswrapper[4632]: I0313 12:34:24.717137 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ae46aa-bead-481b-9416-e5a1a8be8196-utilities\") pod \"certified-operators-4crmg\" (UID: \"99ae46aa-bead-481b-9416-e5a1a8be8196\") " pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:24 crc kubenswrapper[4632]: I0313 12:34:24.717160 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ae46aa-bead-481b-9416-e5a1a8be8196-catalog-content\") pod \"certified-operators-4crmg\" (UID: \"99ae46aa-bead-481b-9416-e5a1a8be8196\") " pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:24 crc kubenswrapper[4632]: I0313 12:34:24.823435 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxvbg\" (UniqueName: \"kubernetes.io/projected/99ae46aa-bead-481b-9416-e5a1a8be8196-kube-api-access-bxvbg\") pod \"certified-operators-4crmg\" (UID: \"99ae46aa-bead-481b-9416-e5a1a8be8196\") " pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:24 crc kubenswrapper[4632]: I0313 12:34:24.824273 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ae46aa-bead-481b-9416-e5a1a8be8196-utilities\") pod \"certified-operators-4crmg\" (UID: \"99ae46aa-bead-481b-9416-e5a1a8be8196\") " pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:24 crc kubenswrapper[4632]: I0313 12:34:24.824313 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ae46aa-bead-481b-9416-e5a1a8be8196-catalog-content\") pod \"certified-operators-4crmg\" (UID: \"99ae46aa-bead-481b-9416-e5a1a8be8196\") " pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:24 crc kubenswrapper[4632]: I0313 12:34:24.882413 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ae46aa-bead-481b-9416-e5a1a8be8196-utilities\") pod \"certified-operators-4crmg\" (UID: \"99ae46aa-bead-481b-9416-e5a1a8be8196\") " pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:24 crc kubenswrapper[4632]: I0313 12:34:24.909363 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ae46aa-bead-481b-9416-e5a1a8be8196-catalog-content\") pod \"certified-operators-4crmg\" (UID: \"99ae46aa-bead-481b-9416-e5a1a8be8196\") " pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:24 crc kubenswrapper[4632]: I0313 12:34:24.918010 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxvbg\" (UniqueName: \"kubernetes.io/projected/99ae46aa-bead-481b-9416-e5a1a8be8196-kube-api-access-bxvbg\") pod \"certified-operators-4crmg\" (UID: \"99ae46aa-bead-481b-9416-e5a1a8be8196\") " pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:25 crc kubenswrapper[4632]: I0313 12:34:25.187220 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:25 crc kubenswrapper[4632]: I0313 12:34:25.814893 4632 generic.go:334] "Generic (PLEG): container finished" podID="aefd0984-736f-4724-96e9-bba46baff210" containerID="5100412017a2a84219c5731b435e390960254cc640066455806e36494b500748" exitCode=0 Mar 13 12:34:25 crc kubenswrapper[4632]: I0313 12:34:25.815207 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvzvt" event={"ID":"aefd0984-736f-4724-96e9-bba46baff210","Type":"ContainerDied","Data":"5100412017a2a84219c5731b435e390960254cc640066455806e36494b500748"} Mar 13 12:34:25 crc kubenswrapper[4632]: I0313 12:34:25.926396 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4crmg"] Mar 13 12:34:26 crc kubenswrapper[4632]: I0313 12:34:26.827534 4632 generic.go:334] "Generic (PLEG): container finished" podID="99ae46aa-bead-481b-9416-e5a1a8be8196" containerID="0f0837940cf88d21f3843c5930040e437ca0684ad084295f166ebe778c422cd2" exitCode=0 Mar 13 12:34:26 crc kubenswrapper[4632]: I0313 12:34:26.827730 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4crmg" event={"ID":"99ae46aa-bead-481b-9416-e5a1a8be8196","Type":"ContainerDied","Data":"0f0837940cf88d21f3843c5930040e437ca0684ad084295f166ebe778c422cd2"} Mar 13 12:34:26 crc kubenswrapper[4632]: I0313 12:34:26.828293 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4crmg" event={"ID":"99ae46aa-bead-481b-9416-e5a1a8be8196","Type":"ContainerStarted","Data":"306a0cb5021e67e31e2f58ee7bbf44297adb063f5cf868144b001c706cab0607"} Mar 13 12:34:26 crc kubenswrapper[4632]: I0313 12:34:26.832101 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvzvt" event={"ID":"aefd0984-736f-4724-96e9-bba46baff210","Type":"ContainerStarted","Data":"30ac8815243eb74ef150559b7957bdc7b42f74dcdd603007dfad8eb2fd4ee057"} Mar 13 12:34:26 crc kubenswrapper[4632]: I0313 12:34:26.928371 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tvzvt" podStartSLOduration=2.388036593 podStartE2EDuration="5.928351872s" podCreationTimestamp="2026-03-13 12:34:21 +0000 UTC" firstStartedPulling="2026-03-13 12:34:22.758598916 +0000 UTC m=+9036.781129049" lastFinishedPulling="2026-03-13 12:34:26.298914195 +0000 UTC m=+9040.321444328" observedRunningTime="2026-03-13 12:34:26.92579664 +0000 UTC m=+9040.948326773" watchObservedRunningTime="2026-03-13 12:34:26.928351872 +0000 UTC m=+9040.950882005" Mar 13 12:34:27 crc kubenswrapper[4632]: I0313 12:34:27.640677 4632 scope.go:117] "RemoveContainer" containerID="d4da70cb5943a7b88f9744961515085bb09badfa367e58e3aee43668a0864bc3" Mar 13 12:34:28 crc kubenswrapper[4632]: I0313 12:34:28.874081 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4crmg" event={"ID":"99ae46aa-bead-481b-9416-e5a1a8be8196","Type":"ContainerStarted","Data":"6dc2cddd4965efa9fc2c49fa5ca691ecbf53012540a04504d6c227a2347e30f4"} Mar 13 12:34:30 crc kubenswrapper[4632]: I0313 12:34:30.894890 4632 generic.go:334] "Generic (PLEG): container finished" podID="99ae46aa-bead-481b-9416-e5a1a8be8196" containerID="6dc2cddd4965efa9fc2c49fa5ca691ecbf53012540a04504d6c227a2347e30f4" exitCode=0 Mar 13 12:34:30 crc kubenswrapper[4632]: I0313 12:34:30.895054 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4crmg" event={"ID":"99ae46aa-bead-481b-9416-e5a1a8be8196","Type":"ContainerDied","Data":"6dc2cddd4965efa9fc2c49fa5ca691ecbf53012540a04504d6c227a2347e30f4"} Mar 13 12:34:31 crc kubenswrapper[4632]: I0313 12:34:31.642964 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:31 crc kubenswrapper[4632]: I0313 12:34:31.643389 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:31 crc kubenswrapper[4632]: I0313 12:34:31.905470 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4crmg" event={"ID":"99ae46aa-bead-481b-9416-e5a1a8be8196","Type":"ContainerStarted","Data":"6e7ba1abc2a88a5e2aebe5114716bbab411d8cc4124cd295d7b37f0ec9fc95e9"} Mar 13 12:34:31 crc kubenswrapper[4632]: I0313 12:34:31.929916 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4crmg" podStartSLOduration=3.233951158 podStartE2EDuration="7.929895809s" podCreationTimestamp="2026-03-13 12:34:24 +0000 UTC" firstStartedPulling="2026-03-13 12:34:26.829318764 +0000 UTC m=+9040.851848897" lastFinishedPulling="2026-03-13 12:34:31.525263415 +0000 UTC m=+9045.547793548" observedRunningTime="2026-03-13 12:34:31.924387964 +0000 UTC m=+9045.946918097" watchObservedRunningTime="2026-03-13 12:34:31.929895809 +0000 UTC m=+9045.952425932" Mar 13 12:34:32 crc kubenswrapper[4632]: I0313 12:34:32.714837 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-tvzvt" podUID="aefd0984-736f-4724-96e9-bba46baff210" containerName="registry-server" probeResult="failure" output=< Mar 13 12:34:32 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:34:32 crc kubenswrapper[4632]: > Mar 13 12:34:35 crc kubenswrapper[4632]: I0313 12:34:35.187433 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:35 crc kubenswrapper[4632]: I0313 12:34:35.187848 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:36 crc kubenswrapper[4632]: I0313 12:34:36.242548 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-4crmg" podUID="99ae46aa-bead-481b-9416-e5a1a8be8196" containerName="registry-server" probeResult="failure" output=< Mar 13 12:34:36 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:34:36 crc kubenswrapper[4632]: > Mar 13 12:34:40 crc kubenswrapper[4632]: I0313 12:34:40.461075 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:34:40 crc kubenswrapper[4632]: I0313 12:34:40.462223 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:34:40 crc kubenswrapper[4632]: I0313 12:34:40.462280 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 12:34:40 crc kubenswrapper[4632]: I0313 12:34:40.465285 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ada2bd3447f81dbcb3c7c10ab1a84d7a61b81476a09d5bccd655ef21929539af"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 12:34:40 crc kubenswrapper[4632]: I0313 12:34:40.466195 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://ada2bd3447f81dbcb3c7c10ab1a84d7a61b81476a09d5bccd655ef21929539af" gracePeriod=600 Mar 13 12:34:41 crc kubenswrapper[4632]: I0313 12:34:41.010844 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="ada2bd3447f81dbcb3c7c10ab1a84d7a61b81476a09d5bccd655ef21929539af" exitCode=0 Mar 13 12:34:41 crc kubenswrapper[4632]: I0313 12:34:41.010912 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"ada2bd3447f81dbcb3c7c10ab1a84d7a61b81476a09d5bccd655ef21929539af"} Mar 13 12:34:41 crc kubenswrapper[4632]: I0313 12:34:41.011233 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9"} Mar 13 12:34:41 crc kubenswrapper[4632]: I0313 12:34:41.011258 4632 scope.go:117] "RemoveContainer" containerID="8741648fe2a67d9da8cf15c1e305dc4ff749d2dd595a90ef09b36ac1d0767d1f" Mar 13 12:34:41 crc kubenswrapper[4632]: I0313 12:34:41.710357 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:41 crc kubenswrapper[4632]: I0313 12:34:41.770636 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:44 crc kubenswrapper[4632]: I0313 12:34:44.605609 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvzvt"] Mar 13 12:34:44 crc kubenswrapper[4632]: I0313 12:34:44.607127 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tvzvt" podUID="aefd0984-736f-4724-96e9-bba46baff210" containerName="registry-server" containerID="cri-o://30ac8815243eb74ef150559b7957bdc7b42f74dcdd603007dfad8eb2fd4ee057" gracePeriod=2 Mar 13 12:34:45 crc kubenswrapper[4632]: I0313 12:34:45.075738 4632 generic.go:334] "Generic (PLEG): container finished" podID="aefd0984-736f-4724-96e9-bba46baff210" containerID="30ac8815243eb74ef150559b7957bdc7b42f74dcdd603007dfad8eb2fd4ee057" exitCode=0 Mar 13 12:34:45 crc kubenswrapper[4632]: I0313 12:34:45.075831 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvzvt" event={"ID":"aefd0984-736f-4724-96e9-bba46baff210","Type":"ContainerDied","Data":"30ac8815243eb74ef150559b7957bdc7b42f74dcdd603007dfad8eb2fd4ee057"} Mar 13 12:34:45 crc kubenswrapper[4632]: I0313 12:34:45.838434 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:45 crc kubenswrapper[4632]: I0313 12:34:45.898597 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aefd0984-736f-4724-96e9-bba46baff210-utilities\") pod \"aefd0984-736f-4724-96e9-bba46baff210\" (UID: \"aefd0984-736f-4724-96e9-bba46baff210\") " Mar 13 12:34:45 crc kubenswrapper[4632]: I0313 12:34:45.898664 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tch8\" (UniqueName: \"kubernetes.io/projected/aefd0984-736f-4724-96e9-bba46baff210-kube-api-access-8tch8\") pod \"aefd0984-736f-4724-96e9-bba46baff210\" (UID: \"aefd0984-736f-4724-96e9-bba46baff210\") " Mar 13 12:34:45 crc kubenswrapper[4632]: I0313 12:34:45.898747 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aefd0984-736f-4724-96e9-bba46baff210-catalog-content\") pod \"aefd0984-736f-4724-96e9-bba46baff210\" (UID: \"aefd0984-736f-4724-96e9-bba46baff210\") " Mar 13 12:34:45 crc kubenswrapper[4632]: I0313 12:34:45.900238 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aefd0984-736f-4724-96e9-bba46baff210-utilities" (OuterVolumeSpecName: "utilities") pod "aefd0984-736f-4724-96e9-bba46baff210" (UID: "aefd0984-736f-4724-96e9-bba46baff210"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:34:45 crc kubenswrapper[4632]: I0313 12:34:45.911710 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aefd0984-736f-4724-96e9-bba46baff210-kube-api-access-8tch8" (OuterVolumeSpecName: "kube-api-access-8tch8") pod "aefd0984-736f-4724-96e9-bba46baff210" (UID: "aefd0984-736f-4724-96e9-bba46baff210"). InnerVolumeSpecName "kube-api-access-8tch8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:34:45 crc kubenswrapper[4632]: I0313 12:34:45.920296 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aefd0984-736f-4724-96e9-bba46baff210-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aefd0984-736f-4724-96e9-bba46baff210" (UID: "aefd0984-736f-4724-96e9-bba46baff210"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:34:46 crc kubenswrapper[4632]: I0313 12:34:46.001050 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aefd0984-736f-4724-96e9-bba46baff210-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:34:46 crc kubenswrapper[4632]: I0313 12:34:46.001317 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tch8\" (UniqueName: \"kubernetes.io/projected/aefd0984-736f-4724-96e9-bba46baff210-kube-api-access-8tch8\") on node \"crc\" DevicePath \"\"" Mar 13 12:34:46 crc kubenswrapper[4632]: I0313 12:34:46.001379 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aefd0984-736f-4724-96e9-bba46baff210-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:34:46 crc kubenswrapper[4632]: I0313 12:34:46.087770 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tvzvt" event={"ID":"aefd0984-736f-4724-96e9-bba46baff210","Type":"ContainerDied","Data":"3864ef386dc747b0066132dd4d6ee7a1758a98ab4c866a3de7ae196d02e95eaa"} Mar 13 12:34:46 crc kubenswrapper[4632]: I0313 12:34:46.088104 4632 scope.go:117] "RemoveContainer" containerID="30ac8815243eb74ef150559b7957bdc7b42f74dcdd603007dfad8eb2fd4ee057" Mar 13 12:34:46 crc kubenswrapper[4632]: I0313 12:34:46.088226 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tvzvt" Mar 13 12:34:46 crc kubenswrapper[4632]: I0313 12:34:46.148438 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvzvt"] Mar 13 12:34:46 crc kubenswrapper[4632]: I0313 12:34:46.154069 4632 scope.go:117] "RemoveContainer" containerID="5100412017a2a84219c5731b435e390960254cc640066455806e36494b500748" Mar 13 12:34:46 crc kubenswrapper[4632]: I0313 12:34:46.158995 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tvzvt"] Mar 13 12:34:46 crc kubenswrapper[4632]: I0313 12:34:46.188236 4632 scope.go:117] "RemoveContainer" containerID="6459338d22358b0519f807dc4afc129c5e2ca57f7e8e5369a57abd6209dfd36f" Mar 13 12:34:46 crc kubenswrapper[4632]: I0313 12:34:46.238188 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-4crmg" podUID="99ae46aa-bead-481b-9416-e5a1a8be8196" containerName="registry-server" probeResult="failure" output=< Mar 13 12:34:46 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:34:46 crc kubenswrapper[4632]: > Mar 13 12:34:48 crc kubenswrapper[4632]: I0313 12:34:48.057960 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aefd0984-736f-4724-96e9-bba46baff210" path="/var/lib/kubelet/pods/aefd0984-736f-4724-96e9-bba46baff210/volumes" Mar 13 12:34:55 crc kubenswrapper[4632]: I0313 12:34:55.264278 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:55 crc kubenswrapper[4632]: I0313 12:34:55.334212 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:55 crc kubenswrapper[4632]: I0313 12:34:55.702341 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4crmg"] Mar 13 12:34:57 crc kubenswrapper[4632]: I0313 12:34:57.192032 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4crmg" podUID="99ae46aa-bead-481b-9416-e5a1a8be8196" containerName="registry-server" containerID="cri-o://6e7ba1abc2a88a5e2aebe5114716bbab411d8cc4124cd295d7b37f0ec9fc95e9" gracePeriod=2 Mar 13 12:34:57 crc kubenswrapper[4632]: I0313 12:34:57.743895 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:57 crc kubenswrapper[4632]: I0313 12:34:57.936310 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ae46aa-bead-481b-9416-e5a1a8be8196-utilities\") pod \"99ae46aa-bead-481b-9416-e5a1a8be8196\" (UID: \"99ae46aa-bead-481b-9416-e5a1a8be8196\") " Mar 13 12:34:57 crc kubenswrapper[4632]: I0313 12:34:57.936455 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ae46aa-bead-481b-9416-e5a1a8be8196-catalog-content\") pod \"99ae46aa-bead-481b-9416-e5a1a8be8196\" (UID: \"99ae46aa-bead-481b-9416-e5a1a8be8196\") " Mar 13 12:34:57 crc kubenswrapper[4632]: I0313 12:34:57.936554 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxvbg\" (UniqueName: \"kubernetes.io/projected/99ae46aa-bead-481b-9416-e5a1a8be8196-kube-api-access-bxvbg\") pod \"99ae46aa-bead-481b-9416-e5a1a8be8196\" (UID: \"99ae46aa-bead-481b-9416-e5a1a8be8196\") " Mar 13 12:34:57 crc kubenswrapper[4632]: I0313 12:34:57.937600 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99ae46aa-bead-481b-9416-e5a1a8be8196-utilities" (OuterVolumeSpecName: "utilities") pod "99ae46aa-bead-481b-9416-e5a1a8be8196" (UID: "99ae46aa-bead-481b-9416-e5a1a8be8196"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:34:57 crc kubenswrapper[4632]: I0313 12:34:57.964105 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99ae46aa-bead-481b-9416-e5a1a8be8196-kube-api-access-bxvbg" (OuterVolumeSpecName: "kube-api-access-bxvbg") pod "99ae46aa-bead-481b-9416-e5a1a8be8196" (UID: "99ae46aa-bead-481b-9416-e5a1a8be8196"). InnerVolumeSpecName "kube-api-access-bxvbg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.018989 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99ae46aa-bead-481b-9416-e5a1a8be8196-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "99ae46aa-bead-481b-9416-e5a1a8be8196" (UID: "99ae46aa-bead-481b-9416-e5a1a8be8196"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.043369 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxvbg\" (UniqueName: \"kubernetes.io/projected/99ae46aa-bead-481b-9416-e5a1a8be8196-kube-api-access-bxvbg\") on node \"crc\" DevicePath \"\"" Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.043693 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99ae46aa-bead-481b-9416-e5a1a8be8196-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.043703 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99ae46aa-bead-481b-9416-e5a1a8be8196-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:34:58 crc kubenswrapper[4632]: E0313 12:34:58.153326 4632 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod99ae46aa_bead_481b_9416_e5a1a8be8196.slice\": RecentStats: unable to find data in memory cache]" Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.203996 4632 generic.go:334] "Generic (PLEG): container finished" podID="99ae46aa-bead-481b-9416-e5a1a8be8196" containerID="6e7ba1abc2a88a5e2aebe5114716bbab411d8cc4124cd295d7b37f0ec9fc95e9" exitCode=0 Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.204039 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4crmg" event={"ID":"99ae46aa-bead-481b-9416-e5a1a8be8196","Type":"ContainerDied","Data":"6e7ba1abc2a88a5e2aebe5114716bbab411d8cc4124cd295d7b37f0ec9fc95e9"} Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.204068 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4crmg" event={"ID":"99ae46aa-bead-481b-9416-e5a1a8be8196","Type":"ContainerDied","Data":"306a0cb5021e67e31e2f58ee7bbf44297adb063f5cf868144b001c706cab0607"} Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.204088 4632 scope.go:117] "RemoveContainer" containerID="6e7ba1abc2a88a5e2aebe5114716bbab411d8cc4124cd295d7b37f0ec9fc95e9" Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.204086 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4crmg" Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.233597 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4crmg"] Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.243621 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4crmg"] Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.246397 4632 scope.go:117] "RemoveContainer" containerID="6dc2cddd4965efa9fc2c49fa5ca691ecbf53012540a04504d6c227a2347e30f4" Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.269434 4632 scope.go:117] "RemoveContainer" containerID="0f0837940cf88d21f3843c5930040e437ca0684ad084295f166ebe778c422cd2" Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.313210 4632 scope.go:117] "RemoveContainer" containerID="6e7ba1abc2a88a5e2aebe5114716bbab411d8cc4124cd295d7b37f0ec9fc95e9" Mar 13 12:34:58 crc kubenswrapper[4632]: E0313 12:34:58.324117 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e7ba1abc2a88a5e2aebe5114716bbab411d8cc4124cd295d7b37f0ec9fc95e9\": container with ID starting with 6e7ba1abc2a88a5e2aebe5114716bbab411d8cc4124cd295d7b37f0ec9fc95e9 not found: ID does not exist" containerID="6e7ba1abc2a88a5e2aebe5114716bbab411d8cc4124cd295d7b37f0ec9fc95e9" Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.324190 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e7ba1abc2a88a5e2aebe5114716bbab411d8cc4124cd295d7b37f0ec9fc95e9"} err="failed to get container status \"6e7ba1abc2a88a5e2aebe5114716bbab411d8cc4124cd295d7b37f0ec9fc95e9\": rpc error: code = NotFound desc = could not find container \"6e7ba1abc2a88a5e2aebe5114716bbab411d8cc4124cd295d7b37f0ec9fc95e9\": container with ID starting with 6e7ba1abc2a88a5e2aebe5114716bbab411d8cc4124cd295d7b37f0ec9fc95e9 not found: ID does not exist" Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.324232 4632 scope.go:117] "RemoveContainer" containerID="6dc2cddd4965efa9fc2c49fa5ca691ecbf53012540a04504d6c227a2347e30f4" Mar 13 12:34:58 crc kubenswrapper[4632]: E0313 12:34:58.325046 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6dc2cddd4965efa9fc2c49fa5ca691ecbf53012540a04504d6c227a2347e30f4\": container with ID starting with 6dc2cddd4965efa9fc2c49fa5ca691ecbf53012540a04504d6c227a2347e30f4 not found: ID does not exist" containerID="6dc2cddd4965efa9fc2c49fa5ca691ecbf53012540a04504d6c227a2347e30f4" Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.325108 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6dc2cddd4965efa9fc2c49fa5ca691ecbf53012540a04504d6c227a2347e30f4"} err="failed to get container status \"6dc2cddd4965efa9fc2c49fa5ca691ecbf53012540a04504d6c227a2347e30f4\": rpc error: code = NotFound desc = could not find container \"6dc2cddd4965efa9fc2c49fa5ca691ecbf53012540a04504d6c227a2347e30f4\": container with ID starting with 6dc2cddd4965efa9fc2c49fa5ca691ecbf53012540a04504d6c227a2347e30f4 not found: ID does not exist" Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.325164 4632 scope.go:117] "RemoveContainer" containerID="0f0837940cf88d21f3843c5930040e437ca0684ad084295f166ebe778c422cd2" Mar 13 12:34:58 crc kubenswrapper[4632]: E0313 12:34:58.325804 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f0837940cf88d21f3843c5930040e437ca0684ad084295f166ebe778c422cd2\": container with ID starting with 0f0837940cf88d21f3843c5930040e437ca0684ad084295f166ebe778c422cd2 not found: ID does not exist" containerID="0f0837940cf88d21f3843c5930040e437ca0684ad084295f166ebe778c422cd2" Mar 13 12:34:58 crc kubenswrapper[4632]: I0313 12:34:58.325869 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f0837940cf88d21f3843c5930040e437ca0684ad084295f166ebe778c422cd2"} err="failed to get container status \"0f0837940cf88d21f3843c5930040e437ca0684ad084295f166ebe778c422cd2\": rpc error: code = NotFound desc = could not find container \"0f0837940cf88d21f3843c5930040e437ca0684ad084295f166ebe778c422cd2\": container with ID starting with 0f0837940cf88d21f3843c5930040e437ca0684ad084295f166ebe778c422cd2 not found: ID does not exist" Mar 13 12:35:00 crc kubenswrapper[4632]: I0313 12:35:00.057708 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99ae46aa-bead-481b-9416-e5a1a8be8196" path="/var/lib/kubelet/pods/99ae46aa-bead-481b-9416-e5a1a8be8196/volumes" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.273738 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556756-4d96k"] Mar 13 12:36:00 crc kubenswrapper[4632]: E0313 12:36:00.293386 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ae46aa-bead-481b-9416-e5a1a8be8196" containerName="registry-server" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.293857 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ae46aa-bead-481b-9416-e5a1a8be8196" containerName="registry-server" Mar 13 12:36:00 crc kubenswrapper[4632]: E0313 12:36:00.293958 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aefd0984-736f-4724-96e9-bba46baff210" containerName="registry-server" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.293969 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="aefd0984-736f-4724-96e9-bba46baff210" containerName="registry-server" Mar 13 12:36:00 crc kubenswrapper[4632]: E0313 12:36:00.293988 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aefd0984-736f-4724-96e9-bba46baff210" containerName="extract-utilities" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.293998 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="aefd0984-736f-4724-96e9-bba46baff210" containerName="extract-utilities" Mar 13 12:36:00 crc kubenswrapper[4632]: E0313 12:36:00.294035 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aefd0984-736f-4724-96e9-bba46baff210" containerName="extract-content" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.294044 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="aefd0984-736f-4724-96e9-bba46baff210" containerName="extract-content" Mar 13 12:36:00 crc kubenswrapper[4632]: E0313 12:36:00.294063 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ae46aa-bead-481b-9416-e5a1a8be8196" containerName="extract-content" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.294072 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ae46aa-bead-481b-9416-e5a1a8be8196" containerName="extract-content" Mar 13 12:36:00 crc kubenswrapper[4632]: E0313 12:36:00.294096 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99ae46aa-bead-481b-9416-e5a1a8be8196" containerName="extract-utilities" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.294105 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="99ae46aa-bead-481b-9416-e5a1a8be8196" containerName="extract-utilities" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.294434 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="99ae46aa-bead-481b-9416-e5a1a8be8196" containerName="registry-server" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.294463 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="aefd0984-736f-4724-96e9-bba46baff210" containerName="registry-server" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.296203 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556756-4d96k" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.302241 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556756-4d96k"] Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.328306 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.328315 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.328406 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.423402 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5gzh\" (UniqueName: \"kubernetes.io/projected/4bab5d92-9e39-4d06-98ae-8b9b50d50214-kube-api-access-f5gzh\") pod \"auto-csr-approver-29556756-4d96k\" (UID: \"4bab5d92-9e39-4d06-98ae-8b9b50d50214\") " pod="openshift-infra/auto-csr-approver-29556756-4d96k" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.525294 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5gzh\" (UniqueName: \"kubernetes.io/projected/4bab5d92-9e39-4d06-98ae-8b9b50d50214-kube-api-access-f5gzh\") pod \"auto-csr-approver-29556756-4d96k\" (UID: \"4bab5d92-9e39-4d06-98ae-8b9b50d50214\") " pod="openshift-infra/auto-csr-approver-29556756-4d96k" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.559805 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5gzh\" (UniqueName: \"kubernetes.io/projected/4bab5d92-9e39-4d06-98ae-8b9b50d50214-kube-api-access-f5gzh\") pod \"auto-csr-approver-29556756-4d96k\" (UID: \"4bab5d92-9e39-4d06-98ae-8b9b50d50214\") " pod="openshift-infra/auto-csr-approver-29556756-4d96k" Mar 13 12:36:00 crc kubenswrapper[4632]: I0313 12:36:00.619915 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556756-4d96k" Mar 13 12:36:01 crc kubenswrapper[4632]: I0313 12:36:01.443101 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556756-4d96k"] Mar 13 12:36:02 crc kubenswrapper[4632]: I0313 12:36:02.006033 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556756-4d96k" event={"ID":"4bab5d92-9e39-4d06-98ae-8b9b50d50214","Type":"ContainerStarted","Data":"8b8fe1afbf64559cb08d91b985e36d5327c22b066ce15bacd4281c3bcd4a5d21"} Mar 13 12:36:04 crc kubenswrapper[4632]: I0313 12:36:04.027505 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556756-4d96k" event={"ID":"4bab5d92-9e39-4d06-98ae-8b9b50d50214","Type":"ContainerStarted","Data":"a8e65f824fd306a694713a170d4b213522c9b5fbd2a9bb06608f463371bdb733"} Mar 13 12:36:04 crc kubenswrapper[4632]: I0313 12:36:04.047496 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556756-4d96k" podStartSLOduration=3.133595994 podStartE2EDuration="4.047476523s" podCreationTimestamp="2026-03-13 12:36:00 +0000 UTC" firstStartedPulling="2026-03-13 12:36:01.464251471 +0000 UTC m=+9135.486781604" lastFinishedPulling="2026-03-13 12:36:02.378132 +0000 UTC m=+9136.400662133" observedRunningTime="2026-03-13 12:36:04.04285304 +0000 UTC m=+9138.065383173" watchObservedRunningTime="2026-03-13 12:36:04.047476523 +0000 UTC m=+9138.070006656" Mar 13 12:36:07 crc kubenswrapper[4632]: I0313 12:36:07.066402 4632 generic.go:334] "Generic (PLEG): container finished" podID="4bab5d92-9e39-4d06-98ae-8b9b50d50214" containerID="a8e65f824fd306a694713a170d4b213522c9b5fbd2a9bb06608f463371bdb733" exitCode=0 Mar 13 12:36:07 crc kubenswrapper[4632]: I0313 12:36:07.066592 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556756-4d96k" event={"ID":"4bab5d92-9e39-4d06-98ae-8b9b50d50214","Type":"ContainerDied","Data":"a8e65f824fd306a694713a170d4b213522c9b5fbd2a9bb06608f463371bdb733"} Mar 13 12:36:08 crc kubenswrapper[4632]: I0313 12:36:08.673547 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556756-4d96k" Mar 13 12:36:08 crc kubenswrapper[4632]: I0313 12:36:08.794457 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5gzh\" (UniqueName: \"kubernetes.io/projected/4bab5d92-9e39-4d06-98ae-8b9b50d50214-kube-api-access-f5gzh\") pod \"4bab5d92-9e39-4d06-98ae-8b9b50d50214\" (UID: \"4bab5d92-9e39-4d06-98ae-8b9b50d50214\") " Mar 13 12:36:08 crc kubenswrapper[4632]: I0313 12:36:08.806208 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bab5d92-9e39-4d06-98ae-8b9b50d50214-kube-api-access-f5gzh" (OuterVolumeSpecName: "kube-api-access-f5gzh") pod "4bab5d92-9e39-4d06-98ae-8b9b50d50214" (UID: "4bab5d92-9e39-4d06-98ae-8b9b50d50214"). InnerVolumeSpecName "kube-api-access-f5gzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:36:08 crc kubenswrapper[4632]: I0313 12:36:08.896694 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5gzh\" (UniqueName: \"kubernetes.io/projected/4bab5d92-9e39-4d06-98ae-8b9b50d50214-kube-api-access-f5gzh\") on node \"crc\" DevicePath \"\"" Mar 13 12:36:09 crc kubenswrapper[4632]: I0313 12:36:09.092723 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556756-4d96k" event={"ID":"4bab5d92-9e39-4d06-98ae-8b9b50d50214","Type":"ContainerDied","Data":"8b8fe1afbf64559cb08d91b985e36d5327c22b066ce15bacd4281c3bcd4a5d21"} Mar 13 12:36:09 crc kubenswrapper[4632]: I0313 12:36:09.092767 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b8fe1afbf64559cb08d91b985e36d5327c22b066ce15bacd4281c3bcd4a5d21" Mar 13 12:36:09 crc kubenswrapper[4632]: I0313 12:36:09.092837 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556756-4d96k" Mar 13 12:36:09 crc kubenswrapper[4632]: I0313 12:36:09.255467 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556750-8fj49"] Mar 13 12:36:09 crc kubenswrapper[4632]: I0313 12:36:09.263522 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556750-8fj49"] Mar 13 12:36:10 crc kubenswrapper[4632]: I0313 12:36:10.058866 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cc083b6-fb70-478e-9824-d9eb3cb1fe5b" path="/var/lib/kubelet/pods/8cc083b6-fb70-478e-9824-d9eb3cb1fe5b/volumes" Mar 13 12:36:16 crc kubenswrapper[4632]: I0313 12:36:16.993889 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-v86rj"] Mar 13 12:36:16 crc kubenswrapper[4632]: E0313 12:36:16.994768 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bab5d92-9e39-4d06-98ae-8b9b50d50214" containerName="oc" Mar 13 12:36:16 crc kubenswrapper[4632]: I0313 12:36:16.994784 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bab5d92-9e39-4d06-98ae-8b9b50d50214" containerName="oc" Mar 13 12:36:16 crc kubenswrapper[4632]: I0313 12:36:16.995020 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bab5d92-9e39-4d06-98ae-8b9b50d50214" containerName="oc" Mar 13 12:36:16 crc kubenswrapper[4632]: I0313 12:36:16.998046 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:36:17 crc kubenswrapper[4632]: I0313 12:36:17.024573 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v86rj"] Mar 13 12:36:17 crc kubenswrapper[4632]: I0313 12:36:17.057735 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwltc\" (UniqueName: \"kubernetes.io/projected/ac09d633-ce71-480d-bc5e-d9be1d416b03-kube-api-access-zwltc\") pod \"redhat-operators-v86rj\" (UID: \"ac09d633-ce71-480d-bc5e-d9be1d416b03\") " pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:36:17 crc kubenswrapper[4632]: I0313 12:36:17.057848 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac09d633-ce71-480d-bc5e-d9be1d416b03-catalog-content\") pod \"redhat-operators-v86rj\" (UID: \"ac09d633-ce71-480d-bc5e-d9be1d416b03\") " pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:36:17 crc kubenswrapper[4632]: I0313 12:36:17.057998 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac09d633-ce71-480d-bc5e-d9be1d416b03-utilities\") pod \"redhat-operators-v86rj\" (UID: \"ac09d633-ce71-480d-bc5e-d9be1d416b03\") " pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:36:17 crc kubenswrapper[4632]: I0313 12:36:17.159611 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac09d633-ce71-480d-bc5e-d9be1d416b03-utilities\") pod \"redhat-operators-v86rj\" (UID: \"ac09d633-ce71-480d-bc5e-d9be1d416b03\") " pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:36:17 crc kubenswrapper[4632]: I0313 12:36:17.159746 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwltc\" (UniqueName: \"kubernetes.io/projected/ac09d633-ce71-480d-bc5e-d9be1d416b03-kube-api-access-zwltc\") pod \"redhat-operators-v86rj\" (UID: \"ac09d633-ce71-480d-bc5e-d9be1d416b03\") " pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:36:17 crc kubenswrapper[4632]: I0313 12:36:17.159853 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac09d633-ce71-480d-bc5e-d9be1d416b03-catalog-content\") pod \"redhat-operators-v86rj\" (UID: \"ac09d633-ce71-480d-bc5e-d9be1d416b03\") " pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:36:17 crc kubenswrapper[4632]: I0313 12:36:17.160323 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac09d633-ce71-480d-bc5e-d9be1d416b03-utilities\") pod \"redhat-operators-v86rj\" (UID: \"ac09d633-ce71-480d-bc5e-d9be1d416b03\") " pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:36:17 crc kubenswrapper[4632]: I0313 12:36:17.160635 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac09d633-ce71-480d-bc5e-d9be1d416b03-catalog-content\") pod \"redhat-operators-v86rj\" (UID: \"ac09d633-ce71-480d-bc5e-d9be1d416b03\") " pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:36:17 crc kubenswrapper[4632]: I0313 12:36:17.191879 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwltc\" (UniqueName: \"kubernetes.io/projected/ac09d633-ce71-480d-bc5e-d9be1d416b03-kube-api-access-zwltc\") pod \"redhat-operators-v86rj\" (UID: \"ac09d633-ce71-480d-bc5e-d9be1d416b03\") " pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:36:17 crc kubenswrapper[4632]: I0313 12:36:17.325426 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:36:17 crc kubenswrapper[4632]: I0313 12:36:17.847838 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-v86rj"] Mar 13 12:36:18 crc kubenswrapper[4632]: I0313 12:36:18.175613 4632 generic.go:334] "Generic (PLEG): container finished" podID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerID="29d8b3a65318b6e6951210b601e841a218c4f0ffd2cbfdd6b7e54cb15bb8b2a2" exitCode=0 Mar 13 12:36:18 crc kubenswrapper[4632]: I0313 12:36:18.175789 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v86rj" event={"ID":"ac09d633-ce71-480d-bc5e-d9be1d416b03","Type":"ContainerDied","Data":"29d8b3a65318b6e6951210b601e841a218c4f0ffd2cbfdd6b7e54cb15bb8b2a2"} Mar 13 12:36:18 crc kubenswrapper[4632]: I0313 12:36:18.175853 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v86rj" event={"ID":"ac09d633-ce71-480d-bc5e-d9be1d416b03","Type":"ContainerStarted","Data":"4cf7e4e180fdee7c7813c8d090f0a4ec39f241c14816db7a866702741112c8ae"} Mar 13 12:36:19 crc kubenswrapper[4632]: I0313 12:36:19.191696 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v86rj" event={"ID":"ac09d633-ce71-480d-bc5e-d9be1d416b03","Type":"ContainerStarted","Data":"1187b21667ee3759adaeaf1e0fd0b393da1afd6f412928f91725d748894e2904"} Mar 13 12:36:26 crc kubenswrapper[4632]: I0313 12:36:26.269884 4632 generic.go:334] "Generic (PLEG): container finished" podID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerID="1187b21667ee3759adaeaf1e0fd0b393da1afd6f412928f91725d748894e2904" exitCode=0 Mar 13 12:36:26 crc kubenswrapper[4632]: I0313 12:36:26.270001 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v86rj" event={"ID":"ac09d633-ce71-480d-bc5e-d9be1d416b03","Type":"ContainerDied","Data":"1187b21667ee3759adaeaf1e0fd0b393da1afd6f412928f91725d748894e2904"} Mar 13 12:36:27 crc kubenswrapper[4632]: I0313 12:36:27.282340 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v86rj" event={"ID":"ac09d633-ce71-480d-bc5e-d9be1d416b03","Type":"ContainerStarted","Data":"21822b608332f376e9e1af2cdb22e1aa5756c02a18555a5fbfad899b33d82cd4"} Mar 13 12:36:27 crc kubenswrapper[4632]: I0313 12:36:27.310425 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-v86rj" podStartSLOduration=2.813055071 podStartE2EDuration="11.310401162s" podCreationTimestamp="2026-03-13 12:36:16 +0000 UTC" firstStartedPulling="2026-03-13 12:36:18.17722794 +0000 UTC m=+9152.199758083" lastFinishedPulling="2026-03-13 12:36:26.674574041 +0000 UTC m=+9160.697104174" observedRunningTime="2026-03-13 12:36:27.30009038 +0000 UTC m=+9161.322620523" watchObservedRunningTime="2026-03-13 12:36:27.310401162 +0000 UTC m=+9161.332931325" Mar 13 12:36:27 crc kubenswrapper[4632]: I0313 12:36:27.326103 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:36:27 crc kubenswrapper[4632]: I0313 12:36:27.326164 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:36:27 crc kubenswrapper[4632]: I0313 12:36:27.843429 4632 scope.go:117] "RemoveContainer" containerID="50ec5eed6591caef46ce66e044fc885293f40a008476cefa9221d3ccb1262877" Mar 13 12:36:28 crc kubenswrapper[4632]: I0313 12:36:28.379695 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v86rj" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="registry-server" probeResult="failure" output=< Mar 13 12:36:28 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:36:28 crc kubenswrapper[4632]: > Mar 13 12:36:38 crc kubenswrapper[4632]: I0313 12:36:38.388928 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v86rj" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="registry-server" probeResult="failure" output=< Mar 13 12:36:38 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:36:38 crc kubenswrapper[4632]: > Mar 13 12:36:40 crc kubenswrapper[4632]: I0313 12:36:40.462673 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:36:40 crc kubenswrapper[4632]: I0313 12:36:40.470435 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:36:48 crc kubenswrapper[4632]: I0313 12:36:48.386520 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v86rj" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="registry-server" probeResult="failure" output=< Mar 13 12:36:48 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:36:48 crc kubenswrapper[4632]: > Mar 13 12:36:58 crc kubenswrapper[4632]: I0313 12:36:58.381573 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v86rj" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="registry-server" probeResult="failure" output=< Mar 13 12:36:58 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:36:58 crc kubenswrapper[4632]: > Mar 13 12:37:08 crc kubenswrapper[4632]: I0313 12:37:08.375140 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v86rj" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="registry-server" probeResult="failure" output=< Mar 13 12:37:08 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:37:08 crc kubenswrapper[4632]: > Mar 13 12:37:10 crc kubenswrapper[4632]: I0313 12:37:10.464142 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:37:10 crc kubenswrapper[4632]: I0313 12:37:10.464404 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:37:18 crc kubenswrapper[4632]: I0313 12:37:18.396613 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v86rj" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="registry-server" probeResult="failure" output=< Mar 13 12:37:18 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:37:18 crc kubenswrapper[4632]: > Mar 13 12:37:28 crc kubenswrapper[4632]: I0313 12:37:28.396601 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-v86rj" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="registry-server" probeResult="failure" output=< Mar 13 12:37:28 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:37:28 crc kubenswrapper[4632]: > Mar 13 12:37:37 crc kubenswrapper[4632]: I0313 12:37:37.389583 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:37:37 crc kubenswrapper[4632]: I0313 12:37:37.452083 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:37:37 crc kubenswrapper[4632]: I0313 12:37:37.670156 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v86rj"] Mar 13 12:37:39 crc kubenswrapper[4632]: I0313 12:37:39.167654 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-v86rj" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="registry-server" containerID="cri-o://21822b608332f376e9e1af2cdb22e1aa5756c02a18555a5fbfad899b33d82cd4" gracePeriod=2 Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.179500 4632 generic.go:334] "Generic (PLEG): container finished" podID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerID="21822b608332f376e9e1af2cdb22e1aa5756c02a18555a5fbfad899b33d82cd4" exitCode=0 Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.180279 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v86rj" event={"ID":"ac09d633-ce71-480d-bc5e-d9be1d416b03","Type":"ContainerDied","Data":"21822b608332f376e9e1af2cdb22e1aa5756c02a18555a5fbfad899b33d82cd4"} Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.426522 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.460674 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.460766 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.460827 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.461772 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.461889 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" gracePeriod=600 Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.480838 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac09d633-ce71-480d-bc5e-d9be1d416b03-utilities\") pod \"ac09d633-ce71-480d-bc5e-d9be1d416b03\" (UID: \"ac09d633-ce71-480d-bc5e-d9be1d416b03\") " Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.480959 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwltc\" (UniqueName: \"kubernetes.io/projected/ac09d633-ce71-480d-bc5e-d9be1d416b03-kube-api-access-zwltc\") pod \"ac09d633-ce71-480d-bc5e-d9be1d416b03\" (UID: \"ac09d633-ce71-480d-bc5e-d9be1d416b03\") " Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.481077 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac09d633-ce71-480d-bc5e-d9be1d416b03-catalog-content\") pod \"ac09d633-ce71-480d-bc5e-d9be1d416b03\" (UID: \"ac09d633-ce71-480d-bc5e-d9be1d416b03\") " Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.491141 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac09d633-ce71-480d-bc5e-d9be1d416b03-utilities" (OuterVolumeSpecName: "utilities") pod "ac09d633-ce71-480d-bc5e-d9be1d416b03" (UID: "ac09d633-ce71-480d-bc5e-d9be1d416b03"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.507502 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac09d633-ce71-480d-bc5e-d9be1d416b03-kube-api-access-zwltc" (OuterVolumeSpecName: "kube-api-access-zwltc") pod "ac09d633-ce71-480d-bc5e-d9be1d416b03" (UID: "ac09d633-ce71-480d-bc5e-d9be1d416b03"). InnerVolumeSpecName "kube-api-access-zwltc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.584247 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac09d633-ce71-480d-bc5e-d9be1d416b03-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.584288 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwltc\" (UniqueName: \"kubernetes.io/projected/ac09d633-ce71-480d-bc5e-d9be1d416b03-kube-api-access-zwltc\") on node \"crc\" DevicePath \"\"" Mar 13 12:37:40 crc kubenswrapper[4632]: E0313 12:37:40.606604 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.763089 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac09d633-ce71-480d-bc5e-d9be1d416b03-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac09d633-ce71-480d-bc5e-d9be1d416b03" (UID: "ac09d633-ce71-480d-bc5e-d9be1d416b03"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:37:40 crc kubenswrapper[4632]: I0313 12:37:40.786775 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac09d633-ce71-480d-bc5e-d9be1d416b03-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:37:41 crc kubenswrapper[4632]: I0313 12:37:41.204495 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" exitCode=0 Mar 13 12:37:41 crc kubenswrapper[4632]: I0313 12:37:41.205513 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9"} Mar 13 12:37:41 crc kubenswrapper[4632]: I0313 12:37:41.207729 4632 scope.go:117] "RemoveContainer" containerID="ada2bd3447f81dbcb3c7c10ab1a84d7a61b81476a09d5bccd655ef21929539af" Mar 13 12:37:41 crc kubenswrapper[4632]: I0313 12:37:41.208794 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:37:41 crc kubenswrapper[4632]: E0313 12:37:41.209812 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:37:41 crc kubenswrapper[4632]: I0313 12:37:41.210700 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-v86rj" event={"ID":"ac09d633-ce71-480d-bc5e-d9be1d416b03","Type":"ContainerDied","Data":"4cf7e4e180fdee7c7813c8d090f0a4ec39f241c14816db7a866702741112c8ae"} Mar 13 12:37:41 crc kubenswrapper[4632]: I0313 12:37:41.210806 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-v86rj" Mar 13 12:37:41 crc kubenswrapper[4632]: I0313 12:37:41.279763 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-v86rj"] Mar 13 12:37:41 crc kubenswrapper[4632]: I0313 12:37:41.290549 4632 scope.go:117] "RemoveContainer" containerID="21822b608332f376e9e1af2cdb22e1aa5756c02a18555a5fbfad899b33d82cd4" Mar 13 12:37:41 crc kubenswrapper[4632]: I0313 12:37:41.291272 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-v86rj"] Mar 13 12:37:41 crc kubenswrapper[4632]: I0313 12:37:41.322384 4632 scope.go:117] "RemoveContainer" containerID="1187b21667ee3759adaeaf1e0fd0b393da1afd6f412928f91725d748894e2904" Mar 13 12:37:41 crc kubenswrapper[4632]: I0313 12:37:41.384126 4632 scope.go:117] "RemoveContainer" containerID="29d8b3a65318b6e6951210b601e841a218c4f0ffd2cbfdd6b7e54cb15bb8b2a2" Mar 13 12:37:42 crc kubenswrapper[4632]: I0313 12:37:42.058325 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" path="/var/lib/kubelet/pods/ac09d633-ce71-480d-bc5e-d9be1d416b03/volumes" Mar 13 12:37:55 crc kubenswrapper[4632]: I0313 12:37:55.044043 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:37:55 crc kubenswrapper[4632]: E0313 12:37:55.045025 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:38:00 crc kubenswrapper[4632]: I0313 12:38:00.225899 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556758-989sl"] Mar 13 12:38:00 crc kubenswrapper[4632]: E0313 12:38:00.229032 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="extract-content" Mar 13 12:38:00 crc kubenswrapper[4632]: I0313 12:38:00.229061 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="extract-content" Mar 13 12:38:00 crc kubenswrapper[4632]: E0313 12:38:00.229086 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="registry-server" Mar 13 12:38:00 crc kubenswrapper[4632]: I0313 12:38:00.229092 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="registry-server" Mar 13 12:38:00 crc kubenswrapper[4632]: E0313 12:38:00.229116 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="extract-utilities" Mar 13 12:38:00 crc kubenswrapper[4632]: I0313 12:38:00.229123 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="extract-utilities" Mar 13 12:38:00 crc kubenswrapper[4632]: I0313 12:38:00.229349 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac09d633-ce71-480d-bc5e-d9be1d416b03" containerName="registry-server" Mar 13 12:38:00 crc kubenswrapper[4632]: I0313 12:38:00.238765 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556758-989sl" Mar 13 12:38:00 crc kubenswrapper[4632]: I0313 12:38:00.261532 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:38:00 crc kubenswrapper[4632]: I0313 12:38:00.263080 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:38:00 crc kubenswrapper[4632]: I0313 12:38:00.263249 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:38:00 crc kubenswrapper[4632]: I0313 12:38:00.322968 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556758-989sl"] Mar 13 12:38:00 crc kubenswrapper[4632]: I0313 12:38:00.402384 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8f58\" (UniqueName: \"kubernetes.io/projected/564cb4aa-8722-4f69-adeb-16bc8b74bff0-kube-api-access-j8f58\") pod \"auto-csr-approver-29556758-989sl\" (UID: \"564cb4aa-8722-4f69-adeb-16bc8b74bff0\") " pod="openshift-infra/auto-csr-approver-29556758-989sl" Mar 13 12:38:00 crc kubenswrapper[4632]: I0313 12:38:00.504999 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8f58\" (UniqueName: \"kubernetes.io/projected/564cb4aa-8722-4f69-adeb-16bc8b74bff0-kube-api-access-j8f58\") pod \"auto-csr-approver-29556758-989sl\" (UID: \"564cb4aa-8722-4f69-adeb-16bc8b74bff0\") " pod="openshift-infra/auto-csr-approver-29556758-989sl" Mar 13 12:38:00 crc kubenswrapper[4632]: I0313 12:38:00.525816 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8f58\" (UniqueName: \"kubernetes.io/projected/564cb4aa-8722-4f69-adeb-16bc8b74bff0-kube-api-access-j8f58\") pod \"auto-csr-approver-29556758-989sl\" (UID: \"564cb4aa-8722-4f69-adeb-16bc8b74bff0\") " pod="openshift-infra/auto-csr-approver-29556758-989sl" Mar 13 12:38:00 crc kubenswrapper[4632]: I0313 12:38:00.576932 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556758-989sl" Mar 13 12:38:01 crc kubenswrapper[4632]: I0313 12:38:01.149109 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556758-989sl"] Mar 13 12:38:01 crc kubenswrapper[4632]: I0313 12:38:01.172447 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 12:38:01 crc kubenswrapper[4632]: I0313 12:38:01.428777 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556758-989sl" event={"ID":"564cb4aa-8722-4f69-adeb-16bc8b74bff0","Type":"ContainerStarted","Data":"5e6178bec38703640c7aad295f130226587694e4d95f5a4c4a588ca736c04131"} Mar 13 12:38:04 crc kubenswrapper[4632]: I0313 12:38:04.455952 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556758-989sl" event={"ID":"564cb4aa-8722-4f69-adeb-16bc8b74bff0","Type":"ContainerStarted","Data":"e9210685ea8bbb4f63a3e5f03db6817cbf29b557c698d4de6b1ae5929063a7c0"} Mar 13 12:38:04 crc kubenswrapper[4632]: I0313 12:38:04.482448 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556758-989sl" podStartSLOduration=2.553739883 podStartE2EDuration="4.480864357s" podCreationTimestamp="2026-03-13 12:38:00 +0000 UTC" firstStartedPulling="2026-03-13 12:38:01.169966432 +0000 UTC m=+9255.192496555" lastFinishedPulling="2026-03-13 12:38:03.097090886 +0000 UTC m=+9257.119621029" observedRunningTime="2026-03-13 12:38:04.468227837 +0000 UTC m=+9258.490757970" watchObservedRunningTime="2026-03-13 12:38:04.480864357 +0000 UTC m=+9258.503394490" Mar 13 12:38:05 crc kubenswrapper[4632]: I0313 12:38:05.467343 4632 generic.go:334] "Generic (PLEG): container finished" podID="564cb4aa-8722-4f69-adeb-16bc8b74bff0" containerID="e9210685ea8bbb4f63a3e5f03db6817cbf29b557c698d4de6b1ae5929063a7c0" exitCode=0 Mar 13 12:38:05 crc kubenswrapper[4632]: I0313 12:38:05.467443 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556758-989sl" event={"ID":"564cb4aa-8722-4f69-adeb-16bc8b74bff0","Type":"ContainerDied","Data":"e9210685ea8bbb4f63a3e5f03db6817cbf29b557c698d4de6b1ae5929063a7c0"} Mar 13 12:38:06 crc kubenswrapper[4632]: I0313 12:38:06.871219 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556758-989sl" Mar 13 12:38:07 crc kubenswrapper[4632]: I0313 12:38:07.032419 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8f58\" (UniqueName: \"kubernetes.io/projected/564cb4aa-8722-4f69-adeb-16bc8b74bff0-kube-api-access-j8f58\") pod \"564cb4aa-8722-4f69-adeb-16bc8b74bff0\" (UID: \"564cb4aa-8722-4f69-adeb-16bc8b74bff0\") " Mar 13 12:38:07 crc kubenswrapper[4632]: I0313 12:38:07.045255 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/564cb4aa-8722-4f69-adeb-16bc8b74bff0-kube-api-access-j8f58" (OuterVolumeSpecName: "kube-api-access-j8f58") pod "564cb4aa-8722-4f69-adeb-16bc8b74bff0" (UID: "564cb4aa-8722-4f69-adeb-16bc8b74bff0"). InnerVolumeSpecName "kube-api-access-j8f58". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:38:07 crc kubenswrapper[4632]: I0313 12:38:07.135546 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8f58\" (UniqueName: \"kubernetes.io/projected/564cb4aa-8722-4f69-adeb-16bc8b74bff0-kube-api-access-j8f58\") on node \"crc\" DevicePath \"\"" Mar 13 12:38:07 crc kubenswrapper[4632]: I0313 12:38:07.487791 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556758-989sl" event={"ID":"564cb4aa-8722-4f69-adeb-16bc8b74bff0","Type":"ContainerDied","Data":"5e6178bec38703640c7aad295f130226587694e4d95f5a4c4a588ca736c04131"} Mar 13 12:38:07 crc kubenswrapper[4632]: I0313 12:38:07.487837 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e6178bec38703640c7aad295f130226587694e4d95f5a4c4a588ca736c04131" Mar 13 12:38:07 crc kubenswrapper[4632]: I0313 12:38:07.487863 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556758-989sl" Mar 13 12:38:07 crc kubenswrapper[4632]: I0313 12:38:07.573744 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556752-48xgt"] Mar 13 12:38:07 crc kubenswrapper[4632]: I0313 12:38:07.584206 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556752-48xgt"] Mar 13 12:38:08 crc kubenswrapper[4632]: I0313 12:38:08.054096 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:38:08 crc kubenswrapper[4632]: E0313 12:38:08.055794 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:38:08 crc kubenswrapper[4632]: I0313 12:38:08.069043 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c311ec54-27ae-4082-bd23-4df180976b2f" path="/var/lib/kubelet/pods/c311ec54-27ae-4082-bd23-4df180976b2f/volumes" Mar 13 12:38:20 crc kubenswrapper[4632]: I0313 12:38:20.044737 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:38:20 crc kubenswrapper[4632]: E0313 12:38:20.045625 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:38:28 crc kubenswrapper[4632]: I0313 12:38:28.103447 4632 scope.go:117] "RemoveContainer" containerID="997fb6aac287ee23705f38733ad6b8cf02cea468d3978ae35a30d75ea0dfec0f" Mar 13 12:38:35 crc kubenswrapper[4632]: I0313 12:38:35.045145 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:38:35 crc kubenswrapper[4632]: E0313 12:38:35.045910 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:38:49 crc kubenswrapper[4632]: I0313 12:38:49.044481 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:38:49 crc kubenswrapper[4632]: E0313 12:38:49.045405 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:39:00 crc kubenswrapper[4632]: I0313 12:39:00.045536 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:39:00 crc kubenswrapper[4632]: E0313 12:39:00.046485 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:39:13 crc kubenswrapper[4632]: I0313 12:39:13.062490 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:39:13 crc kubenswrapper[4632]: E0313 12:39:13.064008 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:39:28 crc kubenswrapper[4632]: I0313 12:39:28.055640 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:39:28 crc kubenswrapper[4632]: E0313 12:39:28.056401 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:39:41 crc kubenswrapper[4632]: I0313 12:39:41.044756 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:39:41 crc kubenswrapper[4632]: E0313 12:39:41.045833 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:39:55 crc kubenswrapper[4632]: I0313 12:39:55.045754 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:39:55 crc kubenswrapper[4632]: E0313 12:39:55.046482 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:40:00 crc kubenswrapper[4632]: I0313 12:40:00.162310 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556760-fxrw2"] Mar 13 12:40:00 crc kubenswrapper[4632]: E0313 12:40:00.167234 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="564cb4aa-8722-4f69-adeb-16bc8b74bff0" containerName="oc" Mar 13 12:40:00 crc kubenswrapper[4632]: I0313 12:40:00.167285 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="564cb4aa-8722-4f69-adeb-16bc8b74bff0" containerName="oc" Mar 13 12:40:00 crc kubenswrapper[4632]: I0313 12:40:00.167661 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="564cb4aa-8722-4f69-adeb-16bc8b74bff0" containerName="oc" Mar 13 12:40:00 crc kubenswrapper[4632]: I0313 12:40:00.169844 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556760-fxrw2" Mar 13 12:40:00 crc kubenswrapper[4632]: I0313 12:40:00.175775 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:40:00 crc kubenswrapper[4632]: I0313 12:40:00.176042 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:40:00 crc kubenswrapper[4632]: I0313 12:40:00.181612 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:40:00 crc kubenswrapper[4632]: I0313 12:40:00.184126 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556760-fxrw2"] Mar 13 12:40:00 crc kubenswrapper[4632]: I0313 12:40:00.238988 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fpcc\" (UniqueName: \"kubernetes.io/projected/58d6a88c-498f-4887-998a-c3e3a1a2fef2-kube-api-access-8fpcc\") pod \"auto-csr-approver-29556760-fxrw2\" (UID: \"58d6a88c-498f-4887-998a-c3e3a1a2fef2\") " pod="openshift-infra/auto-csr-approver-29556760-fxrw2" Mar 13 12:40:00 crc kubenswrapper[4632]: I0313 12:40:00.353118 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fpcc\" (UniqueName: \"kubernetes.io/projected/58d6a88c-498f-4887-998a-c3e3a1a2fef2-kube-api-access-8fpcc\") pod \"auto-csr-approver-29556760-fxrw2\" (UID: \"58d6a88c-498f-4887-998a-c3e3a1a2fef2\") " pod="openshift-infra/auto-csr-approver-29556760-fxrw2" Mar 13 12:40:00 crc kubenswrapper[4632]: I0313 12:40:00.398627 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fpcc\" (UniqueName: \"kubernetes.io/projected/58d6a88c-498f-4887-998a-c3e3a1a2fef2-kube-api-access-8fpcc\") pod \"auto-csr-approver-29556760-fxrw2\" (UID: \"58d6a88c-498f-4887-998a-c3e3a1a2fef2\") " pod="openshift-infra/auto-csr-approver-29556760-fxrw2" Mar 13 12:40:00 crc kubenswrapper[4632]: I0313 12:40:00.498708 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556760-fxrw2" Mar 13 12:40:01 crc kubenswrapper[4632]: I0313 12:40:01.047448 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556760-fxrw2"] Mar 13 12:40:01 crc kubenswrapper[4632]: W0313 12:40:01.061673 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58d6a88c_498f_4887_998a_c3e3a1a2fef2.slice/crio-b5bb002e8e749efb2b27f87a90b9dd24c01d801dd48b684fd281e34cb8a2bc22 WatchSource:0}: Error finding container b5bb002e8e749efb2b27f87a90b9dd24c01d801dd48b684fd281e34cb8a2bc22: Status 404 returned error can't find the container with id b5bb002e8e749efb2b27f87a90b9dd24c01d801dd48b684fd281e34cb8a2bc22 Mar 13 12:40:01 crc kubenswrapper[4632]: I0313 12:40:01.581347 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556760-fxrw2" event={"ID":"58d6a88c-498f-4887-998a-c3e3a1a2fef2","Type":"ContainerStarted","Data":"b5bb002e8e749efb2b27f87a90b9dd24c01d801dd48b684fd281e34cb8a2bc22"} Mar 13 12:40:04 crc kubenswrapper[4632]: I0313 12:40:04.619791 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556760-fxrw2" event={"ID":"58d6a88c-498f-4887-998a-c3e3a1a2fef2","Type":"ContainerStarted","Data":"a5bdb6d7b1972d01ea3faadd8b4d91d40f96718626d20034621dcf3eda3e5f37"} Mar 13 12:40:04 crc kubenswrapper[4632]: I0313 12:40:04.652983 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556760-fxrw2" podStartSLOduration=2.7252013550000003 podStartE2EDuration="4.652917154s" podCreationTimestamp="2026-03-13 12:40:00 +0000 UTC" firstStartedPulling="2026-03-13 12:40:01.069840625 +0000 UTC m=+9375.092370788" lastFinishedPulling="2026-03-13 12:40:02.997556444 +0000 UTC m=+9377.020086587" observedRunningTime="2026-03-13 12:40:04.641502724 +0000 UTC m=+9378.664032867" watchObservedRunningTime="2026-03-13 12:40:04.652917154 +0000 UTC m=+9378.675447287" Mar 13 12:40:05 crc kubenswrapper[4632]: I0313 12:40:05.632898 4632 generic.go:334] "Generic (PLEG): container finished" podID="58d6a88c-498f-4887-998a-c3e3a1a2fef2" containerID="a5bdb6d7b1972d01ea3faadd8b4d91d40f96718626d20034621dcf3eda3e5f37" exitCode=0 Mar 13 12:40:05 crc kubenswrapper[4632]: I0313 12:40:05.633136 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556760-fxrw2" event={"ID":"58d6a88c-498f-4887-998a-c3e3a1a2fef2","Type":"ContainerDied","Data":"a5bdb6d7b1972d01ea3faadd8b4d91d40f96718626d20034621dcf3eda3e5f37"} Mar 13 12:40:07 crc kubenswrapper[4632]: I0313 12:40:07.008823 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556760-fxrw2" Mar 13 12:40:07 crc kubenswrapper[4632]: I0313 12:40:07.097773 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fpcc\" (UniqueName: \"kubernetes.io/projected/58d6a88c-498f-4887-998a-c3e3a1a2fef2-kube-api-access-8fpcc\") pod \"58d6a88c-498f-4887-998a-c3e3a1a2fef2\" (UID: \"58d6a88c-498f-4887-998a-c3e3a1a2fef2\") " Mar 13 12:40:07 crc kubenswrapper[4632]: I0313 12:40:07.106697 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58d6a88c-498f-4887-998a-c3e3a1a2fef2-kube-api-access-8fpcc" (OuterVolumeSpecName: "kube-api-access-8fpcc") pod "58d6a88c-498f-4887-998a-c3e3a1a2fef2" (UID: "58d6a88c-498f-4887-998a-c3e3a1a2fef2"). InnerVolumeSpecName "kube-api-access-8fpcc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:40:07 crc kubenswrapper[4632]: I0313 12:40:07.200931 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fpcc\" (UniqueName: \"kubernetes.io/projected/58d6a88c-498f-4887-998a-c3e3a1a2fef2-kube-api-access-8fpcc\") on node \"crc\" DevicePath \"\"" Mar 13 12:40:07 crc kubenswrapper[4632]: I0313 12:40:07.660661 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556760-fxrw2" event={"ID":"58d6a88c-498f-4887-998a-c3e3a1a2fef2","Type":"ContainerDied","Data":"b5bb002e8e749efb2b27f87a90b9dd24c01d801dd48b684fd281e34cb8a2bc22"} Mar 13 12:40:07 crc kubenswrapper[4632]: I0313 12:40:07.660732 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556760-fxrw2" Mar 13 12:40:07 crc kubenswrapper[4632]: I0313 12:40:07.661171 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5bb002e8e749efb2b27f87a90b9dd24c01d801dd48b684fd281e34cb8a2bc22" Mar 13 12:40:07 crc kubenswrapper[4632]: I0313 12:40:07.737769 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556754-kcjmj"] Mar 13 12:40:07 crc kubenswrapper[4632]: I0313 12:40:07.746631 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556754-kcjmj"] Mar 13 12:40:08 crc kubenswrapper[4632]: I0313 12:40:08.063786 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccc70d73-b58b-4a2c-9bce-dc27405c5710" path="/var/lib/kubelet/pods/ccc70d73-b58b-4a2c-9bce-dc27405c5710/volumes" Mar 13 12:40:09 crc kubenswrapper[4632]: I0313 12:40:09.046148 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:40:09 crc kubenswrapper[4632]: E0313 12:40:09.047020 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:40:20 crc kubenswrapper[4632]: I0313 12:40:20.044049 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:40:20 crc kubenswrapper[4632]: E0313 12:40:20.044640 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:40:28 crc kubenswrapper[4632]: I0313 12:40:28.347137 4632 scope.go:117] "RemoveContainer" containerID="33c9be3390a29151e585ddbf79d6ef390b1a094d7878d4d4c96b9b0bb39d369c" Mar 13 12:40:33 crc kubenswrapper[4632]: I0313 12:40:33.044905 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:40:33 crc kubenswrapper[4632]: E0313 12:40:33.045471 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:40:44 crc kubenswrapper[4632]: I0313 12:40:44.045159 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:40:44 crc kubenswrapper[4632]: E0313 12:40:44.046022 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:40:55 crc kubenswrapper[4632]: I0313 12:40:55.046163 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:40:55 crc kubenswrapper[4632]: E0313 12:40:55.049704 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:41:10 crc kubenswrapper[4632]: I0313 12:41:10.044847 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:41:10 crc kubenswrapper[4632]: E0313 12:41:10.045600 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:41:25 crc kubenswrapper[4632]: I0313 12:41:25.044302 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:41:25 crc kubenswrapper[4632]: E0313 12:41:25.045035 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:41:38 crc kubenswrapper[4632]: I0313 12:41:38.052109 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:41:38 crc kubenswrapper[4632]: E0313 12:41:38.053086 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:41:51 crc kubenswrapper[4632]: I0313 12:41:51.044420 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:41:51 crc kubenswrapper[4632]: E0313 12:41:51.045339 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.163323 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-jp4lw"] Mar 13 12:42:00 crc kubenswrapper[4632]: E0313 12:42:00.164148 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58d6a88c-498f-4887-998a-c3e3a1a2fef2" containerName="oc" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.164160 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="58d6a88c-498f-4887-998a-c3e3a1a2fef2" containerName="oc" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.164367 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="58d6a88c-498f-4887-998a-c3e3a1a2fef2" containerName="oc" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.165836 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.183055 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jp4lw"] Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.251303 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556762-mjtxh"] Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.252746 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556762-mjtxh" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.258803 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.259462 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.264288 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.279511 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556762-mjtxh"] Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.281767 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8fhsr\" (UniqueName: \"kubernetes.io/projected/0325e4fd-6765-4752-afb6-831414e3e532-kube-api-access-8fhsr\") pod \"community-operators-jp4lw\" (UID: \"0325e4fd-6765-4752-afb6-831414e3e532\") " pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.282002 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0325e4fd-6765-4752-afb6-831414e3e532-utilities\") pod \"community-operators-jp4lw\" (UID: \"0325e4fd-6765-4752-afb6-831414e3e532\") " pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.282031 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0325e4fd-6765-4752-afb6-831414e3e532-catalog-content\") pod \"community-operators-jp4lw\" (UID: \"0325e4fd-6765-4752-afb6-831414e3e532\") " pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.384025 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2xcrm\" (UniqueName: \"kubernetes.io/projected/cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae-kube-api-access-2xcrm\") pod \"auto-csr-approver-29556762-mjtxh\" (UID: \"cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae\") " pod="openshift-infra/auto-csr-approver-29556762-mjtxh" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.384362 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0325e4fd-6765-4752-afb6-831414e3e532-utilities\") pod \"community-operators-jp4lw\" (UID: \"0325e4fd-6765-4752-afb6-831414e3e532\") " pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.384384 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0325e4fd-6765-4752-afb6-831414e3e532-catalog-content\") pod \"community-operators-jp4lw\" (UID: \"0325e4fd-6765-4752-afb6-831414e3e532\") " pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.384451 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8fhsr\" (UniqueName: \"kubernetes.io/projected/0325e4fd-6765-4752-afb6-831414e3e532-kube-api-access-8fhsr\") pod \"community-operators-jp4lw\" (UID: \"0325e4fd-6765-4752-afb6-831414e3e532\") " pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.387558 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0325e4fd-6765-4752-afb6-831414e3e532-utilities\") pod \"community-operators-jp4lw\" (UID: \"0325e4fd-6765-4752-afb6-831414e3e532\") " pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.387808 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0325e4fd-6765-4752-afb6-831414e3e532-catalog-content\") pod \"community-operators-jp4lw\" (UID: \"0325e4fd-6765-4752-afb6-831414e3e532\") " pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.410218 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8fhsr\" (UniqueName: \"kubernetes.io/projected/0325e4fd-6765-4752-afb6-831414e3e532-kube-api-access-8fhsr\") pod \"community-operators-jp4lw\" (UID: \"0325e4fd-6765-4752-afb6-831414e3e532\") " pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.482199 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.486610 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2xcrm\" (UniqueName: \"kubernetes.io/projected/cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae-kube-api-access-2xcrm\") pod \"auto-csr-approver-29556762-mjtxh\" (UID: \"cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae\") " pod="openshift-infra/auto-csr-approver-29556762-mjtxh" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.514255 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2xcrm\" (UniqueName: \"kubernetes.io/projected/cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae-kube-api-access-2xcrm\") pod \"auto-csr-approver-29556762-mjtxh\" (UID: \"cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae\") " pod="openshift-infra/auto-csr-approver-29556762-mjtxh" Mar 13 12:42:00 crc kubenswrapper[4632]: I0313 12:42:00.573427 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556762-mjtxh" Mar 13 12:42:01 crc kubenswrapper[4632]: I0313 12:42:01.485671 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-jp4lw"] Mar 13 12:42:01 crc kubenswrapper[4632]: I0313 12:42:01.529461 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jp4lw" event={"ID":"0325e4fd-6765-4752-afb6-831414e3e532","Type":"ContainerStarted","Data":"f753646e3f4b18f0420fff84cc10fc721e6ad55cc5ba98ca0dd48214462119c9"} Mar 13 12:42:01 crc kubenswrapper[4632]: I0313 12:42:01.564567 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556762-mjtxh"] Mar 13 12:42:01 crc kubenswrapper[4632]: W0313 12:42:01.565562 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf13c17d_1ea6_4a0e_bfbd_e3bfc8d453ae.slice/crio-922c1997c080dd1be09c651a5cfc9f28a4d04ab91c7bb78a448553faa0d90e89 WatchSource:0}: Error finding container 922c1997c080dd1be09c651a5cfc9f28a4d04ab91c7bb78a448553faa0d90e89: Status 404 returned error can't find the container with id 922c1997c080dd1be09c651a5cfc9f28a4d04ab91c7bb78a448553faa0d90e89 Mar 13 12:42:02 crc kubenswrapper[4632]: I0313 12:42:02.539862 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556762-mjtxh" event={"ID":"cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae","Type":"ContainerStarted","Data":"922c1997c080dd1be09c651a5cfc9f28a4d04ab91c7bb78a448553faa0d90e89"} Mar 13 12:42:02 crc kubenswrapper[4632]: I0313 12:42:02.542275 4632 generic.go:334] "Generic (PLEG): container finished" podID="0325e4fd-6765-4752-afb6-831414e3e532" containerID="c2b08b03a5ca7bde67bc2372bc69a87126ed72cacb93a81dca4fa17a497ed3c6" exitCode=0 Mar 13 12:42:02 crc kubenswrapper[4632]: I0313 12:42:02.542333 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jp4lw" event={"ID":"0325e4fd-6765-4752-afb6-831414e3e532","Type":"ContainerDied","Data":"c2b08b03a5ca7bde67bc2372bc69a87126ed72cacb93a81dca4fa17a497ed3c6"} Mar 13 12:42:03 crc kubenswrapper[4632]: I0313 12:42:03.554044 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556762-mjtxh" event={"ID":"cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae","Type":"ContainerStarted","Data":"8dfadb29bc36e882b8d8ebf6016fec294107233b5f8602de74595b7d612d371c"} Mar 13 12:42:03 crc kubenswrapper[4632]: I0313 12:42:03.557274 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jp4lw" event={"ID":"0325e4fd-6765-4752-afb6-831414e3e532","Type":"ContainerStarted","Data":"e28102d9e0189fa624d8379759716e19dc6f79055f8b59b2f5724f1145bcc559"} Mar 13 12:42:03 crc kubenswrapper[4632]: I0313 12:42:03.580256 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556762-mjtxh" podStartSLOduration=2.230684517 podStartE2EDuration="3.580234709s" podCreationTimestamp="2026-03-13 12:42:00 +0000 UTC" firstStartedPulling="2026-03-13 12:42:01.567637129 +0000 UTC m=+9495.590167262" lastFinishedPulling="2026-03-13 12:42:02.917187321 +0000 UTC m=+9496.939717454" observedRunningTime="2026-03-13 12:42:03.568574223 +0000 UTC m=+9497.591104366" watchObservedRunningTime="2026-03-13 12:42:03.580234709 +0000 UTC m=+9497.602764842" Mar 13 12:42:05 crc kubenswrapper[4632]: I0313 12:42:05.043854 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:42:05 crc kubenswrapper[4632]: E0313 12:42:05.044640 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:42:05 crc kubenswrapper[4632]: I0313 12:42:05.579898 4632 generic.go:334] "Generic (PLEG): container finished" podID="0325e4fd-6765-4752-afb6-831414e3e532" containerID="e28102d9e0189fa624d8379759716e19dc6f79055f8b59b2f5724f1145bcc559" exitCode=0 Mar 13 12:42:05 crc kubenswrapper[4632]: I0313 12:42:05.580005 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jp4lw" event={"ID":"0325e4fd-6765-4752-afb6-831414e3e532","Type":"ContainerDied","Data":"e28102d9e0189fa624d8379759716e19dc6f79055f8b59b2f5724f1145bcc559"} Mar 13 12:42:05 crc kubenswrapper[4632]: I0313 12:42:05.582164 4632 generic.go:334] "Generic (PLEG): container finished" podID="cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae" containerID="8dfadb29bc36e882b8d8ebf6016fec294107233b5f8602de74595b7d612d371c" exitCode=0 Mar 13 12:42:05 crc kubenswrapper[4632]: I0313 12:42:05.582201 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556762-mjtxh" event={"ID":"cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae","Type":"ContainerDied","Data":"8dfadb29bc36e882b8d8ebf6016fec294107233b5f8602de74595b7d612d371c"} Mar 13 12:42:06 crc kubenswrapper[4632]: I0313 12:42:06.592716 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jp4lw" event={"ID":"0325e4fd-6765-4752-afb6-831414e3e532","Type":"ContainerStarted","Data":"6f6a15708a6bdac7d0aeb9980066237cfba1941652f86fc3be3ae158e72127eb"} Mar 13 12:42:06 crc kubenswrapper[4632]: I0313 12:42:06.616635 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-jp4lw" podStartSLOduration=3.13655522 podStartE2EDuration="6.616615533s" podCreationTimestamp="2026-03-13 12:42:00 +0000 UTC" firstStartedPulling="2026-03-13 12:42:02.544443411 +0000 UTC m=+9496.566973544" lastFinishedPulling="2026-03-13 12:42:06.024503724 +0000 UTC m=+9500.047033857" observedRunningTime="2026-03-13 12:42:06.614507372 +0000 UTC m=+9500.637037515" watchObservedRunningTime="2026-03-13 12:42:06.616615533 +0000 UTC m=+9500.639145666" Mar 13 12:42:07 crc kubenswrapper[4632]: I0313 12:42:07.211307 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556762-mjtxh" Mar 13 12:42:07 crc kubenswrapper[4632]: I0313 12:42:07.325190 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2xcrm\" (UniqueName: \"kubernetes.io/projected/cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae-kube-api-access-2xcrm\") pod \"cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae\" (UID: \"cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae\") " Mar 13 12:42:07 crc kubenswrapper[4632]: I0313 12:42:07.340108 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae-kube-api-access-2xcrm" (OuterVolumeSpecName: "kube-api-access-2xcrm") pod "cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae" (UID: "cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae"). InnerVolumeSpecName "kube-api-access-2xcrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:42:07 crc kubenswrapper[4632]: I0313 12:42:07.427355 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2xcrm\" (UniqueName: \"kubernetes.io/projected/cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae-kube-api-access-2xcrm\") on node \"crc\" DevicePath \"\"" Mar 13 12:42:07 crc kubenswrapper[4632]: I0313 12:42:07.603213 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556762-mjtxh" event={"ID":"cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae","Type":"ContainerDied","Data":"922c1997c080dd1be09c651a5cfc9f28a4d04ab91c7bb78a448553faa0d90e89"} Mar 13 12:42:07 crc kubenswrapper[4632]: I0313 12:42:07.604169 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="922c1997c080dd1be09c651a5cfc9f28a4d04ab91c7bb78a448553faa0d90e89" Mar 13 12:42:07 crc kubenswrapper[4632]: I0313 12:42:07.603302 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556762-mjtxh" Mar 13 12:42:07 crc kubenswrapper[4632]: I0313 12:42:07.693208 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556756-4d96k"] Mar 13 12:42:07 crc kubenswrapper[4632]: I0313 12:42:07.702155 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556756-4d96k"] Mar 13 12:42:08 crc kubenswrapper[4632]: I0313 12:42:08.062539 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bab5d92-9e39-4d06-98ae-8b9b50d50214" path="/var/lib/kubelet/pods/4bab5d92-9e39-4d06-98ae-8b9b50d50214/volumes" Mar 13 12:42:10 crc kubenswrapper[4632]: I0313 12:42:10.483502 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:10 crc kubenswrapper[4632]: I0313 12:42:10.483824 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:11 crc kubenswrapper[4632]: I0313 12:42:11.542532 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-jp4lw" podUID="0325e4fd-6765-4752-afb6-831414e3e532" containerName="registry-server" probeResult="failure" output=< Mar 13 12:42:11 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:42:11 crc kubenswrapper[4632]: > Mar 13 12:42:19 crc kubenswrapper[4632]: I0313 12:42:19.045664 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:42:19 crc kubenswrapper[4632]: E0313 12:42:19.046593 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:42:20 crc kubenswrapper[4632]: I0313 12:42:20.577527 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:20 crc kubenswrapper[4632]: I0313 12:42:20.635459 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:20 crc kubenswrapper[4632]: I0313 12:42:20.820955 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jp4lw"] Mar 13 12:42:21 crc kubenswrapper[4632]: I0313 12:42:21.744358 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-jp4lw" podUID="0325e4fd-6765-4752-afb6-831414e3e532" containerName="registry-server" containerID="cri-o://6f6a15708a6bdac7d0aeb9980066237cfba1941652f86fc3be3ae158e72127eb" gracePeriod=2 Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.600191 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.696437 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0325e4fd-6765-4752-afb6-831414e3e532-utilities\") pod \"0325e4fd-6765-4752-afb6-831414e3e532\" (UID: \"0325e4fd-6765-4752-afb6-831414e3e532\") " Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.696510 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8fhsr\" (UniqueName: \"kubernetes.io/projected/0325e4fd-6765-4752-afb6-831414e3e532-kube-api-access-8fhsr\") pod \"0325e4fd-6765-4752-afb6-831414e3e532\" (UID: \"0325e4fd-6765-4752-afb6-831414e3e532\") " Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.696821 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0325e4fd-6765-4752-afb6-831414e3e532-catalog-content\") pod \"0325e4fd-6765-4752-afb6-831414e3e532\" (UID: \"0325e4fd-6765-4752-afb6-831414e3e532\") " Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.698112 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0325e4fd-6765-4752-afb6-831414e3e532-utilities" (OuterVolumeSpecName: "utilities") pod "0325e4fd-6765-4752-afb6-831414e3e532" (UID: "0325e4fd-6765-4752-afb6-831414e3e532"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.738079 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0325e4fd-6765-4752-afb6-831414e3e532-kube-api-access-8fhsr" (OuterVolumeSpecName: "kube-api-access-8fhsr") pod "0325e4fd-6765-4752-afb6-831414e3e532" (UID: "0325e4fd-6765-4752-afb6-831414e3e532"). InnerVolumeSpecName "kube-api-access-8fhsr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.757396 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0325e4fd-6765-4752-afb6-831414e3e532-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0325e4fd-6765-4752-afb6-831414e3e532" (UID: "0325e4fd-6765-4752-afb6-831414e3e532"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.761341 4632 generic.go:334] "Generic (PLEG): container finished" podID="0325e4fd-6765-4752-afb6-831414e3e532" containerID="6f6a15708a6bdac7d0aeb9980066237cfba1941652f86fc3be3ae158e72127eb" exitCode=0 Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.761364 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-jp4lw" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.761387 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jp4lw" event={"ID":"0325e4fd-6765-4752-afb6-831414e3e532","Type":"ContainerDied","Data":"6f6a15708a6bdac7d0aeb9980066237cfba1941652f86fc3be3ae158e72127eb"} Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.761416 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-jp4lw" event={"ID":"0325e4fd-6765-4752-afb6-831414e3e532","Type":"ContainerDied","Data":"f753646e3f4b18f0420fff84cc10fc721e6ad55cc5ba98ca0dd48214462119c9"} Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.761436 4632 scope.go:117] "RemoveContainer" containerID="6f6a15708a6bdac7d0aeb9980066237cfba1941652f86fc3be3ae158e72127eb" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.798635 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0325e4fd-6765-4752-afb6-831414e3e532-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.802236 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0325e4fd-6765-4752-afb6-831414e3e532-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.802322 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8fhsr\" (UniqueName: \"kubernetes.io/projected/0325e4fd-6765-4752-afb6-831414e3e532-kube-api-access-8fhsr\") on node \"crc\" DevicePath \"\"" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.802813 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-jp4lw"] Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.812222 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-jp4lw"] Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.822174 4632 scope.go:117] "RemoveContainer" containerID="e28102d9e0189fa624d8379759716e19dc6f79055f8b59b2f5724f1145bcc559" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.849526 4632 scope.go:117] "RemoveContainer" containerID="c2b08b03a5ca7bde67bc2372bc69a87126ed72cacb93a81dca4fa17a497ed3c6" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.928755 4632 scope.go:117] "RemoveContainer" containerID="6f6a15708a6bdac7d0aeb9980066237cfba1941652f86fc3be3ae158e72127eb" Mar 13 12:42:22 crc kubenswrapper[4632]: E0313 12:42:22.933051 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f6a15708a6bdac7d0aeb9980066237cfba1941652f86fc3be3ae158e72127eb\": container with ID starting with 6f6a15708a6bdac7d0aeb9980066237cfba1941652f86fc3be3ae158e72127eb not found: ID does not exist" containerID="6f6a15708a6bdac7d0aeb9980066237cfba1941652f86fc3be3ae158e72127eb" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.933098 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f6a15708a6bdac7d0aeb9980066237cfba1941652f86fc3be3ae158e72127eb"} err="failed to get container status \"6f6a15708a6bdac7d0aeb9980066237cfba1941652f86fc3be3ae158e72127eb\": rpc error: code = NotFound desc = could not find container \"6f6a15708a6bdac7d0aeb9980066237cfba1941652f86fc3be3ae158e72127eb\": container with ID starting with 6f6a15708a6bdac7d0aeb9980066237cfba1941652f86fc3be3ae158e72127eb not found: ID does not exist" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.933122 4632 scope.go:117] "RemoveContainer" containerID="e28102d9e0189fa624d8379759716e19dc6f79055f8b59b2f5724f1145bcc559" Mar 13 12:42:22 crc kubenswrapper[4632]: E0313 12:42:22.933730 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e28102d9e0189fa624d8379759716e19dc6f79055f8b59b2f5724f1145bcc559\": container with ID starting with e28102d9e0189fa624d8379759716e19dc6f79055f8b59b2f5724f1145bcc559 not found: ID does not exist" containerID="e28102d9e0189fa624d8379759716e19dc6f79055f8b59b2f5724f1145bcc559" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.933753 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e28102d9e0189fa624d8379759716e19dc6f79055f8b59b2f5724f1145bcc559"} err="failed to get container status \"e28102d9e0189fa624d8379759716e19dc6f79055f8b59b2f5724f1145bcc559\": rpc error: code = NotFound desc = could not find container \"e28102d9e0189fa624d8379759716e19dc6f79055f8b59b2f5724f1145bcc559\": container with ID starting with e28102d9e0189fa624d8379759716e19dc6f79055f8b59b2f5724f1145bcc559 not found: ID does not exist" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.933767 4632 scope.go:117] "RemoveContainer" containerID="c2b08b03a5ca7bde67bc2372bc69a87126ed72cacb93a81dca4fa17a497ed3c6" Mar 13 12:42:22 crc kubenswrapper[4632]: E0313 12:42:22.934425 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2b08b03a5ca7bde67bc2372bc69a87126ed72cacb93a81dca4fa17a497ed3c6\": container with ID starting with c2b08b03a5ca7bde67bc2372bc69a87126ed72cacb93a81dca4fa17a497ed3c6 not found: ID does not exist" containerID="c2b08b03a5ca7bde67bc2372bc69a87126ed72cacb93a81dca4fa17a497ed3c6" Mar 13 12:42:22 crc kubenswrapper[4632]: I0313 12:42:22.934453 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2b08b03a5ca7bde67bc2372bc69a87126ed72cacb93a81dca4fa17a497ed3c6"} err="failed to get container status \"c2b08b03a5ca7bde67bc2372bc69a87126ed72cacb93a81dca4fa17a497ed3c6\": rpc error: code = NotFound desc = could not find container \"c2b08b03a5ca7bde67bc2372bc69a87126ed72cacb93a81dca4fa17a497ed3c6\": container with ID starting with c2b08b03a5ca7bde67bc2372bc69a87126ed72cacb93a81dca4fa17a497ed3c6 not found: ID does not exist" Mar 13 12:42:24 crc kubenswrapper[4632]: I0313 12:42:24.055158 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0325e4fd-6765-4752-afb6-831414e3e532" path="/var/lib/kubelet/pods/0325e4fd-6765-4752-afb6-831414e3e532/volumes" Mar 13 12:42:28 crc kubenswrapper[4632]: I0313 12:42:28.472101 4632 scope.go:117] "RemoveContainer" containerID="a8e65f824fd306a694713a170d4b213522c9b5fbd2a9bb06608f463371bdb733" Mar 13 12:42:33 crc kubenswrapper[4632]: I0313 12:42:33.043920 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:42:33 crc kubenswrapper[4632]: E0313 12:42:33.044868 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:42:44 crc kubenswrapper[4632]: I0313 12:42:44.044911 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:42:44 crc kubenswrapper[4632]: I0313 12:42:44.969521 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"ac0aa587db0bc6f14a810b1c0a407933497eafb76c5051481d9814592d0380b3"} Mar 13 12:42:56 crc kubenswrapper[4632]: I0313 12:42:56.073025 4632 generic.go:334] "Generic (PLEG): container finished" podID="611401cc-04fe-4276-82fa-a896182802d4" containerID="e5092a16adcd02c327c069b34afdd26aca8018f63ed747e3778a6c696a0e6a3c" exitCode=0 Mar 13 12:42:56 crc kubenswrapper[4632]: I0313 12:42:56.073605 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" event={"ID":"611401cc-04fe-4276-82fa-a896182802d4","Type":"ContainerDied","Data":"e5092a16adcd02c327c069b34afdd26aca8018f63ed747e3778a6c696a0e6a3c"} Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.864441 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.933700 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-openstack-config-secret\") pod \"611401cc-04fe-4276-82fa-a896182802d4\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.934035 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"611401cc-04fe-4276-82fa-a896182802d4\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.934066 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/611401cc-04fe-4276-82fa-a896182802d4-test-operator-ephemeral-workdir\") pod \"611401cc-04fe-4276-82fa-a896182802d4\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.934091 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-ssh-key\") pod \"611401cc-04fe-4276-82fa-a896182802d4\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.934866 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/611401cc-04fe-4276-82fa-a896182802d4-config-data\") pod \"611401cc-04fe-4276-82fa-a896182802d4\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.934921 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-ca-certs\") pod \"611401cc-04fe-4276-82fa-a896182802d4\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.935099 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/611401cc-04fe-4276-82fa-a896182802d4-openstack-config\") pod \"611401cc-04fe-4276-82fa-a896182802d4\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.935175 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrwf2\" (UniqueName: \"kubernetes.io/projected/611401cc-04fe-4276-82fa-a896182802d4-kube-api-access-hrwf2\") pod \"611401cc-04fe-4276-82fa-a896182802d4\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.935203 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/611401cc-04fe-4276-82fa-a896182802d4-test-operator-ephemeral-temporary\") pod \"611401cc-04fe-4276-82fa-a896182802d4\" (UID: \"611401cc-04fe-4276-82fa-a896182802d4\") " Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.935675 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/611401cc-04fe-4276-82fa-a896182802d4-config-data" (OuterVolumeSpecName: "config-data") pod "611401cc-04fe-4276-82fa-a896182802d4" (UID: "611401cc-04fe-4276-82fa-a896182802d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.938069 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/611401cc-04fe-4276-82fa-a896182802d4-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "611401cc-04fe-4276-82fa-a896182802d4" (UID: "611401cc-04fe-4276-82fa-a896182802d4"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.952470 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/611401cc-04fe-4276-82fa-a896182802d4-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "611401cc-04fe-4276-82fa-a896182802d4" (UID: "611401cc-04fe-4276-82fa-a896182802d4"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.952762 4632 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/611401cc-04fe-4276-82fa-a896182802d4-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.952780 4632 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/611401cc-04fe-4276-82fa-a896182802d4-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.952793 4632 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/611401cc-04fe-4276-82fa-a896182802d4-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.954099 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "611401cc-04fe-4276-82fa-a896182802d4" (UID: "611401cc-04fe-4276-82fa-a896182802d4"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 12:42:57 crc kubenswrapper[4632]: I0313 12:42:57.995231 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/611401cc-04fe-4276-82fa-a896182802d4-kube-api-access-hrwf2" (OuterVolumeSpecName: "kube-api-access-hrwf2") pod "611401cc-04fe-4276-82fa-a896182802d4" (UID: "611401cc-04fe-4276-82fa-a896182802d4"). InnerVolumeSpecName "kube-api-access-hrwf2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.075322 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/611401cc-04fe-4276-82fa-a896182802d4-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "611401cc-04fe-4276-82fa-a896182802d4" (UID: "611401cc-04fe-4276-82fa-a896182802d4"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.075543 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrwf2\" (UniqueName: \"kubernetes.io/projected/611401cc-04fe-4276-82fa-a896182802d4-kube-api-access-hrwf2\") on node \"crc\" DevicePath \"\"" Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.077908 4632 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.083226 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "611401cc-04fe-4276-82fa-a896182802d4" (UID: "611401cc-04fe-4276-82fa-a896182802d4"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.089734 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "611401cc-04fe-4276-82fa-a896182802d4" (UID: "611401cc-04fe-4276-82fa-a896182802d4"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.097596 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.099023 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-thread-testing" event={"ID":"611401cc-04fe-4276-82fa-a896182802d4","Type":"ContainerDied","Data":"95b1b1d6a519cb7b9bfef154cebb6e4b73104a8706f52af49a8997ffa20ebd91"} Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.099053 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95b1b1d6a519cb7b9bfef154cebb6e4b73104a8706f52af49a8997ffa20ebd91" Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.100785 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "611401cc-04fe-4276-82fa-a896182802d4" (UID: "611401cc-04fe-4276-82fa-a896182802d4"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.115840 4632 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.179397 4632 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-ca-certs\") on node \"crc\" DevicePath \"\"" Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.179500 4632 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/611401cc-04fe-4276-82fa-a896182802d4-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.179514 4632 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.179523 4632 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Mar 13 12:42:58 crc kubenswrapper[4632]: I0313 12:42:58.179534 4632 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/611401cc-04fe-4276-82fa-a896182802d4-ssh-key\") on node \"crc\" DevicePath \"\"" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.499952 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 13 12:43:00 crc kubenswrapper[4632]: E0313 12:43:00.500845 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0325e4fd-6765-4752-afb6-831414e3e532" containerName="extract-content" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.500858 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="0325e4fd-6765-4752-afb6-831414e3e532" containerName="extract-content" Mar 13 12:43:00 crc kubenswrapper[4632]: E0313 12:43:00.500877 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0325e4fd-6765-4752-afb6-831414e3e532" containerName="registry-server" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.500883 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="0325e4fd-6765-4752-afb6-831414e3e532" containerName="registry-server" Mar 13 12:43:00 crc kubenswrapper[4632]: E0313 12:43:00.500913 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0325e4fd-6765-4752-afb6-831414e3e532" containerName="extract-utilities" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.500920 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="0325e4fd-6765-4752-afb6-831414e3e532" containerName="extract-utilities" Mar 13 12:43:00 crc kubenswrapper[4632]: E0313 12:43:00.500951 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae" containerName="oc" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.500959 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae" containerName="oc" Mar 13 12:43:00 crc kubenswrapper[4632]: E0313 12:43:00.500970 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="611401cc-04fe-4276-82fa-a896182802d4" containerName="tempest-tests-tempest-tests-runner" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.500977 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="611401cc-04fe-4276-82fa-a896182802d4" containerName="tempest-tests-tempest-tests-runner" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.501138 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae" containerName="oc" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.501158 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="0325e4fd-6765-4752-afb6-831414e3e532" containerName="registry-server" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.501171 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="611401cc-04fe-4276-82fa-a896182802d4" containerName="tempest-tests-tempest-tests-runner" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.501801 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.507963 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-9w9qk" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.521323 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.622798 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c4836490-7b24-4245-bf50-7d590576f21e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.622952 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqlzz\" (UniqueName: \"kubernetes.io/projected/c4836490-7b24-4245-bf50-7d590576f21e-kube-api-access-vqlzz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c4836490-7b24-4245-bf50-7d590576f21e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.724951 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqlzz\" (UniqueName: \"kubernetes.io/projected/c4836490-7b24-4245-bf50-7d590576f21e-kube-api-access-vqlzz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c4836490-7b24-4245-bf50-7d590576f21e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.725094 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c4836490-7b24-4245-bf50-7d590576f21e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.728741 4632 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c4836490-7b24-4245-bf50-7d590576f21e\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.755857 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqlzz\" (UniqueName: \"kubernetes.io/projected/c4836490-7b24-4245-bf50-7d590576f21e-kube-api-access-vqlzz\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c4836490-7b24-4245-bf50-7d590576f21e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.764508 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"c4836490-7b24-4245-bf50-7d590576f21e\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 12:43:00 crc kubenswrapper[4632]: I0313 12:43:00.823819 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 12:43:01 crc kubenswrapper[4632]: I0313 12:43:01.369676 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 12:43:01 crc kubenswrapper[4632]: I0313 12:43:01.371253 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 13 12:43:02 crc kubenswrapper[4632]: I0313 12:43:02.136230 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"c4836490-7b24-4245-bf50-7d590576f21e","Type":"ContainerStarted","Data":"fcb3b56a670015e8f1282bd22f7ebb710e98ea5f9b1d93e0919fac38dd6d6288"} Mar 13 12:43:03 crc kubenswrapper[4632]: I0313 12:43:03.149389 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"c4836490-7b24-4245-bf50-7d590576f21e","Type":"ContainerStarted","Data":"467eab617c540776869c8f7c1778fe8bb6f1a0f79a6bdc7565116731eee00756"} Mar 13 12:43:03 crc kubenswrapper[4632]: I0313 12:43:03.172795 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.733172398 podStartE2EDuration="3.172779618s" podCreationTimestamp="2026-03-13 12:43:00 +0000 UTC" firstStartedPulling="2026-03-13 12:43:01.368746762 +0000 UTC m=+9555.391276895" lastFinishedPulling="2026-03-13 12:43:02.808353982 +0000 UTC m=+9556.830884115" observedRunningTime="2026-03-13 12:43:03.171060806 +0000 UTC m=+9557.193590939" watchObservedRunningTime="2026-03-13 12:43:03.172779618 +0000 UTC m=+9557.195309751" Mar 13 12:43:50 crc kubenswrapper[4632]: I0313 12:43:50.679199 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jfn52/must-gather-9gqfn"] Mar 13 12:43:50 crc kubenswrapper[4632]: I0313 12:43:50.682374 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/must-gather-9gqfn" Mar 13 12:43:50 crc kubenswrapper[4632]: I0313 12:43:50.684775 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jfn52"/"openshift-service-ca.crt" Mar 13 12:43:50 crc kubenswrapper[4632]: I0313 12:43:50.687060 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-jfn52"/"kube-root-ca.crt" Mar 13 12:43:50 crc kubenswrapper[4632]: I0313 12:43:50.690539 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-jfn52"/"default-dockercfg-74jfd" Mar 13 12:43:50 crc kubenswrapper[4632]: I0313 12:43:50.711215 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/252f97d9-adeb-4cce-858d-eb0bdb151871-must-gather-output\") pod \"must-gather-9gqfn\" (UID: \"252f97d9-adeb-4cce-858d-eb0bdb151871\") " pod="openshift-must-gather-jfn52/must-gather-9gqfn" Mar 13 12:43:50 crc kubenswrapper[4632]: I0313 12:43:50.711505 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwxgx\" (UniqueName: \"kubernetes.io/projected/252f97d9-adeb-4cce-858d-eb0bdb151871-kube-api-access-kwxgx\") pod \"must-gather-9gqfn\" (UID: \"252f97d9-adeb-4cce-858d-eb0bdb151871\") " pod="openshift-must-gather-jfn52/must-gather-9gqfn" Mar 13 12:43:50 crc kubenswrapper[4632]: I0313 12:43:50.813554 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/252f97d9-adeb-4cce-858d-eb0bdb151871-must-gather-output\") pod \"must-gather-9gqfn\" (UID: \"252f97d9-adeb-4cce-858d-eb0bdb151871\") " pod="openshift-must-gather-jfn52/must-gather-9gqfn" Mar 13 12:43:50 crc kubenswrapper[4632]: I0313 12:43:50.813671 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwxgx\" (UniqueName: \"kubernetes.io/projected/252f97d9-adeb-4cce-858d-eb0bdb151871-kube-api-access-kwxgx\") pod \"must-gather-9gqfn\" (UID: \"252f97d9-adeb-4cce-858d-eb0bdb151871\") " pod="openshift-must-gather-jfn52/must-gather-9gqfn" Mar 13 12:43:50 crc kubenswrapper[4632]: I0313 12:43:50.814239 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/252f97d9-adeb-4cce-858d-eb0bdb151871-must-gather-output\") pod \"must-gather-9gqfn\" (UID: \"252f97d9-adeb-4cce-858d-eb0bdb151871\") " pod="openshift-must-gather-jfn52/must-gather-9gqfn" Mar 13 12:43:50 crc kubenswrapper[4632]: I0313 12:43:50.835635 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwxgx\" (UniqueName: \"kubernetes.io/projected/252f97d9-adeb-4cce-858d-eb0bdb151871-kube-api-access-kwxgx\") pod \"must-gather-9gqfn\" (UID: \"252f97d9-adeb-4cce-858d-eb0bdb151871\") " pod="openshift-must-gather-jfn52/must-gather-9gqfn" Mar 13 12:43:50 crc kubenswrapper[4632]: I0313 12:43:50.891172 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jfn52/must-gather-9gqfn"] Mar 13 12:43:51 crc kubenswrapper[4632]: I0313 12:43:51.006463 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/must-gather-9gqfn" Mar 13 12:43:51 crc kubenswrapper[4632]: I0313 12:43:51.491372 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-jfn52/must-gather-9gqfn"] Mar 13 12:43:51 crc kubenswrapper[4632]: I0313 12:43:51.638681 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfn52/must-gather-9gqfn" event={"ID":"252f97d9-adeb-4cce-858d-eb0bdb151871","Type":"ContainerStarted","Data":"eb201c9262110becebc0a17449ee45af81bfd07a68b8b76aec66f1966a55fe23"} Mar 13 12:44:00 crc kubenswrapper[4632]: I0313 12:44:00.209573 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556764-9lxh8"] Mar 13 12:44:00 crc kubenswrapper[4632]: I0313 12:44:00.211834 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556764-9lxh8" Mar 13 12:44:00 crc kubenswrapper[4632]: I0313 12:44:00.215902 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:44:00 crc kubenswrapper[4632]: I0313 12:44:00.216132 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:44:00 crc kubenswrapper[4632]: I0313 12:44:00.220396 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556764-9lxh8"] Mar 13 12:44:00 crc kubenswrapper[4632]: I0313 12:44:00.222964 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:44:00 crc kubenswrapper[4632]: I0313 12:44:00.338018 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rqnq\" (UniqueName: \"kubernetes.io/projected/437f55ff-c573-4944-a680-6ac2d168cb0f-kube-api-access-9rqnq\") pod \"auto-csr-approver-29556764-9lxh8\" (UID: \"437f55ff-c573-4944-a680-6ac2d168cb0f\") " pod="openshift-infra/auto-csr-approver-29556764-9lxh8" Mar 13 12:44:00 crc kubenswrapper[4632]: I0313 12:44:00.440056 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rqnq\" (UniqueName: \"kubernetes.io/projected/437f55ff-c573-4944-a680-6ac2d168cb0f-kube-api-access-9rqnq\") pod \"auto-csr-approver-29556764-9lxh8\" (UID: \"437f55ff-c573-4944-a680-6ac2d168cb0f\") " pod="openshift-infra/auto-csr-approver-29556764-9lxh8" Mar 13 12:44:00 crc kubenswrapper[4632]: I0313 12:44:00.463604 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rqnq\" (UniqueName: \"kubernetes.io/projected/437f55ff-c573-4944-a680-6ac2d168cb0f-kube-api-access-9rqnq\") pod \"auto-csr-approver-29556764-9lxh8\" (UID: \"437f55ff-c573-4944-a680-6ac2d168cb0f\") " pod="openshift-infra/auto-csr-approver-29556764-9lxh8" Mar 13 12:44:00 crc kubenswrapper[4632]: I0313 12:44:00.539967 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556764-9lxh8" Mar 13 12:44:01 crc kubenswrapper[4632]: I0313 12:44:01.686130 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556764-9lxh8"] Mar 13 12:44:01 crc kubenswrapper[4632]: W0313 12:44:01.693872 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod437f55ff_c573_4944_a680_6ac2d168cb0f.slice/crio-0a3427c255939c243894f261514eaf59b4474ffa7b09ba5549c6a6b3a1fad4af WatchSource:0}: Error finding container 0a3427c255939c243894f261514eaf59b4474ffa7b09ba5549c6a6b3a1fad4af: Status 404 returned error can't find the container with id 0a3427c255939c243894f261514eaf59b4474ffa7b09ba5549c6a6b3a1fad4af Mar 13 12:44:01 crc kubenswrapper[4632]: I0313 12:44:01.749096 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfn52/must-gather-9gqfn" event={"ID":"252f97d9-adeb-4cce-858d-eb0bdb151871","Type":"ContainerStarted","Data":"8c2cc6936b125830f781b6c13d49ab8294ed023e1773b93feeb9bf3339c7d42f"} Mar 13 12:44:01 crc kubenswrapper[4632]: I0313 12:44:01.749169 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfn52/must-gather-9gqfn" event={"ID":"252f97d9-adeb-4cce-858d-eb0bdb151871","Type":"ContainerStarted","Data":"ce4c408f0cb872f87b4909caa3af1013273fb7803c492e30e5b30666a913955d"} Mar 13 12:44:01 crc kubenswrapper[4632]: I0313 12:44:01.750932 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556764-9lxh8" event={"ID":"437f55ff-c573-4944-a680-6ac2d168cb0f","Type":"ContainerStarted","Data":"0a3427c255939c243894f261514eaf59b4474ffa7b09ba5549c6a6b3a1fad4af"} Mar 13 12:44:01 crc kubenswrapper[4632]: I0313 12:44:01.770776 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jfn52/must-gather-9gqfn" podStartSLOduration=2.198756595 podStartE2EDuration="11.770759827s" podCreationTimestamp="2026-03-13 12:43:50 +0000 UTC" firstStartedPulling="2026-03-13 12:43:51.498559226 +0000 UTC m=+9605.521089359" lastFinishedPulling="2026-03-13 12:44:01.070562458 +0000 UTC m=+9615.093092591" observedRunningTime="2026-03-13 12:44:01.766463182 +0000 UTC m=+9615.788993335" watchObservedRunningTime="2026-03-13 12:44:01.770759827 +0000 UTC m=+9615.793289960" Mar 13 12:44:04 crc kubenswrapper[4632]: I0313 12:44:04.799672 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556764-9lxh8" event={"ID":"437f55ff-c573-4944-a680-6ac2d168cb0f","Type":"ContainerStarted","Data":"aa02f726269acb1a95d7a68005cbfbbe4f481bb481f4612470d155ee5bde6649"} Mar 13 12:44:04 crc kubenswrapper[4632]: I0313 12:44:04.827040 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556764-9lxh8" podStartSLOduration=3.912939635 podStartE2EDuration="4.827016839s" podCreationTimestamp="2026-03-13 12:44:00 +0000 UTC" firstStartedPulling="2026-03-13 12:44:01.696129017 +0000 UTC m=+9615.718659150" lastFinishedPulling="2026-03-13 12:44:02.610206221 +0000 UTC m=+9616.632736354" observedRunningTime="2026-03-13 12:44:04.819819292 +0000 UTC m=+9618.842349465" watchObservedRunningTime="2026-03-13 12:44:04.827016839 +0000 UTC m=+9618.849546982" Mar 13 12:44:05 crc kubenswrapper[4632]: I0313 12:44:05.809923 4632 generic.go:334] "Generic (PLEG): container finished" podID="437f55ff-c573-4944-a680-6ac2d168cb0f" containerID="aa02f726269acb1a95d7a68005cbfbbe4f481bb481f4612470d155ee5bde6649" exitCode=0 Mar 13 12:44:05 crc kubenswrapper[4632]: I0313 12:44:05.810104 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556764-9lxh8" event={"ID":"437f55ff-c573-4944-a680-6ac2d168cb0f","Type":"ContainerDied","Data":"aa02f726269acb1a95d7a68005cbfbbe4f481bb481f4612470d155ee5bde6649"} Mar 13 12:44:07 crc kubenswrapper[4632]: I0313 12:44:07.514327 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556764-9lxh8" Mar 13 12:44:07 crc kubenswrapper[4632]: I0313 12:44:07.686358 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9rqnq\" (UniqueName: \"kubernetes.io/projected/437f55ff-c573-4944-a680-6ac2d168cb0f-kube-api-access-9rqnq\") pod \"437f55ff-c573-4944-a680-6ac2d168cb0f\" (UID: \"437f55ff-c573-4944-a680-6ac2d168cb0f\") " Mar 13 12:44:07 crc kubenswrapper[4632]: I0313 12:44:07.695977 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/437f55ff-c573-4944-a680-6ac2d168cb0f-kube-api-access-9rqnq" (OuterVolumeSpecName: "kube-api-access-9rqnq") pod "437f55ff-c573-4944-a680-6ac2d168cb0f" (UID: "437f55ff-c573-4944-a680-6ac2d168cb0f"). InnerVolumeSpecName "kube-api-access-9rqnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:44:07 crc kubenswrapper[4632]: I0313 12:44:07.789486 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9rqnq\" (UniqueName: \"kubernetes.io/projected/437f55ff-c573-4944-a680-6ac2d168cb0f-kube-api-access-9rqnq\") on node \"crc\" DevicePath \"\"" Mar 13 12:44:07 crc kubenswrapper[4632]: I0313 12:44:07.831092 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556764-9lxh8" event={"ID":"437f55ff-c573-4944-a680-6ac2d168cb0f","Type":"ContainerDied","Data":"0a3427c255939c243894f261514eaf59b4474ffa7b09ba5549c6a6b3a1fad4af"} Mar 13 12:44:07 crc kubenswrapper[4632]: I0313 12:44:07.831357 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a3427c255939c243894f261514eaf59b4474ffa7b09ba5549c6a6b3a1fad4af" Mar 13 12:44:07 crc kubenswrapper[4632]: I0313 12:44:07.831190 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556764-9lxh8" Mar 13 12:44:07 crc kubenswrapper[4632]: I0313 12:44:07.904585 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556758-989sl"] Mar 13 12:44:07 crc kubenswrapper[4632]: I0313 12:44:07.913083 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556758-989sl"] Mar 13 12:44:08 crc kubenswrapper[4632]: I0313 12:44:08.057518 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="564cb4aa-8722-4f69-adeb-16bc8b74bff0" path="/var/lib/kubelet/pods/564cb4aa-8722-4f69-adeb-16bc8b74bff0/volumes" Mar 13 12:44:08 crc kubenswrapper[4632]: I0313 12:44:08.349244 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jfn52/crc-debug-9w6fr"] Mar 13 12:44:08 crc kubenswrapper[4632]: E0313 12:44:08.349729 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="437f55ff-c573-4944-a680-6ac2d168cb0f" containerName="oc" Mar 13 12:44:08 crc kubenswrapper[4632]: I0313 12:44:08.349755 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="437f55ff-c573-4944-a680-6ac2d168cb0f" containerName="oc" Mar 13 12:44:08 crc kubenswrapper[4632]: I0313 12:44:08.350038 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="437f55ff-c573-4944-a680-6ac2d168cb0f" containerName="oc" Mar 13 12:44:08 crc kubenswrapper[4632]: I0313 12:44:08.350884 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/crc-debug-9w6fr" Mar 13 12:44:08 crc kubenswrapper[4632]: I0313 12:44:08.405272 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3ccc7848-e5be-416d-95de-b621b5cc770d-host\") pod \"crc-debug-9w6fr\" (UID: \"3ccc7848-e5be-416d-95de-b621b5cc770d\") " pod="openshift-must-gather-jfn52/crc-debug-9w6fr" Mar 13 12:44:08 crc kubenswrapper[4632]: I0313 12:44:08.405369 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bfjh\" (UniqueName: \"kubernetes.io/projected/3ccc7848-e5be-416d-95de-b621b5cc770d-kube-api-access-7bfjh\") pod \"crc-debug-9w6fr\" (UID: \"3ccc7848-e5be-416d-95de-b621b5cc770d\") " pod="openshift-must-gather-jfn52/crc-debug-9w6fr" Mar 13 12:44:08 crc kubenswrapper[4632]: I0313 12:44:08.506753 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3ccc7848-e5be-416d-95de-b621b5cc770d-host\") pod \"crc-debug-9w6fr\" (UID: \"3ccc7848-e5be-416d-95de-b621b5cc770d\") " pod="openshift-must-gather-jfn52/crc-debug-9w6fr" Mar 13 12:44:08 crc kubenswrapper[4632]: I0313 12:44:08.506880 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bfjh\" (UniqueName: \"kubernetes.io/projected/3ccc7848-e5be-416d-95de-b621b5cc770d-kube-api-access-7bfjh\") pod \"crc-debug-9w6fr\" (UID: \"3ccc7848-e5be-416d-95de-b621b5cc770d\") " pod="openshift-must-gather-jfn52/crc-debug-9w6fr" Mar 13 12:44:08 crc kubenswrapper[4632]: I0313 12:44:08.508169 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3ccc7848-e5be-416d-95de-b621b5cc770d-host\") pod \"crc-debug-9w6fr\" (UID: \"3ccc7848-e5be-416d-95de-b621b5cc770d\") " pod="openshift-must-gather-jfn52/crc-debug-9w6fr" Mar 13 12:44:08 crc kubenswrapper[4632]: I0313 12:44:08.526309 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bfjh\" (UniqueName: \"kubernetes.io/projected/3ccc7848-e5be-416d-95de-b621b5cc770d-kube-api-access-7bfjh\") pod \"crc-debug-9w6fr\" (UID: \"3ccc7848-e5be-416d-95de-b621b5cc770d\") " pod="openshift-must-gather-jfn52/crc-debug-9w6fr" Mar 13 12:44:08 crc kubenswrapper[4632]: I0313 12:44:08.671079 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/crc-debug-9w6fr" Mar 13 12:44:08 crc kubenswrapper[4632]: I0313 12:44:08.842666 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfn52/crc-debug-9w6fr" event={"ID":"3ccc7848-e5be-416d-95de-b621b5cc770d","Type":"ContainerStarted","Data":"b1c855fe5ef0d99083c07be960529478cbb1140696aa5e6315932510eb584567"} Mar 13 12:44:21 crc kubenswrapper[4632]: I0313 12:44:21.978577 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfn52/crc-debug-9w6fr" event={"ID":"3ccc7848-e5be-416d-95de-b621b5cc770d","Type":"ContainerStarted","Data":"ef8f3145eaca72f4029fdc60f2fb5cc46e463f526ce49ed7c2f109c93d6646f5"} Mar 13 12:44:22 crc kubenswrapper[4632]: I0313 12:44:22.003148 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jfn52/crc-debug-9w6fr" podStartSLOduration=1.288398085 podStartE2EDuration="14.003127299s" podCreationTimestamp="2026-03-13 12:44:08 +0000 UTC" firstStartedPulling="2026-03-13 12:44:08.725595785 +0000 UTC m=+9622.748125918" lastFinishedPulling="2026-03-13 12:44:21.440324999 +0000 UTC m=+9635.462855132" observedRunningTime="2026-03-13 12:44:21.99583384 +0000 UTC m=+9636.018363973" watchObservedRunningTime="2026-03-13 12:44:22.003127299 +0000 UTC m=+9636.025657432" Mar 13 12:44:28 crc kubenswrapper[4632]: I0313 12:44:28.703047 4632 scope.go:117] "RemoveContainer" containerID="e9210685ea8bbb4f63a3e5f03db6817cbf29b557c698d4de6b1ae5929063a7c0" Mar 13 12:44:33 crc kubenswrapper[4632]: I0313 12:44:33.186777 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-gzkxx"] Mar 13 12:44:33 crc kubenswrapper[4632]: I0313 12:44:33.193698 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:44:33 crc kubenswrapper[4632]: I0313 12:44:33.195576 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gzkxx"] Mar 13 12:44:33 crc kubenswrapper[4632]: I0313 12:44:33.238434 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/850e7777-b942-4a4a-85ca-355a2ebd2ec9-utilities\") pod \"certified-operators-gzkxx\" (UID: \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\") " pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:44:33 crc kubenswrapper[4632]: I0313 12:44:33.239151 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/850e7777-b942-4a4a-85ca-355a2ebd2ec9-catalog-content\") pod \"certified-operators-gzkxx\" (UID: \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\") " pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:44:33 crc kubenswrapper[4632]: I0313 12:44:33.239267 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmcqj\" (UniqueName: \"kubernetes.io/projected/850e7777-b942-4a4a-85ca-355a2ebd2ec9-kube-api-access-tmcqj\") pod \"certified-operators-gzkxx\" (UID: \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\") " pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:44:33 crc kubenswrapper[4632]: I0313 12:44:33.340368 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/850e7777-b942-4a4a-85ca-355a2ebd2ec9-utilities\") pod \"certified-operators-gzkxx\" (UID: \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\") " pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:44:33 crc kubenswrapper[4632]: I0313 12:44:33.340452 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/850e7777-b942-4a4a-85ca-355a2ebd2ec9-catalog-content\") pod \"certified-operators-gzkxx\" (UID: \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\") " pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:44:33 crc kubenswrapper[4632]: I0313 12:44:33.340541 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmcqj\" (UniqueName: \"kubernetes.io/projected/850e7777-b942-4a4a-85ca-355a2ebd2ec9-kube-api-access-tmcqj\") pod \"certified-operators-gzkxx\" (UID: \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\") " pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:44:33 crc kubenswrapper[4632]: I0313 12:44:33.341015 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/850e7777-b942-4a4a-85ca-355a2ebd2ec9-utilities\") pod \"certified-operators-gzkxx\" (UID: \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\") " pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:44:33 crc kubenswrapper[4632]: I0313 12:44:33.341108 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/850e7777-b942-4a4a-85ca-355a2ebd2ec9-catalog-content\") pod \"certified-operators-gzkxx\" (UID: \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\") " pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:44:33 crc kubenswrapper[4632]: I0313 12:44:33.565272 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmcqj\" (UniqueName: \"kubernetes.io/projected/850e7777-b942-4a4a-85ca-355a2ebd2ec9-kube-api-access-tmcqj\") pod \"certified-operators-gzkxx\" (UID: \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\") " pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:44:33 crc kubenswrapper[4632]: I0313 12:44:33.817741 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:44:34 crc kubenswrapper[4632]: I0313 12:44:34.966251 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-gzkxx"] Mar 13 12:44:36 crc kubenswrapper[4632]: I0313 12:44:36.161922 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gzkxx" event={"ID":"850e7777-b942-4a4a-85ca-355a2ebd2ec9","Type":"ContainerStarted","Data":"3fe91ad8e992364755fc6ec2dea5c46c659b3d35b19fd4b30b1afd69688bec39"} Mar 13 12:44:38 crc kubenswrapper[4632]: I0313 12:44:38.181577 4632 generic.go:334] "Generic (PLEG): container finished" podID="850e7777-b942-4a4a-85ca-355a2ebd2ec9" containerID="88c9506140b859200b8b7485eae4c4fe840edad812655ef77eef068e2e2b3959" exitCode=0 Mar 13 12:44:38 crc kubenswrapper[4632]: I0313 12:44:38.181682 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gzkxx" event={"ID":"850e7777-b942-4a4a-85ca-355a2ebd2ec9","Type":"ContainerDied","Data":"88c9506140b859200b8b7485eae4c4fe840edad812655ef77eef068e2e2b3959"} Mar 13 12:44:39 crc kubenswrapper[4632]: I0313 12:44:39.217111 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gzkxx" event={"ID":"850e7777-b942-4a4a-85ca-355a2ebd2ec9","Type":"ContainerStarted","Data":"58931899bc81756bd316a55670c1933e22c756c6165c637e6a284e4a8b934ab3"} Mar 13 12:44:42 crc kubenswrapper[4632]: I0313 12:44:42.251251 4632 generic.go:334] "Generic (PLEG): container finished" podID="850e7777-b942-4a4a-85ca-355a2ebd2ec9" containerID="58931899bc81756bd316a55670c1933e22c756c6165c637e6a284e4a8b934ab3" exitCode=0 Mar 13 12:44:42 crc kubenswrapper[4632]: I0313 12:44:42.251341 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gzkxx" event={"ID":"850e7777-b942-4a4a-85ca-355a2ebd2ec9","Type":"ContainerDied","Data":"58931899bc81756bd316a55670c1933e22c756c6165c637e6a284e4a8b934ab3"} Mar 13 12:44:43 crc kubenswrapper[4632]: I0313 12:44:43.269555 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gzkxx" event={"ID":"850e7777-b942-4a4a-85ca-355a2ebd2ec9","Type":"ContainerStarted","Data":"5b7aa43a76c781e9b29d218c7062f057d96c8d06109d34357a6ad258bc4c6803"} Mar 13 12:44:43 crc kubenswrapper[4632]: I0313 12:44:43.304250 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-gzkxx" podStartSLOduration=5.709656055 podStartE2EDuration="10.304228027s" podCreationTimestamp="2026-03-13 12:44:33 +0000 UTC" firstStartedPulling="2026-03-13 12:44:38.183213186 +0000 UTC m=+9652.205743319" lastFinishedPulling="2026-03-13 12:44:42.777785158 +0000 UTC m=+9656.800315291" observedRunningTime="2026-03-13 12:44:43.293711249 +0000 UTC m=+9657.316241382" watchObservedRunningTime="2026-03-13 12:44:43.304228027 +0000 UTC m=+9657.326758160" Mar 13 12:44:43 crc kubenswrapper[4632]: I0313 12:44:43.818439 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:44:43 crc kubenswrapper[4632]: I0313 12:44:43.818873 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:44:44 crc kubenswrapper[4632]: I0313 12:44:44.878100 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-gzkxx" podUID="850e7777-b942-4a4a-85ca-355a2ebd2ec9" containerName="registry-server" probeResult="failure" output=< Mar 13 12:44:44 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:44:44 crc kubenswrapper[4632]: > Mar 13 12:44:55 crc kubenswrapper[4632]: I0313 12:44:55.024283 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-gzkxx" podUID="850e7777-b942-4a4a-85ca-355a2ebd2ec9" containerName="registry-server" probeResult="failure" output=< Mar 13 12:44:55 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:44:55 crc kubenswrapper[4632]: > Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.157062 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4"] Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.158755 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.160833 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.160963 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.174756 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4"] Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.314371 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/34a8177f-225c-4996-a654-1e50907b3249-config-volume\") pod \"collect-profiles-29556765-2tqx4\" (UID: \"34a8177f-225c-4996-a654-1e50907b3249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.314470 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/34a8177f-225c-4996-a654-1e50907b3249-secret-volume\") pod \"collect-profiles-29556765-2tqx4\" (UID: \"34a8177f-225c-4996-a654-1e50907b3249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.314577 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvvqn\" (UniqueName: \"kubernetes.io/projected/34a8177f-225c-4996-a654-1e50907b3249-kube-api-access-mvvqn\") pod \"collect-profiles-29556765-2tqx4\" (UID: \"34a8177f-225c-4996-a654-1e50907b3249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.415764 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/34a8177f-225c-4996-a654-1e50907b3249-secret-volume\") pod \"collect-profiles-29556765-2tqx4\" (UID: \"34a8177f-225c-4996-a654-1e50907b3249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.415910 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvvqn\" (UniqueName: \"kubernetes.io/projected/34a8177f-225c-4996-a654-1e50907b3249-kube-api-access-mvvqn\") pod \"collect-profiles-29556765-2tqx4\" (UID: \"34a8177f-225c-4996-a654-1e50907b3249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.416001 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/34a8177f-225c-4996-a654-1e50907b3249-config-volume\") pod \"collect-profiles-29556765-2tqx4\" (UID: \"34a8177f-225c-4996-a654-1e50907b3249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.417146 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/34a8177f-225c-4996-a654-1e50907b3249-config-volume\") pod \"collect-profiles-29556765-2tqx4\" (UID: \"34a8177f-225c-4996-a654-1e50907b3249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.426725 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/34a8177f-225c-4996-a654-1e50907b3249-secret-volume\") pod \"collect-profiles-29556765-2tqx4\" (UID: \"34a8177f-225c-4996-a654-1e50907b3249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.442094 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvvqn\" (UniqueName: \"kubernetes.io/projected/34a8177f-225c-4996-a654-1e50907b3249-kube-api-access-mvvqn\") pod \"collect-profiles-29556765-2tqx4\" (UID: \"34a8177f-225c-4996-a654-1e50907b3249\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" Mar 13 12:45:00 crc kubenswrapper[4632]: I0313 12:45:00.493520 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" Mar 13 12:45:01 crc kubenswrapper[4632]: I0313 12:45:01.109010 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4"] Mar 13 12:45:01 crc kubenswrapper[4632]: I0313 12:45:01.433617 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" event={"ID":"34a8177f-225c-4996-a654-1e50907b3249","Type":"ContainerStarted","Data":"b0423b15daf52bcd3217d9d6cdc18f717ea2ddda79e23e83c7754ec9daf0796a"} Mar 13 12:45:01 crc kubenswrapper[4632]: I0313 12:45:01.436014 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" event={"ID":"34a8177f-225c-4996-a654-1e50907b3249","Type":"ContainerStarted","Data":"2c03855a67ca775082742718ba4da7ba63b64bd75f781b18a29c8d5763aa88db"} Mar 13 12:45:01 crc kubenswrapper[4632]: I0313 12:45:01.456618 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" podStartSLOduration=1.456587015 podStartE2EDuration="1.456587015s" podCreationTimestamp="2026-03-13 12:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 12:45:01.448965098 +0000 UTC m=+9675.471495251" watchObservedRunningTime="2026-03-13 12:45:01.456587015 +0000 UTC m=+9675.479117138" Mar 13 12:45:02 crc kubenswrapper[4632]: I0313 12:45:02.444957 4632 generic.go:334] "Generic (PLEG): container finished" podID="34a8177f-225c-4996-a654-1e50907b3249" containerID="b0423b15daf52bcd3217d9d6cdc18f717ea2ddda79e23e83c7754ec9daf0796a" exitCode=0 Mar 13 12:45:02 crc kubenswrapper[4632]: I0313 12:45:02.445086 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" event={"ID":"34a8177f-225c-4996-a654-1e50907b3249","Type":"ContainerDied","Data":"b0423b15daf52bcd3217d9d6cdc18f717ea2ddda79e23e83c7754ec9daf0796a"} Mar 13 12:45:03 crc kubenswrapper[4632]: I0313 12:45:03.882078 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" Mar 13 12:45:03 crc kubenswrapper[4632]: I0313 12:45:03.919446 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:45:03 crc kubenswrapper[4632]: I0313 12:45:03.985147 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:45:03 crc kubenswrapper[4632]: I0313 12:45:03.988950 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/34a8177f-225c-4996-a654-1e50907b3249-config-volume\") pod \"34a8177f-225c-4996-a654-1e50907b3249\" (UID: \"34a8177f-225c-4996-a654-1e50907b3249\") " Mar 13 12:45:03 crc kubenswrapper[4632]: I0313 12:45:03.989016 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/34a8177f-225c-4996-a654-1e50907b3249-secret-volume\") pod \"34a8177f-225c-4996-a654-1e50907b3249\" (UID: \"34a8177f-225c-4996-a654-1e50907b3249\") " Mar 13 12:45:03 crc kubenswrapper[4632]: I0313 12:45:03.989078 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvvqn\" (UniqueName: \"kubernetes.io/projected/34a8177f-225c-4996-a654-1e50907b3249-kube-api-access-mvvqn\") pod \"34a8177f-225c-4996-a654-1e50907b3249\" (UID: \"34a8177f-225c-4996-a654-1e50907b3249\") " Mar 13 12:45:03 crc kubenswrapper[4632]: I0313 12:45:03.990696 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34a8177f-225c-4996-a654-1e50907b3249-config-volume" (OuterVolumeSpecName: "config-volume") pod "34a8177f-225c-4996-a654-1e50907b3249" (UID: "34a8177f-225c-4996-a654-1e50907b3249"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 12:45:03 crc kubenswrapper[4632]: I0313 12:45:03.998251 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/34a8177f-225c-4996-a654-1e50907b3249-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "34a8177f-225c-4996-a654-1e50907b3249" (UID: "34a8177f-225c-4996-a654-1e50907b3249"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 12:45:03 crc kubenswrapper[4632]: I0313 12:45:03.999947 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34a8177f-225c-4996-a654-1e50907b3249-kube-api-access-mvvqn" (OuterVolumeSpecName: "kube-api-access-mvvqn") pod "34a8177f-225c-4996-a654-1e50907b3249" (UID: "34a8177f-225c-4996-a654-1e50907b3249"). InnerVolumeSpecName "kube-api-access-mvvqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:45:04 crc kubenswrapper[4632]: I0313 12:45:04.091291 4632 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/34a8177f-225c-4996-a654-1e50907b3249-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:04 crc kubenswrapper[4632]: I0313 12:45:04.091337 4632 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/34a8177f-225c-4996-a654-1e50907b3249-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:04 crc kubenswrapper[4632]: I0313 12:45:04.091351 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvvqn\" (UniqueName: \"kubernetes.io/projected/34a8177f-225c-4996-a654-1e50907b3249-kube-api-access-mvvqn\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:04 crc kubenswrapper[4632]: I0313 12:45:04.410927 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gzkxx"] Mar 13 12:45:04 crc kubenswrapper[4632]: I0313 12:45:04.465665 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" event={"ID":"34a8177f-225c-4996-a654-1e50907b3249","Type":"ContainerDied","Data":"2c03855a67ca775082742718ba4da7ba63b64bd75f781b18a29c8d5763aa88db"} Mar 13 12:45:04 crc kubenswrapper[4632]: I0313 12:45:04.465714 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c03855a67ca775082742718ba4da7ba63b64bd75f781b18a29c8d5763aa88db" Mar 13 12:45:04 crc kubenswrapper[4632]: I0313 12:45:04.465715 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556765-2tqx4" Mar 13 12:45:04 crc kubenswrapper[4632]: I0313 12:45:04.549966 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7"] Mar 13 12:45:04 crc kubenswrapper[4632]: I0313 12:45:04.559416 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556720-9qfd7"] Mar 13 12:45:05 crc kubenswrapper[4632]: I0313 12:45:05.473866 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-gzkxx" podUID="850e7777-b942-4a4a-85ca-355a2ebd2ec9" containerName="registry-server" containerID="cri-o://5b7aa43a76c781e9b29d218c7062f057d96c8d06109d34357a6ad258bc4c6803" gracePeriod=2 Mar 13 12:45:05 crc kubenswrapper[4632]: I0313 12:45:05.973752 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.058158 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="453f2bd4-a723-4b7f-9b06-05d75e8df7b8" path="/var/lib/kubelet/pods/453f2bd4-a723-4b7f-9b06-05d75e8df7b8/volumes" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.128696 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/850e7777-b942-4a4a-85ca-355a2ebd2ec9-utilities\") pod \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\" (UID: \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\") " Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.128871 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmcqj\" (UniqueName: \"kubernetes.io/projected/850e7777-b942-4a4a-85ca-355a2ebd2ec9-kube-api-access-tmcqj\") pod \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\" (UID: \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\") " Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.128959 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/850e7777-b942-4a4a-85ca-355a2ebd2ec9-catalog-content\") pod \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\" (UID: \"850e7777-b942-4a4a-85ca-355a2ebd2ec9\") " Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.129470 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/850e7777-b942-4a4a-85ca-355a2ebd2ec9-utilities" (OuterVolumeSpecName: "utilities") pod "850e7777-b942-4a4a-85ca-355a2ebd2ec9" (UID: "850e7777-b942-4a4a-85ca-355a2ebd2ec9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.190003 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/850e7777-b942-4a4a-85ca-355a2ebd2ec9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "850e7777-b942-4a4a-85ca-355a2ebd2ec9" (UID: "850e7777-b942-4a4a-85ca-355a2ebd2ec9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.230726 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/850e7777-b942-4a4a-85ca-355a2ebd2ec9-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.230763 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/850e7777-b942-4a4a-85ca-355a2ebd2ec9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.464215 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/850e7777-b942-4a4a-85ca-355a2ebd2ec9-kube-api-access-tmcqj" (OuterVolumeSpecName: "kube-api-access-tmcqj") pod "850e7777-b942-4a4a-85ca-355a2ebd2ec9" (UID: "850e7777-b942-4a4a-85ca-355a2ebd2ec9"). InnerVolumeSpecName "kube-api-access-tmcqj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.486381 4632 generic.go:334] "Generic (PLEG): container finished" podID="850e7777-b942-4a4a-85ca-355a2ebd2ec9" containerID="5b7aa43a76c781e9b29d218c7062f057d96c8d06109d34357a6ad258bc4c6803" exitCode=0 Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.486425 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gzkxx" event={"ID":"850e7777-b942-4a4a-85ca-355a2ebd2ec9","Type":"ContainerDied","Data":"5b7aa43a76c781e9b29d218c7062f057d96c8d06109d34357a6ad258bc4c6803"} Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.486458 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-gzkxx" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.486477 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-gzkxx" event={"ID":"850e7777-b942-4a4a-85ca-355a2ebd2ec9","Type":"ContainerDied","Data":"3fe91ad8e992364755fc6ec2dea5c46c659b3d35b19fd4b30b1afd69688bec39"} Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.486500 4632 scope.go:117] "RemoveContainer" containerID="5b7aa43a76c781e9b29d218c7062f057d96c8d06109d34357a6ad258bc4c6803" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.536399 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmcqj\" (UniqueName: \"kubernetes.io/projected/850e7777-b942-4a4a-85ca-355a2ebd2ec9-kube-api-access-tmcqj\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.546949 4632 scope.go:117] "RemoveContainer" containerID="58931899bc81756bd316a55670c1933e22c756c6165c637e6a284e4a8b934ab3" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.558146 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-gzkxx"] Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.572770 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-gzkxx"] Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.583637 4632 scope.go:117] "RemoveContainer" containerID="88c9506140b859200b8b7485eae4c4fe840edad812655ef77eef068e2e2b3959" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.641975 4632 scope.go:117] "RemoveContainer" containerID="5b7aa43a76c781e9b29d218c7062f057d96c8d06109d34357a6ad258bc4c6803" Mar 13 12:45:06 crc kubenswrapper[4632]: E0313 12:45:06.644747 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b7aa43a76c781e9b29d218c7062f057d96c8d06109d34357a6ad258bc4c6803\": container with ID starting with 5b7aa43a76c781e9b29d218c7062f057d96c8d06109d34357a6ad258bc4c6803 not found: ID does not exist" containerID="5b7aa43a76c781e9b29d218c7062f057d96c8d06109d34357a6ad258bc4c6803" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.644807 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b7aa43a76c781e9b29d218c7062f057d96c8d06109d34357a6ad258bc4c6803"} err="failed to get container status \"5b7aa43a76c781e9b29d218c7062f057d96c8d06109d34357a6ad258bc4c6803\": rpc error: code = NotFound desc = could not find container \"5b7aa43a76c781e9b29d218c7062f057d96c8d06109d34357a6ad258bc4c6803\": container with ID starting with 5b7aa43a76c781e9b29d218c7062f057d96c8d06109d34357a6ad258bc4c6803 not found: ID does not exist" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.644840 4632 scope.go:117] "RemoveContainer" containerID="58931899bc81756bd316a55670c1933e22c756c6165c637e6a284e4a8b934ab3" Mar 13 12:45:06 crc kubenswrapper[4632]: E0313 12:45:06.645322 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58931899bc81756bd316a55670c1933e22c756c6165c637e6a284e4a8b934ab3\": container with ID starting with 58931899bc81756bd316a55670c1933e22c756c6165c637e6a284e4a8b934ab3 not found: ID does not exist" containerID="58931899bc81756bd316a55670c1933e22c756c6165c637e6a284e4a8b934ab3" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.645363 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58931899bc81756bd316a55670c1933e22c756c6165c637e6a284e4a8b934ab3"} err="failed to get container status \"58931899bc81756bd316a55670c1933e22c756c6165c637e6a284e4a8b934ab3\": rpc error: code = NotFound desc = could not find container \"58931899bc81756bd316a55670c1933e22c756c6165c637e6a284e4a8b934ab3\": container with ID starting with 58931899bc81756bd316a55670c1933e22c756c6165c637e6a284e4a8b934ab3 not found: ID does not exist" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.645388 4632 scope.go:117] "RemoveContainer" containerID="88c9506140b859200b8b7485eae4c4fe840edad812655ef77eef068e2e2b3959" Mar 13 12:45:06 crc kubenswrapper[4632]: E0313 12:45:06.645811 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"88c9506140b859200b8b7485eae4c4fe840edad812655ef77eef068e2e2b3959\": container with ID starting with 88c9506140b859200b8b7485eae4c4fe840edad812655ef77eef068e2e2b3959 not found: ID does not exist" containerID="88c9506140b859200b8b7485eae4c4fe840edad812655ef77eef068e2e2b3959" Mar 13 12:45:06 crc kubenswrapper[4632]: I0313 12:45:06.645833 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"88c9506140b859200b8b7485eae4c4fe840edad812655ef77eef068e2e2b3959"} err="failed to get container status \"88c9506140b859200b8b7485eae4c4fe840edad812655ef77eef068e2e2b3959\": rpc error: code = NotFound desc = could not find container \"88c9506140b859200b8b7485eae4c4fe840edad812655ef77eef068e2e2b3959\": container with ID starting with 88c9506140b859200b8b7485eae4c4fe840edad812655ef77eef068e2e2b3959 not found: ID does not exist" Mar 13 12:45:08 crc kubenswrapper[4632]: I0313 12:45:08.058244 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="850e7777-b942-4a4a-85ca-355a2ebd2ec9" path="/var/lib/kubelet/pods/850e7777-b942-4a4a-85ca-355a2ebd2ec9/volumes" Mar 13 12:45:10 crc kubenswrapper[4632]: I0313 12:45:10.460735 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:45:10 crc kubenswrapper[4632]: I0313 12:45:10.462207 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:45:18 crc kubenswrapper[4632]: I0313 12:45:18.600039 4632 generic.go:334] "Generic (PLEG): container finished" podID="3ccc7848-e5be-416d-95de-b621b5cc770d" containerID="ef8f3145eaca72f4029fdc60f2fb5cc46e463f526ce49ed7c2f109c93d6646f5" exitCode=0 Mar 13 12:45:18 crc kubenswrapper[4632]: I0313 12:45:18.600274 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfn52/crc-debug-9w6fr" event={"ID":"3ccc7848-e5be-416d-95de-b621b5cc770d","Type":"ContainerDied","Data":"ef8f3145eaca72f4029fdc60f2fb5cc46e463f526ce49ed7c2f109c93d6646f5"} Mar 13 12:45:19 crc kubenswrapper[4632]: I0313 12:45:19.745139 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/crc-debug-9w6fr" Mar 13 12:45:19 crc kubenswrapper[4632]: I0313 12:45:19.791701 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jfn52/crc-debug-9w6fr"] Mar 13 12:45:19 crc kubenswrapper[4632]: I0313 12:45:19.803493 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jfn52/crc-debug-9w6fr"] Mar 13 12:45:19 crc kubenswrapper[4632]: I0313 12:45:19.873905 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3ccc7848-e5be-416d-95de-b621b5cc770d-host\") pod \"3ccc7848-e5be-416d-95de-b621b5cc770d\" (UID: \"3ccc7848-e5be-416d-95de-b621b5cc770d\") " Mar 13 12:45:19 crc kubenswrapper[4632]: I0313 12:45:19.874041 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3ccc7848-e5be-416d-95de-b621b5cc770d-host" (OuterVolumeSpecName: "host") pod "3ccc7848-e5be-416d-95de-b621b5cc770d" (UID: "3ccc7848-e5be-416d-95de-b621b5cc770d"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 12:45:19 crc kubenswrapper[4632]: I0313 12:45:19.874080 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7bfjh\" (UniqueName: \"kubernetes.io/projected/3ccc7848-e5be-416d-95de-b621b5cc770d-kube-api-access-7bfjh\") pod \"3ccc7848-e5be-416d-95de-b621b5cc770d\" (UID: \"3ccc7848-e5be-416d-95de-b621b5cc770d\") " Mar 13 12:45:19 crc kubenswrapper[4632]: I0313 12:45:19.874536 4632 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/3ccc7848-e5be-416d-95de-b621b5cc770d-host\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:19 crc kubenswrapper[4632]: I0313 12:45:19.882547 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ccc7848-e5be-416d-95de-b621b5cc770d-kube-api-access-7bfjh" (OuterVolumeSpecName: "kube-api-access-7bfjh") pod "3ccc7848-e5be-416d-95de-b621b5cc770d" (UID: "3ccc7848-e5be-416d-95de-b621b5cc770d"). InnerVolumeSpecName "kube-api-access-7bfjh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:45:19 crc kubenswrapper[4632]: I0313 12:45:19.975956 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7bfjh\" (UniqueName: \"kubernetes.io/projected/3ccc7848-e5be-416d-95de-b621b5cc770d-kube-api-access-7bfjh\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:20 crc kubenswrapper[4632]: I0313 12:45:20.059301 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ccc7848-e5be-416d-95de-b621b5cc770d" path="/var/lib/kubelet/pods/3ccc7848-e5be-416d-95de-b621b5cc770d/volumes" Mar 13 12:45:20 crc kubenswrapper[4632]: I0313 12:45:20.629120 4632 scope.go:117] "RemoveContainer" containerID="ef8f3145eaca72f4029fdc60f2fb5cc46e463f526ce49ed7c2f109c93d6646f5" Mar 13 12:45:20 crc kubenswrapper[4632]: I0313 12:45:20.629131 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/crc-debug-9w6fr" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.033089 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jfn52/crc-debug-mm9wk"] Mar 13 12:45:21 crc kubenswrapper[4632]: E0313 12:45:21.033607 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ccc7848-e5be-416d-95de-b621b5cc770d" containerName="container-00" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.034296 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ccc7848-e5be-416d-95de-b621b5cc770d" containerName="container-00" Mar 13 12:45:21 crc kubenswrapper[4632]: E0313 12:45:21.034334 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34a8177f-225c-4996-a654-1e50907b3249" containerName="collect-profiles" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.034346 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="34a8177f-225c-4996-a654-1e50907b3249" containerName="collect-profiles" Mar 13 12:45:21 crc kubenswrapper[4632]: E0313 12:45:21.034369 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="850e7777-b942-4a4a-85ca-355a2ebd2ec9" containerName="extract-content" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.034378 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="850e7777-b942-4a4a-85ca-355a2ebd2ec9" containerName="extract-content" Mar 13 12:45:21 crc kubenswrapper[4632]: E0313 12:45:21.034403 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="850e7777-b942-4a4a-85ca-355a2ebd2ec9" containerName="extract-utilities" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.034412 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="850e7777-b942-4a4a-85ca-355a2ebd2ec9" containerName="extract-utilities" Mar 13 12:45:21 crc kubenswrapper[4632]: E0313 12:45:21.034434 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="850e7777-b942-4a4a-85ca-355a2ebd2ec9" containerName="registry-server" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.034443 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="850e7777-b942-4a4a-85ca-355a2ebd2ec9" containerName="registry-server" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.034761 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="850e7777-b942-4a4a-85ca-355a2ebd2ec9" containerName="registry-server" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.034805 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ccc7848-e5be-416d-95de-b621b5cc770d" containerName="container-00" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.034817 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="34a8177f-225c-4996-a654-1e50907b3249" containerName="collect-profiles" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.035699 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/crc-debug-mm9wk" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.096276 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae-host\") pod \"crc-debug-mm9wk\" (UID: \"7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae\") " pod="openshift-must-gather-jfn52/crc-debug-mm9wk" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.096329 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gc7bg\" (UniqueName: \"kubernetes.io/projected/7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae-kube-api-access-gc7bg\") pod \"crc-debug-mm9wk\" (UID: \"7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae\") " pod="openshift-must-gather-jfn52/crc-debug-mm9wk" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.197695 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae-host\") pod \"crc-debug-mm9wk\" (UID: \"7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae\") " pod="openshift-must-gather-jfn52/crc-debug-mm9wk" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.198039 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gc7bg\" (UniqueName: \"kubernetes.io/projected/7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae-kube-api-access-gc7bg\") pod \"crc-debug-mm9wk\" (UID: \"7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae\") " pod="openshift-must-gather-jfn52/crc-debug-mm9wk" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.199760 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae-host\") pod \"crc-debug-mm9wk\" (UID: \"7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae\") " pod="openshift-must-gather-jfn52/crc-debug-mm9wk" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.217405 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gc7bg\" (UniqueName: \"kubernetes.io/projected/7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae-kube-api-access-gc7bg\") pod \"crc-debug-mm9wk\" (UID: \"7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae\") " pod="openshift-must-gather-jfn52/crc-debug-mm9wk" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.351412 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/crc-debug-mm9wk" Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.639554 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfn52/crc-debug-mm9wk" event={"ID":"7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae","Type":"ContainerStarted","Data":"f9576f389f75db79c6cd02f685bff29c0e4ed007b62591df661e2a8ee57c8ce2"} Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.639599 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfn52/crc-debug-mm9wk" event={"ID":"7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae","Type":"ContainerStarted","Data":"f533186bae84ac6e618c29157ca171a61d7d3023adf769459e9d86f07e720694"} Mar 13 12:45:21 crc kubenswrapper[4632]: I0313 12:45:21.672740 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-jfn52/crc-debug-mm9wk" podStartSLOduration=0.672717328 podStartE2EDuration="672.717328ms" podCreationTimestamp="2026-03-13 12:45:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 12:45:21.659213896 +0000 UTC m=+9695.681744049" watchObservedRunningTime="2026-03-13 12:45:21.672717328 +0000 UTC m=+9695.695247471" Mar 13 12:45:22 crc kubenswrapper[4632]: I0313 12:45:22.695320 4632 generic.go:334] "Generic (PLEG): container finished" podID="7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae" containerID="f9576f389f75db79c6cd02f685bff29c0e4ed007b62591df661e2a8ee57c8ce2" exitCode=0 Mar 13 12:45:22 crc kubenswrapper[4632]: I0313 12:45:22.696668 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfn52/crc-debug-mm9wk" event={"ID":"7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae","Type":"ContainerDied","Data":"f9576f389f75db79c6cd02f685bff29c0e4ed007b62591df661e2a8ee57c8ce2"} Mar 13 12:45:23 crc kubenswrapper[4632]: I0313 12:45:23.862476 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/crc-debug-mm9wk" Mar 13 12:45:24 crc kubenswrapper[4632]: I0313 12:45:24.005007 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae-host\") pod \"7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae\" (UID: \"7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae\") " Mar 13 12:45:24 crc kubenswrapper[4632]: I0313 12:45:24.005172 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae-host" (OuterVolumeSpecName: "host") pod "7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae" (UID: "7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 12:45:24 crc kubenswrapper[4632]: I0313 12:45:24.005207 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gc7bg\" (UniqueName: \"kubernetes.io/projected/7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae-kube-api-access-gc7bg\") pod \"7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae\" (UID: \"7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae\") " Mar 13 12:45:24 crc kubenswrapper[4632]: I0313 12:45:24.005776 4632 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae-host\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:24 crc kubenswrapper[4632]: I0313 12:45:24.016096 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae-kube-api-access-gc7bg" (OuterVolumeSpecName: "kube-api-access-gc7bg") pod "7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae" (UID: "7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae"). InnerVolumeSpecName "kube-api-access-gc7bg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:45:24 crc kubenswrapper[4632]: I0313 12:45:24.124533 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gc7bg\" (UniqueName: \"kubernetes.io/projected/7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae-kube-api-access-gc7bg\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:24 crc kubenswrapper[4632]: I0313 12:45:24.222009 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jfn52/crc-debug-mm9wk"] Mar 13 12:45:24 crc kubenswrapper[4632]: I0313 12:45:24.233711 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jfn52/crc-debug-mm9wk"] Mar 13 12:45:24 crc kubenswrapper[4632]: I0313 12:45:24.713673 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f533186bae84ac6e618c29157ca171a61d7d3023adf769459e9d86f07e720694" Mar 13 12:45:24 crc kubenswrapper[4632]: I0313 12:45:24.713891 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/crc-debug-mm9wk" Mar 13 12:45:25 crc kubenswrapper[4632]: I0313 12:45:25.453862 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-jfn52/crc-debug-wv7nx"] Mar 13 12:45:25 crc kubenswrapper[4632]: E0313 12:45:25.454295 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae" containerName="container-00" Mar 13 12:45:25 crc kubenswrapper[4632]: I0313 12:45:25.454308 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae" containerName="container-00" Mar 13 12:45:25 crc kubenswrapper[4632]: I0313 12:45:25.454482 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae" containerName="container-00" Mar 13 12:45:25 crc kubenswrapper[4632]: I0313 12:45:25.455349 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/crc-debug-wv7nx" Mar 13 12:45:25 crc kubenswrapper[4632]: I0313 12:45:25.552523 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84dbf8f9-b22a-4cd1-8589-6c437bc73a36-host\") pod \"crc-debug-wv7nx\" (UID: \"84dbf8f9-b22a-4cd1-8589-6c437bc73a36\") " pod="openshift-must-gather-jfn52/crc-debug-wv7nx" Mar 13 12:45:25 crc kubenswrapper[4632]: I0313 12:45:25.552594 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jgfj\" (UniqueName: \"kubernetes.io/projected/84dbf8f9-b22a-4cd1-8589-6c437bc73a36-kube-api-access-7jgfj\") pod \"crc-debug-wv7nx\" (UID: \"84dbf8f9-b22a-4cd1-8589-6c437bc73a36\") " pod="openshift-must-gather-jfn52/crc-debug-wv7nx" Mar 13 12:45:25 crc kubenswrapper[4632]: I0313 12:45:25.654730 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84dbf8f9-b22a-4cd1-8589-6c437bc73a36-host\") pod \"crc-debug-wv7nx\" (UID: \"84dbf8f9-b22a-4cd1-8589-6c437bc73a36\") " pod="openshift-must-gather-jfn52/crc-debug-wv7nx" Mar 13 12:45:25 crc kubenswrapper[4632]: I0313 12:45:25.655167 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jgfj\" (UniqueName: \"kubernetes.io/projected/84dbf8f9-b22a-4cd1-8589-6c437bc73a36-kube-api-access-7jgfj\") pod \"crc-debug-wv7nx\" (UID: \"84dbf8f9-b22a-4cd1-8589-6c437bc73a36\") " pod="openshift-must-gather-jfn52/crc-debug-wv7nx" Mar 13 12:45:25 crc kubenswrapper[4632]: I0313 12:45:25.654882 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84dbf8f9-b22a-4cd1-8589-6c437bc73a36-host\") pod \"crc-debug-wv7nx\" (UID: \"84dbf8f9-b22a-4cd1-8589-6c437bc73a36\") " pod="openshift-must-gather-jfn52/crc-debug-wv7nx" Mar 13 12:45:25 crc kubenswrapper[4632]: I0313 12:45:25.680698 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jgfj\" (UniqueName: \"kubernetes.io/projected/84dbf8f9-b22a-4cd1-8589-6c437bc73a36-kube-api-access-7jgfj\") pod \"crc-debug-wv7nx\" (UID: \"84dbf8f9-b22a-4cd1-8589-6c437bc73a36\") " pod="openshift-must-gather-jfn52/crc-debug-wv7nx" Mar 13 12:45:25 crc kubenswrapper[4632]: I0313 12:45:25.777402 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/crc-debug-wv7nx" Mar 13 12:45:26 crc kubenswrapper[4632]: I0313 12:45:26.083725 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae" path="/var/lib/kubelet/pods/7a5eb284-6ea3-45af-bd45-6ba4aa90b9ae/volumes" Mar 13 12:45:26 crc kubenswrapper[4632]: I0313 12:45:26.736354 4632 generic.go:334] "Generic (PLEG): container finished" podID="84dbf8f9-b22a-4cd1-8589-6c437bc73a36" containerID="8c80829077510548bea49d9ed53848049ef8179bb726bd5562c95f37f2977880" exitCode=0 Mar 13 12:45:26 crc kubenswrapper[4632]: I0313 12:45:26.736449 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfn52/crc-debug-wv7nx" event={"ID":"84dbf8f9-b22a-4cd1-8589-6c437bc73a36","Type":"ContainerDied","Data":"8c80829077510548bea49d9ed53848049ef8179bb726bd5562c95f37f2977880"} Mar 13 12:45:26 crc kubenswrapper[4632]: I0313 12:45:26.736655 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfn52/crc-debug-wv7nx" event={"ID":"84dbf8f9-b22a-4cd1-8589-6c437bc73a36","Type":"ContainerStarted","Data":"adceec6a78211534843923e49837eb34cf6d7f0ef86a1e17c8ecd2fc03018512"} Mar 13 12:45:26 crc kubenswrapper[4632]: I0313 12:45:26.778307 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jfn52/crc-debug-wv7nx"] Mar 13 12:45:26 crc kubenswrapper[4632]: I0313 12:45:26.789921 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jfn52/crc-debug-wv7nx"] Mar 13 12:45:27 crc kubenswrapper[4632]: I0313 12:45:27.847658 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/crc-debug-wv7nx" Mar 13 12:45:28 crc kubenswrapper[4632]: I0313 12:45:28.015507 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jgfj\" (UniqueName: \"kubernetes.io/projected/84dbf8f9-b22a-4cd1-8589-6c437bc73a36-kube-api-access-7jgfj\") pod \"84dbf8f9-b22a-4cd1-8589-6c437bc73a36\" (UID: \"84dbf8f9-b22a-4cd1-8589-6c437bc73a36\") " Mar 13 12:45:28 crc kubenswrapper[4632]: I0313 12:45:28.015854 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84dbf8f9-b22a-4cd1-8589-6c437bc73a36-host\") pod \"84dbf8f9-b22a-4cd1-8589-6c437bc73a36\" (UID: \"84dbf8f9-b22a-4cd1-8589-6c437bc73a36\") " Mar 13 12:45:28 crc kubenswrapper[4632]: I0313 12:45:28.016000 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/84dbf8f9-b22a-4cd1-8589-6c437bc73a36-host" (OuterVolumeSpecName: "host") pod "84dbf8f9-b22a-4cd1-8589-6c437bc73a36" (UID: "84dbf8f9-b22a-4cd1-8589-6c437bc73a36"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 12:45:28 crc kubenswrapper[4632]: I0313 12:45:28.016619 4632 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/84dbf8f9-b22a-4cd1-8589-6c437bc73a36-host\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:28 crc kubenswrapper[4632]: I0313 12:45:28.023208 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/84dbf8f9-b22a-4cd1-8589-6c437bc73a36-kube-api-access-7jgfj" (OuterVolumeSpecName: "kube-api-access-7jgfj") pod "84dbf8f9-b22a-4cd1-8589-6c437bc73a36" (UID: "84dbf8f9-b22a-4cd1-8589-6c437bc73a36"). InnerVolumeSpecName "kube-api-access-7jgfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:45:28 crc kubenswrapper[4632]: I0313 12:45:28.055325 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="84dbf8f9-b22a-4cd1-8589-6c437bc73a36" path="/var/lib/kubelet/pods/84dbf8f9-b22a-4cd1-8589-6c437bc73a36/volumes" Mar 13 12:45:28 crc kubenswrapper[4632]: I0313 12:45:28.118231 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jgfj\" (UniqueName: \"kubernetes.io/projected/84dbf8f9-b22a-4cd1-8589-6c437bc73a36-kube-api-access-7jgfj\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:28 crc kubenswrapper[4632]: I0313 12:45:28.755165 4632 scope.go:117] "RemoveContainer" containerID="8c80829077510548bea49d9ed53848049ef8179bb726bd5562c95f37f2977880" Mar 13 12:45:28 crc kubenswrapper[4632]: I0313 12:45:28.755192 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/crc-debug-wv7nx" Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.137363 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jbtjg"] Mar 13 12:45:29 crc kubenswrapper[4632]: E0313 12:45:29.138223 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="84dbf8f9-b22a-4cd1-8589-6c437bc73a36" containerName="container-00" Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.138246 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="84dbf8f9-b22a-4cd1-8589-6c437bc73a36" containerName="container-00" Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.138517 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="84dbf8f9-b22a-4cd1-8589-6c437bc73a36" containerName="container-00" Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.139865 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.157153 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbtjg"] Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.239743 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90e1bcb1-81b1-42c2-a625-fe691fe60434-utilities\") pod \"redhat-marketplace-jbtjg\" (UID: \"90e1bcb1-81b1-42c2-a625-fe691fe60434\") " pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.239788 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrqsw\" (UniqueName: \"kubernetes.io/projected/90e1bcb1-81b1-42c2-a625-fe691fe60434-kube-api-access-nrqsw\") pod \"redhat-marketplace-jbtjg\" (UID: \"90e1bcb1-81b1-42c2-a625-fe691fe60434\") " pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.239835 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90e1bcb1-81b1-42c2-a625-fe691fe60434-catalog-content\") pod \"redhat-marketplace-jbtjg\" (UID: \"90e1bcb1-81b1-42c2-a625-fe691fe60434\") " pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.341708 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90e1bcb1-81b1-42c2-a625-fe691fe60434-catalog-content\") pod \"redhat-marketplace-jbtjg\" (UID: \"90e1bcb1-81b1-42c2-a625-fe691fe60434\") " pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.341911 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90e1bcb1-81b1-42c2-a625-fe691fe60434-utilities\") pod \"redhat-marketplace-jbtjg\" (UID: \"90e1bcb1-81b1-42c2-a625-fe691fe60434\") " pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.341976 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrqsw\" (UniqueName: \"kubernetes.io/projected/90e1bcb1-81b1-42c2-a625-fe691fe60434-kube-api-access-nrqsw\") pod \"redhat-marketplace-jbtjg\" (UID: \"90e1bcb1-81b1-42c2-a625-fe691fe60434\") " pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.342320 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90e1bcb1-81b1-42c2-a625-fe691fe60434-catalog-content\") pod \"redhat-marketplace-jbtjg\" (UID: \"90e1bcb1-81b1-42c2-a625-fe691fe60434\") " pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.342369 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90e1bcb1-81b1-42c2-a625-fe691fe60434-utilities\") pod \"redhat-marketplace-jbtjg\" (UID: \"90e1bcb1-81b1-42c2-a625-fe691fe60434\") " pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.362337 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrqsw\" (UniqueName: \"kubernetes.io/projected/90e1bcb1-81b1-42c2-a625-fe691fe60434-kube-api-access-nrqsw\") pod \"redhat-marketplace-jbtjg\" (UID: \"90e1bcb1-81b1-42c2-a625-fe691fe60434\") " pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:29 crc kubenswrapper[4632]: I0313 12:45:29.461663 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:30 crc kubenswrapper[4632]: I0313 12:45:30.084557 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbtjg"] Mar 13 12:45:30 crc kubenswrapper[4632]: I0313 12:45:30.135425 4632 scope.go:117] "RemoveContainer" containerID="513ba32a6f64209e9e7a4b86369065ec16320243702d6e9f6899a7182c651338" Mar 13 12:45:30 crc kubenswrapper[4632]: I0313 12:45:30.778560 4632 generic.go:334] "Generic (PLEG): container finished" podID="90e1bcb1-81b1-42c2-a625-fe691fe60434" containerID="a5eb8f564f7bb06a4d39ae737d43ff00b69617f7796a97fc2d384634398638cc" exitCode=0 Mar 13 12:45:30 crc kubenswrapper[4632]: I0313 12:45:30.778621 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbtjg" event={"ID":"90e1bcb1-81b1-42c2-a625-fe691fe60434","Type":"ContainerDied","Data":"a5eb8f564f7bb06a4d39ae737d43ff00b69617f7796a97fc2d384634398638cc"} Mar 13 12:45:30 crc kubenswrapper[4632]: I0313 12:45:30.778686 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbtjg" event={"ID":"90e1bcb1-81b1-42c2-a625-fe691fe60434","Type":"ContainerStarted","Data":"7c9d33658cf93236a127bb3fe33f8e8067d984a174cb915b8ecf156bd22eb63f"} Mar 13 12:45:31 crc kubenswrapper[4632]: I0313 12:45:31.788898 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbtjg" event={"ID":"90e1bcb1-81b1-42c2-a625-fe691fe60434","Type":"ContainerStarted","Data":"c40b8467f0fbf11bb8bc8bea09fc62d937cebf42c2af6303174926f814c485aa"} Mar 13 12:45:33 crc kubenswrapper[4632]: I0313 12:45:33.816679 4632 generic.go:334] "Generic (PLEG): container finished" podID="90e1bcb1-81b1-42c2-a625-fe691fe60434" containerID="c40b8467f0fbf11bb8bc8bea09fc62d937cebf42c2af6303174926f814c485aa" exitCode=0 Mar 13 12:45:33 crc kubenswrapper[4632]: I0313 12:45:33.817348 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbtjg" event={"ID":"90e1bcb1-81b1-42c2-a625-fe691fe60434","Type":"ContainerDied","Data":"c40b8467f0fbf11bb8bc8bea09fc62d937cebf42c2af6303174926f814c485aa"} Mar 13 12:45:34 crc kubenswrapper[4632]: I0313 12:45:34.833614 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbtjg" event={"ID":"90e1bcb1-81b1-42c2-a625-fe691fe60434","Type":"ContainerStarted","Data":"b40e35ab8281b9feedc9a3c0e3cb1babc5421097a7d08144dee8b68a590ba85f"} Mar 13 12:45:34 crc kubenswrapper[4632]: I0313 12:45:34.864398 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jbtjg" podStartSLOduration=2.39969714 podStartE2EDuration="5.864378367s" podCreationTimestamp="2026-03-13 12:45:29 +0000 UTC" firstStartedPulling="2026-03-13 12:45:30.781230444 +0000 UTC m=+9704.803760577" lastFinishedPulling="2026-03-13 12:45:34.245911671 +0000 UTC m=+9708.268441804" observedRunningTime="2026-03-13 12:45:34.858881402 +0000 UTC m=+9708.881411535" watchObservedRunningTime="2026-03-13 12:45:34.864378367 +0000 UTC m=+9708.886908510" Mar 13 12:45:39 crc kubenswrapper[4632]: I0313 12:45:39.462853 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:39 crc kubenswrapper[4632]: I0313 12:45:39.463495 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:40 crc kubenswrapper[4632]: I0313 12:45:40.460769 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:45:40 crc kubenswrapper[4632]: I0313 12:45:40.461365 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:45:40 crc kubenswrapper[4632]: I0313 12:45:40.794083 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-jbtjg" podUID="90e1bcb1-81b1-42c2-a625-fe691fe60434" containerName="registry-server" probeResult="failure" output=< Mar 13 12:45:40 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:45:40 crc kubenswrapper[4632]: > Mar 13 12:45:49 crc kubenswrapper[4632]: I0313 12:45:49.523315 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:49 crc kubenswrapper[4632]: I0313 12:45:49.583730 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:49 crc kubenswrapper[4632]: I0313 12:45:49.762075 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbtjg"] Mar 13 12:45:50 crc kubenswrapper[4632]: I0313 12:45:50.987144 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jbtjg" podUID="90e1bcb1-81b1-42c2-a625-fe691fe60434" containerName="registry-server" containerID="cri-o://b40e35ab8281b9feedc9a3c0e3cb1babc5421097a7d08144dee8b68a590ba85f" gracePeriod=2 Mar 13 12:45:51 crc kubenswrapper[4632]: I0313 12:45:51.491650 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:51 crc kubenswrapper[4632]: I0313 12:45:51.651229 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90e1bcb1-81b1-42c2-a625-fe691fe60434-utilities\") pod \"90e1bcb1-81b1-42c2-a625-fe691fe60434\" (UID: \"90e1bcb1-81b1-42c2-a625-fe691fe60434\") " Mar 13 12:45:51 crc kubenswrapper[4632]: I0313 12:45:51.651926 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90e1bcb1-81b1-42c2-a625-fe691fe60434-utilities" (OuterVolumeSpecName: "utilities") pod "90e1bcb1-81b1-42c2-a625-fe691fe60434" (UID: "90e1bcb1-81b1-42c2-a625-fe691fe60434"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:45:51 crc kubenswrapper[4632]: I0313 12:45:51.657106 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90e1bcb1-81b1-42c2-a625-fe691fe60434-catalog-content\") pod \"90e1bcb1-81b1-42c2-a625-fe691fe60434\" (UID: \"90e1bcb1-81b1-42c2-a625-fe691fe60434\") " Mar 13 12:45:51 crc kubenswrapper[4632]: I0313 12:45:51.657279 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nrqsw\" (UniqueName: \"kubernetes.io/projected/90e1bcb1-81b1-42c2-a625-fe691fe60434-kube-api-access-nrqsw\") pod \"90e1bcb1-81b1-42c2-a625-fe691fe60434\" (UID: \"90e1bcb1-81b1-42c2-a625-fe691fe60434\") " Mar 13 12:45:51 crc kubenswrapper[4632]: I0313 12:45:51.658390 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/90e1bcb1-81b1-42c2-a625-fe691fe60434-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:51 crc kubenswrapper[4632]: I0313 12:45:51.676359 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90e1bcb1-81b1-42c2-a625-fe691fe60434-kube-api-access-nrqsw" (OuterVolumeSpecName: "kube-api-access-nrqsw") pod "90e1bcb1-81b1-42c2-a625-fe691fe60434" (UID: "90e1bcb1-81b1-42c2-a625-fe691fe60434"). InnerVolumeSpecName "kube-api-access-nrqsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:45:51 crc kubenswrapper[4632]: I0313 12:45:51.683074 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/90e1bcb1-81b1-42c2-a625-fe691fe60434-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "90e1bcb1-81b1-42c2-a625-fe691fe60434" (UID: "90e1bcb1-81b1-42c2-a625-fe691fe60434"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:45:51 crc kubenswrapper[4632]: I0313 12:45:51.760762 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/90e1bcb1-81b1-42c2-a625-fe691fe60434-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:51 crc kubenswrapper[4632]: I0313 12:45:51.760804 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nrqsw\" (UniqueName: \"kubernetes.io/projected/90e1bcb1-81b1-42c2-a625-fe691fe60434-kube-api-access-nrqsw\") on node \"crc\" DevicePath \"\"" Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.001090 4632 generic.go:334] "Generic (PLEG): container finished" podID="90e1bcb1-81b1-42c2-a625-fe691fe60434" containerID="b40e35ab8281b9feedc9a3c0e3cb1babc5421097a7d08144dee8b68a590ba85f" exitCode=0 Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.001139 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbtjg" event={"ID":"90e1bcb1-81b1-42c2-a625-fe691fe60434","Type":"ContainerDied","Data":"b40e35ab8281b9feedc9a3c0e3cb1babc5421097a7d08144dee8b68a590ba85f"} Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.001173 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jbtjg" event={"ID":"90e1bcb1-81b1-42c2-a625-fe691fe60434","Type":"ContainerDied","Data":"7c9d33658cf93236a127bb3fe33f8e8067d984a174cb915b8ecf156bd22eb63f"} Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.001196 4632 scope.go:117] "RemoveContainer" containerID="b40e35ab8281b9feedc9a3c0e3cb1babc5421097a7d08144dee8b68a590ba85f" Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.003514 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jbtjg" Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.036082 4632 scope.go:117] "RemoveContainer" containerID="c40b8467f0fbf11bb8bc8bea09fc62d937cebf42c2af6303174926f814c485aa" Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.063258 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbtjg"] Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.072055 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jbtjg"] Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.101108 4632 scope.go:117] "RemoveContainer" containerID="a5eb8f564f7bb06a4d39ae737d43ff00b69617f7796a97fc2d384634398638cc" Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.125647 4632 scope.go:117] "RemoveContainer" containerID="b40e35ab8281b9feedc9a3c0e3cb1babc5421097a7d08144dee8b68a590ba85f" Mar 13 12:45:52 crc kubenswrapper[4632]: E0313 12:45:52.126320 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b40e35ab8281b9feedc9a3c0e3cb1babc5421097a7d08144dee8b68a590ba85f\": container with ID starting with b40e35ab8281b9feedc9a3c0e3cb1babc5421097a7d08144dee8b68a590ba85f not found: ID does not exist" containerID="b40e35ab8281b9feedc9a3c0e3cb1babc5421097a7d08144dee8b68a590ba85f" Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.126363 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b40e35ab8281b9feedc9a3c0e3cb1babc5421097a7d08144dee8b68a590ba85f"} err="failed to get container status \"b40e35ab8281b9feedc9a3c0e3cb1babc5421097a7d08144dee8b68a590ba85f\": rpc error: code = NotFound desc = could not find container \"b40e35ab8281b9feedc9a3c0e3cb1babc5421097a7d08144dee8b68a590ba85f\": container with ID starting with b40e35ab8281b9feedc9a3c0e3cb1babc5421097a7d08144dee8b68a590ba85f not found: ID does not exist" Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.126390 4632 scope.go:117] "RemoveContainer" containerID="c40b8467f0fbf11bb8bc8bea09fc62d937cebf42c2af6303174926f814c485aa" Mar 13 12:45:52 crc kubenswrapper[4632]: E0313 12:45:52.126613 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c40b8467f0fbf11bb8bc8bea09fc62d937cebf42c2af6303174926f814c485aa\": container with ID starting with c40b8467f0fbf11bb8bc8bea09fc62d937cebf42c2af6303174926f814c485aa not found: ID does not exist" containerID="c40b8467f0fbf11bb8bc8bea09fc62d937cebf42c2af6303174926f814c485aa" Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.126636 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c40b8467f0fbf11bb8bc8bea09fc62d937cebf42c2af6303174926f814c485aa"} err="failed to get container status \"c40b8467f0fbf11bb8bc8bea09fc62d937cebf42c2af6303174926f814c485aa\": rpc error: code = NotFound desc = could not find container \"c40b8467f0fbf11bb8bc8bea09fc62d937cebf42c2af6303174926f814c485aa\": container with ID starting with c40b8467f0fbf11bb8bc8bea09fc62d937cebf42c2af6303174926f814c485aa not found: ID does not exist" Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.126650 4632 scope.go:117] "RemoveContainer" containerID="a5eb8f564f7bb06a4d39ae737d43ff00b69617f7796a97fc2d384634398638cc" Mar 13 12:45:52 crc kubenswrapper[4632]: E0313 12:45:52.126842 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5eb8f564f7bb06a4d39ae737d43ff00b69617f7796a97fc2d384634398638cc\": container with ID starting with a5eb8f564f7bb06a4d39ae737d43ff00b69617f7796a97fc2d384634398638cc not found: ID does not exist" containerID="a5eb8f564f7bb06a4d39ae737d43ff00b69617f7796a97fc2d384634398638cc" Mar 13 12:45:52 crc kubenswrapper[4632]: I0313 12:45:52.126871 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5eb8f564f7bb06a4d39ae737d43ff00b69617f7796a97fc2d384634398638cc"} err="failed to get container status \"a5eb8f564f7bb06a4d39ae737d43ff00b69617f7796a97fc2d384634398638cc\": rpc error: code = NotFound desc = could not find container \"a5eb8f564f7bb06a4d39ae737d43ff00b69617f7796a97fc2d384634398638cc\": container with ID starting with a5eb8f564f7bb06a4d39ae737d43ff00b69617f7796a97fc2d384634398638cc not found: ID does not exist" Mar 13 12:45:54 crc kubenswrapper[4632]: I0313 12:45:54.060054 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90e1bcb1-81b1-42c2-a625-fe691fe60434" path="/var/lib/kubelet/pods/90e1bcb1-81b1-42c2-a625-fe691fe60434/volumes" Mar 13 12:45:57 crc kubenswrapper[4632]: I0313 12:45:57.241015 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-756c4b86c6-rm274_dbc1c989-5fa1-46dc-818e-8d609c069e34/barbican-api/0.log" Mar 13 12:45:57 crc kubenswrapper[4632]: I0313 12:45:57.464791 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-756c4b86c6-rm274_dbc1c989-5fa1-46dc-818e-8d609c069e34/barbican-api-log/0.log" Mar 13 12:45:57 crc kubenswrapper[4632]: I0313 12:45:57.622743 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6c97cdfb86-z2dqq_58332dcc-b1a6-4550-9c8b-8bbb82c04ff0/barbican-keystone-listener/0.log" Mar 13 12:45:57 crc kubenswrapper[4632]: I0313 12:45:57.922333 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-6c97cdfb86-z2dqq_58332dcc-b1a6-4550-9c8b-8bbb82c04ff0/barbican-keystone-listener-log/0.log" Mar 13 12:45:57 crc kubenswrapper[4632]: I0313 12:45:57.942375 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5fc9b6f5b5-6ps9m_51b847ef-ada2-456f-819d-0084fbb17185/barbican-worker-log/0.log" Mar 13 12:45:57 crc kubenswrapper[4632]: I0313 12:45:57.962260 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5fc9b6f5b5-6ps9m_51b847ef-ada2-456f-819d-0084fbb17185/barbican-worker/0.log" Mar 13 12:45:58 crc kubenswrapper[4632]: I0313 12:45:58.221484 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-fvfjp_684a2658-ba02-40cf-a371-ec2a8934c0d3/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:45:58 crc kubenswrapper[4632]: I0313 12:45:58.383415 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_046f071d-f091-4681-8a9b-06c7e7dc2192/ceilometer-central-agent/0.log" Mar 13 12:45:59 crc kubenswrapper[4632]: I0313 12:45:59.423601 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_046f071d-f091-4681-8a9b-06c7e7dc2192/ceilometer-notification-agent/0.log" Mar 13 12:45:59 crc kubenswrapper[4632]: I0313 12:45:59.470606 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_046f071d-f091-4681-8a9b-06c7e7dc2192/sg-core/0.log" Mar 13 12:45:59 crc kubenswrapper[4632]: I0313 12:45:59.554565 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_046f071d-f091-4681-8a9b-06c7e7dc2192/proxy-httpd/0.log" Mar 13 12:45:59 crc kubenswrapper[4632]: I0313 12:45:59.848193 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6785ba8c-a47b-4851-945e-c07ccecb9911/cinder-api/0.log" Mar 13 12:45:59 crc kubenswrapper[4632]: I0313 12:45:59.882927 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_6785ba8c-a47b-4851-945e-c07ccecb9911/cinder-api-log/0.log" Mar 13 12:45:59 crc kubenswrapper[4632]: I0313 12:45:59.988152 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d2c1c19b-95a5-4db1-8e54-36fe83704b25/cinder-scheduler/1.log" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.159837 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556766-lklf4"] Mar 13 12:46:00 crc kubenswrapper[4632]: E0313 12:46:00.160244 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90e1bcb1-81b1-42c2-a625-fe691fe60434" containerName="registry-server" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.160261 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="90e1bcb1-81b1-42c2-a625-fe691fe60434" containerName="registry-server" Mar 13 12:46:00 crc kubenswrapper[4632]: E0313 12:46:00.160279 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90e1bcb1-81b1-42c2-a625-fe691fe60434" containerName="extract-utilities" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.160286 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="90e1bcb1-81b1-42c2-a625-fe691fe60434" containerName="extract-utilities" Mar 13 12:46:00 crc kubenswrapper[4632]: E0313 12:46:00.160300 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90e1bcb1-81b1-42c2-a625-fe691fe60434" containerName="extract-content" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.160306 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="90e1bcb1-81b1-42c2-a625-fe691fe60434" containerName="extract-content" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.160503 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="90e1bcb1-81b1-42c2-a625-fe691fe60434" containerName="registry-server" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.161228 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556766-lklf4" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.163375 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.167296 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.167321 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.188134 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556766-lklf4"] Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.279874 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d2c1c19b-95a5-4db1-8e54-36fe83704b25/cinder-scheduler/0.log" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.320803 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ltmx4\" (UniqueName: \"kubernetes.io/projected/e38323af-ae58-48ce-979e-c8905218b4fe-kube-api-access-ltmx4\") pod \"auto-csr-approver-29556766-lklf4\" (UID: \"e38323af-ae58-48ce-979e-c8905218b4fe\") " pod="openshift-infra/auto-csr-approver-29556766-lklf4" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.400279 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_d2c1c19b-95a5-4db1-8e54-36fe83704b25/probe/0.log" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.422273 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ltmx4\" (UniqueName: \"kubernetes.io/projected/e38323af-ae58-48ce-979e-c8905218b4fe-kube-api-access-ltmx4\") pod \"auto-csr-approver-29556766-lklf4\" (UID: \"e38323af-ae58-48ce-979e-c8905218b4fe\") " pod="openshift-infra/auto-csr-approver-29556766-lklf4" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.430683 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-tpk84_bcd0e6df-81c2-4541-b0b5-d5c539f03451/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.760908 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ltmx4\" (UniqueName: \"kubernetes.io/projected/e38323af-ae58-48ce-979e-c8905218b4fe-kube-api-access-ltmx4\") pod \"auto-csr-approver-29556766-lklf4\" (UID: \"e38323af-ae58-48ce-979e-c8905218b4fe\") " pod="openshift-infra/auto-csr-approver-29556766-lklf4" Mar 13 12:46:00 crc kubenswrapper[4632]: I0313 12:46:00.785012 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556766-lklf4" Mar 13 12:46:01 crc kubenswrapper[4632]: I0313 12:46:01.049540 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-5fcgw_4931647b-bba4-489f-b5c1-cbe714834388/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:01 crc kubenswrapper[4632]: I0313 12:46:01.304281 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7b457785b5-7hzp6_1aca78bb-c923-4964-9b4c-5f7fb50badba/init/0.log" Mar 13 12:46:01 crc kubenswrapper[4632]: I0313 12:46:01.358451 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556766-lklf4"] Mar 13 12:46:01 crc kubenswrapper[4632]: I0313 12:46:01.594241 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7b457785b5-7hzp6_1aca78bb-c923-4964-9b4c-5f7fb50badba/init/0.log" Mar 13 12:46:01 crc kubenswrapper[4632]: I0313 12:46:01.717651 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-754cp_0d75181a-4c91-485e-8bcd-02e2aedd4d45/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:01 crc kubenswrapper[4632]: I0313 12:46:01.888536 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-7b457785b5-7hzp6_1aca78bb-c923-4964-9b4c-5f7fb50badba/dnsmasq-dns/0.log" Mar 13 12:46:02 crc kubenswrapper[4632]: I0313 12:46:02.052039 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_a2394af9-fd85-4291-8d57-c2bff02eccce/glance-httpd/0.log" Mar 13 12:46:02 crc kubenswrapper[4632]: I0313 12:46:02.097819 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556766-lklf4" event={"ID":"e38323af-ae58-48ce-979e-c8905218b4fe","Type":"ContainerStarted","Data":"9f75319750315cb6a2a63bbf725562683577f1266cb27af1c861282ce8fe51d0"} Mar 13 12:46:02 crc kubenswrapper[4632]: I0313 12:46:02.125650 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_a2394af9-fd85-4291-8d57-c2bff02eccce/glance-log/0.log" Mar 13 12:46:02 crc kubenswrapper[4632]: I0313 12:46:02.644579 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_97cf3e4a-cbe1-441c-8652-281a30fcf432/glance-httpd/0.log" Mar 13 12:46:02 crc kubenswrapper[4632]: I0313 12:46:02.846283 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_97cf3e4a-cbe1-441c-8652-281a30fcf432/glance-log/0.log" Mar 13 12:46:03 crc kubenswrapper[4632]: I0313 12:46:03.631709 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-c959f64fb-hx4t8_53145947-4584-4cef-b085-a0e0f550dde9/heat-engine/0.log" Mar 13 12:46:04 crc kubenswrapper[4632]: I0313 12:46:04.125294 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556766-lklf4" event={"ID":"e38323af-ae58-48ce-979e-c8905218b4fe","Type":"ContainerStarted","Data":"e2d392c178854d8d02c1d90a74a70ca0dce9ae28135802be619d355191eb7f40"} Mar 13 12:46:04 crc kubenswrapper[4632]: I0313 12:46:04.149167 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556766-lklf4" podStartSLOduration=2.9709213070000002 podStartE2EDuration="4.149147718s" podCreationTimestamp="2026-03-13 12:46:00 +0000 UTC" firstStartedPulling="2026-03-13 12:46:01.354593544 +0000 UTC m=+9735.377123667" lastFinishedPulling="2026-03-13 12:46:02.532819945 +0000 UTC m=+9736.555350078" observedRunningTime="2026-03-13 12:46:04.142988828 +0000 UTC m=+9738.165518971" watchObservedRunningTime="2026-03-13 12:46:04.149147718 +0000 UTC m=+9738.171677851" Mar 13 12:46:04 crc kubenswrapper[4632]: I0313 12:46:04.269549 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-689764498d-rg7vt_5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c/horizon/3.log" Mar 13 12:46:04 crc kubenswrapper[4632]: I0313 12:46:04.327138 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-689764498d-rg7vt_5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c/horizon/2.log" Mar 13 12:46:04 crc kubenswrapper[4632]: I0313 12:46:04.684653 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-n8wkn_41861d23-3e34-4f91-bafc-1b7eeee125db/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:04 crc kubenswrapper[4632]: I0313 12:46:04.754848 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-7fcc47f8dc-lhqhx_00b138c6-9e7c-4782-8454-1a4c035b1fbc/heat-api/0.log" Mar 13 12:46:05 crc kubenswrapper[4632]: I0313 12:46:05.077452 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-86bb565f45-ntq5k_de2e3cc7-c5cb-449a-a19c-2d671f08c656/heat-cfnapi/0.log" Mar 13 12:46:05 crc kubenswrapper[4632]: I0313 12:46:05.246961 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-srdvg_78de7f45-2a11-4cbe-84bf-46c4307a1459/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:05 crc kubenswrapper[4632]: I0313 12:46:05.413485 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29556661-2p4wf_6c20fa3e-2873-4076-b17a-3ee171199959/keystone-cron/0.log" Mar 13 12:46:05 crc kubenswrapper[4632]: I0313 12:46:05.619373 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-689764498d-rg7vt_5a03baf3-a8ba-4b13-9fa9-17eafe7b9b7c/horizon-log/0.log" Mar 13 12:46:05 crc kubenswrapper[4632]: I0313 12:46:05.649678 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29556721-l54tk_c8daf4c2-f012-4d18-b11a-e666e00d6a03/keystone-cron/0.log" Mar 13 12:46:06 crc kubenswrapper[4632]: I0313 12:46:06.154615 4632 generic.go:334] "Generic (PLEG): container finished" podID="e38323af-ae58-48ce-979e-c8905218b4fe" containerID="e2d392c178854d8d02c1d90a74a70ca0dce9ae28135802be619d355191eb7f40" exitCode=0 Mar 13 12:46:06 crc kubenswrapper[4632]: I0313 12:46:06.154658 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556766-lklf4" event={"ID":"e38323af-ae58-48ce-979e-c8905218b4fe","Type":"ContainerDied","Data":"e2d392c178854d8d02c1d90a74a70ca0dce9ae28135802be619d355191eb7f40"} Mar 13 12:46:06 crc kubenswrapper[4632]: I0313 12:46:06.187914 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_26ce3314-15f1-490c-83e5-a1c609212437/kube-state-metrics/0.log" Mar 13 12:46:06 crc kubenswrapper[4632]: I0313 12:46:06.326144 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-skjrh_ed1a2c50-a476-43ca-9764-e0ebffb14134/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:07 crc kubenswrapper[4632]: I0313 12:46:07.251269 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-695f666b49-nw48z_3a5c1185-e64b-44a9-b4b8-0108d4e80f9a/neutron-httpd/0.log" Mar 13 12:46:07 crc kubenswrapper[4632]: I0313 12:46:07.708549 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-c45jq_96e4ce1c-8f09-4563-864f-da1f95bdd500/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:07 crc kubenswrapper[4632]: I0313 12:46:07.733853 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556766-lklf4" Mar 13 12:46:07 crc kubenswrapper[4632]: I0313 12:46:07.920459 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ltmx4\" (UniqueName: \"kubernetes.io/projected/e38323af-ae58-48ce-979e-c8905218b4fe-kube-api-access-ltmx4\") pod \"e38323af-ae58-48ce-979e-c8905218b4fe\" (UID: \"e38323af-ae58-48ce-979e-c8905218b4fe\") " Mar 13 12:46:07 crc kubenswrapper[4632]: I0313 12:46:07.956674 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e38323af-ae58-48ce-979e-c8905218b4fe-kube-api-access-ltmx4" (OuterVolumeSpecName: "kube-api-access-ltmx4") pod "e38323af-ae58-48ce-979e-c8905218b4fe" (UID: "e38323af-ae58-48ce-979e-c8905218b4fe"). InnerVolumeSpecName "kube-api-access-ltmx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:46:08 crc kubenswrapper[4632]: I0313 12:46:08.022246 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ltmx4\" (UniqueName: \"kubernetes.io/projected/e38323af-ae58-48ce-979e-c8905218b4fe-kube-api-access-ltmx4\") on node \"crc\" DevicePath \"\"" Mar 13 12:46:08 crc kubenswrapper[4632]: I0313 12:46:08.163985 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-f664b756d-8fxf4_df64dbf7-8526-4fab-950a-4afefe47ec77/keystone-api/0.log" Mar 13 12:46:08 crc kubenswrapper[4632]: I0313 12:46:08.183813 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556766-lklf4" event={"ID":"e38323af-ae58-48ce-979e-c8905218b4fe","Type":"ContainerDied","Data":"9f75319750315cb6a2a63bbf725562683577f1266cb27af1c861282ce8fe51d0"} Mar 13 12:46:08 crc kubenswrapper[4632]: I0313 12:46:08.183850 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f75319750315cb6a2a63bbf725562683577f1266cb27af1c861282ce8fe51d0" Mar 13 12:46:08 crc kubenswrapper[4632]: I0313 12:46:08.183911 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556766-lklf4" Mar 13 12:46:08 crc kubenswrapper[4632]: I0313 12:46:08.321171 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556760-fxrw2"] Mar 13 12:46:08 crc kubenswrapper[4632]: I0313 12:46:08.373117 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556760-fxrw2"] Mar 13 12:46:09 crc kubenswrapper[4632]: I0313 12:46:09.083499 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-695f666b49-nw48z_3a5c1185-e64b-44a9-b4b8-0108d4e80f9a/neutron-api/0.log" Mar 13 12:46:09 crc kubenswrapper[4632]: I0313 12:46:09.237232 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_dbe53f0a-8bf3-4572-b5c8-01d5ed72c426/nova-cell0-conductor-conductor/0.log" Mar 13 12:46:09 crc kubenswrapper[4632]: I0313 12:46:09.788841 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_febcbdc5-25a6-46f7-8c06-d6f45624a466/nova-cell1-conductor-conductor/0.log" Mar 13 12:46:10 crc kubenswrapper[4632]: I0313 12:46:10.059414 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58d6a88c-498f-4887-998a-c3e3a1a2fef2" path="/var/lib/kubelet/pods/58d6a88c-498f-4887-998a-c3e3a1a2fef2/volumes" Mar 13 12:46:10 crc kubenswrapper[4632]: I0313 12:46:10.172155 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3ef77ea1-fee2-432d-9aba-c0acfedb4e69/nova-api-log/0.log" Mar 13 12:46:10 crc kubenswrapper[4632]: I0313 12:46:10.266089 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_bf01307f-1529-4aa7-95fc-8af84b061970/nova-cell1-novncproxy-novncproxy/0.log" Mar 13 12:46:10 crc kubenswrapper[4632]: I0313 12:46:10.467205 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:46:10 crc kubenswrapper[4632]: I0313 12:46:10.467535 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:46:10 crc kubenswrapper[4632]: I0313 12:46:10.467584 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 12:46:10 crc kubenswrapper[4632]: I0313 12:46:10.472559 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ac0aa587db0bc6f14a810b1c0a407933497eafb76c5051481d9814592d0380b3"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 12:46:10 crc kubenswrapper[4632]: I0313 12:46:10.472704 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://ac0aa587db0bc6f14a810b1c0a407933497eafb76c5051481d9814592d0380b3" gracePeriod=600 Mar 13 12:46:10 crc kubenswrapper[4632]: I0313 12:46:10.651702 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-dl4cq_c897af06-c467-4ec3-aa76-c29a3ea3a462/nova-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:10 crc kubenswrapper[4632]: I0313 12:46:10.721915 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b75084d0-782c-4f7e-8cc0-62ac424eec6f/nova-metadata-log/0.log" Mar 13 12:46:11 crc kubenswrapper[4632]: I0313 12:46:11.223356 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="ac0aa587db0bc6f14a810b1c0a407933497eafb76c5051481d9814592d0380b3" exitCode=0 Mar 13 12:46:11 crc kubenswrapper[4632]: I0313 12:46:11.223651 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"ac0aa587db0bc6f14a810b1c0a407933497eafb76c5051481d9814592d0380b3"} Mar 13 12:46:11 crc kubenswrapper[4632]: I0313 12:46:11.223712 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086"} Mar 13 12:46:11 crc kubenswrapper[4632]: I0313 12:46:11.223777 4632 scope.go:117] "RemoveContainer" containerID="23322afa11cd9b7f3f2b893b0662422a41461a3fe0777a0243c232f90f5a4eb9" Mar 13 12:46:11 crc kubenswrapper[4632]: I0313 12:46:11.533017 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1761ca69-46fd-4375-af60-22b3e77c19a2/mysql-bootstrap/0.log" Mar 13 12:46:11 crc kubenswrapper[4632]: I0313 12:46:11.720542 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_bd274a76-bf05-4f69-8d56-4844012a1fd1/nova-scheduler-scheduler/0.log" Mar 13 12:46:11 crc kubenswrapper[4632]: I0313 12:46:11.753663 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_3ef77ea1-fee2-432d-9aba-c0acfedb4e69/nova-api-api/0.log" Mar 13 12:46:11 crc kubenswrapper[4632]: I0313 12:46:11.835821 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1761ca69-46fd-4375-af60-22b3e77c19a2/mysql-bootstrap/0.log" Mar 13 12:46:12 crc kubenswrapper[4632]: I0313 12:46:12.132517 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1761ca69-46fd-4375-af60-22b3e77c19a2/galera/1.log" Mar 13 12:46:12 crc kubenswrapper[4632]: I0313 12:46:12.158035 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_1761ca69-46fd-4375-af60-22b3e77c19a2/galera/0.log" Mar 13 12:46:12 crc kubenswrapper[4632]: I0313 12:46:12.509736 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2cb2f546-c8c5-4ec9-aba8-d3782431de10/mysql-bootstrap/0.log" Mar 13 12:46:13 crc kubenswrapper[4632]: I0313 12:46:13.255316 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2cb2f546-c8c5-4ec9-aba8-d3782431de10/galera/1.log" Mar 13 12:46:13 crc kubenswrapper[4632]: I0313 12:46:13.293724 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2cb2f546-c8c5-4ec9-aba8-d3782431de10/galera/0.log" Mar 13 12:46:13 crc kubenswrapper[4632]: I0313 12:46:13.306010 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_2cb2f546-c8c5-4ec9-aba8-d3782431de10/mysql-bootstrap/0.log" Mar 13 12:46:13 crc kubenswrapper[4632]: I0313 12:46:13.600289 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_aef9680f-df77-4e2e-ac53-9d7530c2270c/openstackclient/0.log" Mar 13 12:46:13 crc kubenswrapper[4632]: I0313 12:46:13.960836 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-9kd7r_eab798dd-482a-4c66-983b-908966cd1f94/ovn-controller/0.log" Mar 13 12:46:14 crc kubenswrapper[4632]: I0313 12:46:14.095290 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-798sf_9246fc4f-3716-4a8b-9854-52137cf04e9a/openstack-network-exporter/0.log" Mar 13 12:46:14 crc kubenswrapper[4632]: I0313 12:46:14.291471 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c5xnp_d2677b19-4860-497e-a473-6d52d4901d8c/ovsdb-server-init/0.log" Mar 13 12:46:14 crc kubenswrapper[4632]: I0313 12:46:14.516689 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c5xnp_d2677b19-4860-497e-a473-6d52d4901d8c/ovsdb-server-init/0.log" Mar 13 12:46:14 crc kubenswrapper[4632]: I0313 12:46:14.990131 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c5xnp_d2677b19-4860-497e-a473-6d52d4901d8c/ovs-vswitchd/0.log" Mar 13 12:46:15 crc kubenswrapper[4632]: I0313 12:46:15.109875 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-c5xnp_d2677b19-4860-497e-a473-6d52d4901d8c/ovsdb-server/0.log" Mar 13 12:46:15 crc kubenswrapper[4632]: I0313 12:46:15.307256 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-6t9b6_96ca1247-6625-4b08-b155-34c56f02ec04/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:15 crc kubenswrapper[4632]: I0313 12:46:15.351989 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9a169306-9d47-41ae-8667-1efb89c43d82/openstack-network-exporter/0.log" Mar 13 12:46:15 crc kubenswrapper[4632]: I0313 12:46:15.576907 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_9a169306-9d47-41ae-8667-1efb89c43d82/ovn-northd/0.log" Mar 13 12:46:15 crc kubenswrapper[4632]: I0313 12:46:15.605539 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4ee148f1-cc66-4aa0-b603-c8a70f3554f5/openstack-network-exporter/0.log" Mar 13 12:46:15 crc kubenswrapper[4632]: I0313 12:46:15.858769 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_4ee148f1-cc66-4aa0-b603-c8a70f3554f5/ovsdbserver-nb/0.log" Mar 13 12:46:15 crc kubenswrapper[4632]: I0313 12:46:15.981535 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_b75084d0-782c-4f7e-8cc0-62ac424eec6f/nova-metadata-metadata/0.log" Mar 13 12:46:16 crc kubenswrapper[4632]: I0313 12:46:16.065062 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5529a725-48d8-4a60-91cd-775a4b520c20/openstack-network-exporter/0.log" Mar 13 12:46:16 crc kubenswrapper[4632]: I0313 12:46:16.203633 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_5529a725-48d8-4a60-91cd-775a4b520c20/ovsdbserver-sb/0.log" Mar 13 12:46:16 crc kubenswrapper[4632]: I0313 12:46:16.733008 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a3d80d9f-c956-40f5-b2e1-8aea2f136b6e/setup-container/0.log" Mar 13 12:46:16 crc kubenswrapper[4632]: I0313 12:46:16.913885 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6db55c595b-pwgcg_ab896d5b-a5b6-46a3-84d8-c3a8c968eac0/placement-api/0.log" Mar 13 12:46:16 crc kubenswrapper[4632]: I0313 12:46:16.966741 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a3d80d9f-c956-40f5-b2e1-8aea2f136b6e/setup-container/0.log" Mar 13 12:46:17 crc kubenswrapper[4632]: I0313 12:46:17.087062 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-6db55c595b-pwgcg_ab896d5b-a5b6-46a3-84d8-c3a8c968eac0/placement-log/0.log" Mar 13 12:46:17 crc kubenswrapper[4632]: I0313 12:46:17.099177 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_a3d80d9f-c956-40f5-b2e1-8aea2f136b6e/rabbitmq/0.log" Mar 13 12:46:17 crc kubenswrapper[4632]: I0313 12:46:17.321485 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e/setup-container/0.log" Mar 13 12:46:17 crc kubenswrapper[4632]: I0313 12:46:17.567344 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e/rabbitmq/0.log" Mar 13 12:46:17 crc kubenswrapper[4632]: I0313 12:46:17.573198 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_c8eb7c17-3ca6-4538-9d8b-b46cdfafb69e/setup-container/0.log" Mar 13 12:46:17 crc kubenswrapper[4632]: I0313 12:46:17.734779 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-fzxsz_4eaeef27-fa4c-41d9-a197-a780a6a6cebd/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:17 crc kubenswrapper[4632]: I0313 12:46:17.848746 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-5s64h_1dc9191f-32b9-45b9-b49f-fd704075f0a5/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:18 crc kubenswrapper[4632]: I0313 12:46:18.033804 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-r6nh9_0ea59acf-3206-492e-a7a8-bf855823d92c/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:18 crc kubenswrapper[4632]: I0313 12:46:18.391259 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-f9rbk_f69a3b21-eb1c-4300-91dc-55766900da95/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:18 crc kubenswrapper[4632]: I0313 12:46:18.423144 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-55n7g_9ff4122d-b9f1-4dd0-80dc-deb9d84760e1/ssh-known-hosts-edpm-deployment/0.log" Mar 13 12:46:18 crc kubenswrapper[4632]: I0313 12:46:18.931981 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7dbf8b9ddc-6p5vh_03ca050c-63a7-4b37-91fe-fe5c322cca78/proxy-server/0.log" Mar 13 12:46:18 crc kubenswrapper[4632]: I0313 12:46:18.988916 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-mkdcg_bc39c52e-008f-40c1-b93b-532707127fcd/swift-ring-rebalance/0.log" Mar 13 12:46:19 crc kubenswrapper[4632]: I0313 12:46:19.111320 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-7dbf8b9ddc-6p5vh_03ca050c-63a7-4b37-91fe-fe5c322cca78/proxy-httpd/0.log" Mar 13 12:46:19 crc kubenswrapper[4632]: I0313 12:46:19.232248 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/account-auditor/0.log" Mar 13 12:46:19 crc kubenswrapper[4632]: I0313 12:46:19.298281 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/account-reaper/0.log" Mar 13 12:46:19 crc kubenswrapper[4632]: I0313 12:46:19.500081 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/container-auditor/0.log" Mar 13 12:46:19 crc kubenswrapper[4632]: I0313 12:46:19.542089 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/account-replicator/0.log" Mar 13 12:46:19 crc kubenswrapper[4632]: I0313 12:46:19.624918 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/account-server/0.log" Mar 13 12:46:19 crc kubenswrapper[4632]: I0313 12:46:19.743504 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/container-replicator/0.log" Mar 13 12:46:19 crc kubenswrapper[4632]: I0313 12:46:19.812475 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/container-server/0.log" Mar 13 12:46:19 crc kubenswrapper[4632]: I0313 12:46:19.890260 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/container-updater/0.log" Mar 13 12:46:19 crc kubenswrapper[4632]: I0313 12:46:19.960829 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/object-auditor/0.log" Mar 13 12:46:20 crc kubenswrapper[4632]: I0313 12:46:20.070672 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/object-expirer/0.log" Mar 13 12:46:20 crc kubenswrapper[4632]: I0313 12:46:20.166386 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/object-server/0.log" Mar 13 12:46:20 crc kubenswrapper[4632]: I0313 12:46:20.212161 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/object-replicator/0.log" Mar 13 12:46:20 crc kubenswrapper[4632]: I0313 12:46:20.359209 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/object-updater/0.log" Mar 13 12:46:20 crc kubenswrapper[4632]: I0313 12:46:20.461026 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/swift-recon-cron/0.log" Mar 13 12:46:20 crc kubenswrapper[4632]: I0313 12:46:20.478179 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_e37b3d77-de2e-4be9-9984-550d4ba0f2f0/rsync/0.log" Mar 13 12:46:20 crc kubenswrapper[4632]: I0313 12:46:20.778308 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-tq6dw_4656b24f-4b10-481a-ba5b-1c17e5f2f7ef/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:21 crc kubenswrapper[4632]: I0313 12:46:21.095720 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s01-single-thread-testing_611401cc-04fe-4276-82fa-a896182802d4/tempest-tests-tempest-tests-runner/0.log" Mar 13 12:46:21 crc kubenswrapper[4632]: I0313 12:46:21.175676 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_c4836490-7b24-4245-bf50-7d590576f21e/test-operator-logs-container/0.log" Mar 13 12:46:21 crc kubenswrapper[4632]: I0313 12:46:21.279772 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s00-multi-thread-testing_a62e0eae-95dd-40a3-a489-80646fde4301/tempest-tests-tempest-tests-runner/0.log" Mar 13 12:46:21 crc kubenswrapper[4632]: I0313 12:46:21.485122 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-6w5gg_a1c30ff2-4a23-4fb1-b689-59318014bf57/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 13 12:46:30 crc kubenswrapper[4632]: I0313 12:46:30.364507 4632 scope.go:117] "RemoveContainer" containerID="a5bdb6d7b1972d01ea3faadd8b4d91d40f96718626d20034621dcf3eda3e5f37" Mar 13 12:46:38 crc kubenswrapper[4632]: I0313 12:46:38.838071 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_d9100748-6b15-4ccf-b961-aab1135f08d1/memcached/0.log" Mar 13 12:46:59 crc kubenswrapper[4632]: I0313 12:46:59.841929 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-wcgkr"] Mar 13 12:46:59 crc kubenswrapper[4632]: E0313 12:46:59.842927 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e38323af-ae58-48ce-979e-c8905218b4fe" containerName="oc" Mar 13 12:46:59 crc kubenswrapper[4632]: I0313 12:46:59.842963 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="e38323af-ae58-48ce-979e-c8905218b4fe" containerName="oc" Mar 13 12:46:59 crc kubenswrapper[4632]: I0313 12:46:59.843178 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="e38323af-ae58-48ce-979e-c8905218b4fe" containerName="oc" Mar 13 12:46:59 crc kubenswrapper[4632]: I0313 12:46:59.848006 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:46:59 crc kubenswrapper[4632]: I0313 12:46:59.884285 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wcgkr"] Mar 13 12:46:59 crc kubenswrapper[4632]: I0313 12:46:59.990662 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/964e423d-b9e8-4f29-af5d-84b106ae8159-catalog-content\") pod \"redhat-operators-wcgkr\" (UID: \"964e423d-b9e8-4f29-af5d-84b106ae8159\") " pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:46:59 crc kubenswrapper[4632]: I0313 12:46:59.990809 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/964e423d-b9e8-4f29-af5d-84b106ae8159-utilities\") pod \"redhat-operators-wcgkr\" (UID: \"964e423d-b9e8-4f29-af5d-84b106ae8159\") " pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:46:59 crc kubenswrapper[4632]: I0313 12:46:59.990844 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4frc4\" (UniqueName: \"kubernetes.io/projected/964e423d-b9e8-4f29-af5d-84b106ae8159-kube-api-access-4frc4\") pod \"redhat-operators-wcgkr\" (UID: \"964e423d-b9e8-4f29-af5d-84b106ae8159\") " pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:47:00 crc kubenswrapper[4632]: I0313 12:47:00.104109 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4frc4\" (UniqueName: \"kubernetes.io/projected/964e423d-b9e8-4f29-af5d-84b106ae8159-kube-api-access-4frc4\") pod \"redhat-operators-wcgkr\" (UID: \"964e423d-b9e8-4f29-af5d-84b106ae8159\") " pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:47:00 crc kubenswrapper[4632]: I0313 12:47:00.104328 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/964e423d-b9e8-4f29-af5d-84b106ae8159-catalog-content\") pod \"redhat-operators-wcgkr\" (UID: \"964e423d-b9e8-4f29-af5d-84b106ae8159\") " pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:47:00 crc kubenswrapper[4632]: I0313 12:47:00.104746 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/964e423d-b9e8-4f29-af5d-84b106ae8159-utilities\") pod \"redhat-operators-wcgkr\" (UID: \"964e423d-b9e8-4f29-af5d-84b106ae8159\") " pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:47:00 crc kubenswrapper[4632]: I0313 12:47:00.105505 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/964e423d-b9e8-4f29-af5d-84b106ae8159-utilities\") pod \"redhat-operators-wcgkr\" (UID: \"964e423d-b9e8-4f29-af5d-84b106ae8159\") " pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:47:00 crc kubenswrapper[4632]: I0313 12:47:00.105932 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/964e423d-b9e8-4f29-af5d-84b106ae8159-catalog-content\") pod \"redhat-operators-wcgkr\" (UID: \"964e423d-b9e8-4f29-af5d-84b106ae8159\") " pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:47:00 crc kubenswrapper[4632]: I0313 12:47:00.144970 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4frc4\" (UniqueName: \"kubernetes.io/projected/964e423d-b9e8-4f29-af5d-84b106ae8159-kube-api-access-4frc4\") pod \"redhat-operators-wcgkr\" (UID: \"964e423d-b9e8-4f29-af5d-84b106ae8159\") " pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:47:00 crc kubenswrapper[4632]: I0313 12:47:00.217578 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:47:00 crc kubenswrapper[4632]: I0313 12:47:00.342806 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m_13abf84a-b499-4439-ab4e-1c34bcf07308/util/0.log" Mar 13 12:47:00 crc kubenswrapper[4632]: I0313 12:47:00.523090 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m_13abf84a-b499-4439-ab4e-1c34bcf07308/util/0.log" Mar 13 12:47:00 crc kubenswrapper[4632]: I0313 12:47:00.820222 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m_13abf84a-b499-4439-ab4e-1c34bcf07308/pull/0.log" Mar 13 12:47:00 crc kubenswrapper[4632]: I0313 12:47:00.953319 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m_13abf84a-b499-4439-ab4e-1c34bcf07308/pull/0.log" Mar 13 12:47:01 crc kubenswrapper[4632]: I0313 12:47:01.004445 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-wcgkr"] Mar 13 12:47:01 crc kubenswrapper[4632]: I0313 12:47:01.181617 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m_13abf84a-b499-4439-ab4e-1c34bcf07308/util/0.log" Mar 13 12:47:01 crc kubenswrapper[4632]: I0313 12:47:01.326302 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m_13abf84a-b499-4439-ab4e-1c34bcf07308/pull/0.log" Mar 13 12:47:01 crc kubenswrapper[4632]: I0313 12:47:01.522536 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cdce3512d6dd40d0fa1f5a9460ef5ddb632791dec16c7770c73458169cskc7m_13abf84a-b499-4439-ab4e-1c34bcf07308/extract/0.log" Mar 13 12:47:01 crc kubenswrapper[4632]: I0313 12:47:01.756130 4632 generic.go:334] "Generic (PLEG): container finished" podID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerID="4814d54398dddc7491e0a4c9f868011d9742ed90e2d05a245b23e35c28be791e" exitCode=0 Mar 13 12:47:01 crc kubenswrapper[4632]: I0313 12:47:01.756170 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wcgkr" event={"ID":"964e423d-b9e8-4f29-af5d-84b106ae8159","Type":"ContainerDied","Data":"4814d54398dddc7491e0a4c9f868011d9742ed90e2d05a245b23e35c28be791e"} Mar 13 12:47:01 crc kubenswrapper[4632]: I0313 12:47:01.756197 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wcgkr" event={"ID":"964e423d-b9e8-4f29-af5d-84b106ae8159","Type":"ContainerStarted","Data":"38d0595c24aa0a1cd0e1355e6ab1c56e237ac066c9da90382b7cbb6e3fba6db4"} Mar 13 12:47:02 crc kubenswrapper[4632]: I0313 12:47:02.293316 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-66d56f6ff4-cfcgn_75d652c7-8521-4039-913a-fa625f89b094/manager/0.log" Mar 13 12:47:03 crc kubenswrapper[4632]: I0313 12:47:03.065068 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5964f64c48-qg79l_20f92131-aca4-41ea-9144-a23bd9216f49/manager/0.log" Mar 13 12:47:03 crc kubenswrapper[4632]: I0313 12:47:03.784800 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wcgkr" event={"ID":"964e423d-b9e8-4f29-af5d-84b106ae8159","Type":"ContainerStarted","Data":"219da7f0798a57241a68dc972a8e6cf63665a59509f96ff776be2e82e493c3c5"} Mar 13 12:47:03 crc kubenswrapper[4632]: I0313 12:47:03.960400 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-77b6666d85-cgh6c_ff6d4dcb-9eb8-44fc-951e-f2aecd77a639/manager/0.log" Mar 13 12:47:04 crc kubenswrapper[4632]: I0313 12:47:04.760448 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d9d6b584d-2rv7s_9a963f9c-ac58-4e21-abfa-fca1279a192d/manager/0.log" Mar 13 12:47:06 crc kubenswrapper[4632]: I0313 12:47:06.153128 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6bbb499bbc-wtzrw_c8fc6f03-c43b-4ade-92a8-acc5537a4eeb/manager/0.log" Mar 13 12:47:06 crc kubenswrapper[4632]: I0313 12:47:06.580293 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5995f4446f-flfxh_1542a9c8-92f6-4bc9-8231-829f649b0b8f/manager/0.log" Mar 13 12:47:06 crc kubenswrapper[4632]: I0313 12:47:06.957581 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-984cd4dcf-f6c87_3f3a462e-4d89-45b3-8611-181aca5f8558/manager/0.log" Mar 13 12:47:06 crc kubenswrapper[4632]: I0313 12:47:06.974825 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-684f77d66d-6nb82_f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda/manager/1.log" Mar 13 12:47:07 crc kubenswrapper[4632]: I0313 12:47:07.142238 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-684f77d66d-6nb82_f0be4a6b-e3ac-4141-b5bf-b3fafcca5fda/manager/0.log" Mar 13 12:47:07 crc kubenswrapper[4632]: I0313 12:47:07.579928 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-68f45f9d9f-sxw8d_7b491335-6a73-46de-8098-f27ff4c6f795/manager/0.log" Mar 13 12:47:08 crc kubenswrapper[4632]: I0313 12:47:08.001848 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-658d4cdd5-szd7c_9040a0e0-2a56-4331-ba50-b19ff05ef0c0/manager/0.log" Mar 13 12:47:08 crc kubenswrapper[4632]: I0313 12:47:08.193520 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-776c5696bf-bkmbn_c33d0da9-5a04-42d6-80d3-2f558b4a90b0/manager/0.log" Mar 13 12:47:08 crc kubenswrapper[4632]: I0313 12:47:08.609551 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5f4f55cb5c-62gpm_9e1d6ac6-c4ee-4381-86b8-c337f8c2d6a5/manager/0.log" Mar 13 12:47:08 crc kubenswrapper[4632]: I0313 12:47:08.679318 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-569cc54c5-628ss_d04e9aa6-f234-4ffa-81e2-1a2407addb77/manager/0.log" Mar 13 12:47:08 crc kubenswrapper[4632]: I0313 12:47:08.868022 4632 generic.go:334] "Generic (PLEG): container finished" podID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerID="219da7f0798a57241a68dc972a8e6cf63665a59509f96ff776be2e82e493c3c5" exitCode=0 Mar 13 12:47:08 crc kubenswrapper[4632]: I0313 12:47:08.868069 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wcgkr" event={"ID":"964e423d-b9e8-4f29-af5d-84b106ae8159","Type":"ContainerDied","Data":"219da7f0798a57241a68dc972a8e6cf63665a59509f96ff776be2e82e493c3c5"} Mar 13 12:47:08 crc kubenswrapper[4632]: I0313 12:47:08.911837 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-557ccf57b7v927j_2d221857-ee77-4165-a351-ecd5fc424970/manager/0.log" Mar 13 12:47:09 crc kubenswrapper[4632]: I0313 12:47:09.702639 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-865685cd99-ls9jq_82fe7ef6-50a5-41d4-9419-787812e16bd6/operator/0.log" Mar 13 12:47:09 crc kubenswrapper[4632]: I0313 12:47:09.870017 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-2jqnk_7de02b7f-4e1c-4ba1-9659-c864e9080092/registry-server/1.log" Mar 13 12:47:09 crc kubenswrapper[4632]: I0313 12:47:09.890435 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wcgkr" event={"ID":"964e423d-b9e8-4f29-af5d-84b106ae8159","Type":"ContainerStarted","Data":"537f35c912c9c8057ecc3ec80663f0a0d3c386360eb374b59f3e50a3f8bd59ee"} Mar 13 12:47:09 crc kubenswrapper[4632]: I0313 12:47:09.916121 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-wcgkr" podStartSLOduration=3.329566062 podStartE2EDuration="10.916101569s" podCreationTimestamp="2026-03-13 12:46:59 +0000 UTC" firstStartedPulling="2026-03-13 12:47:01.760069158 +0000 UTC m=+9795.782599291" lastFinishedPulling="2026-03-13 12:47:09.346604665 +0000 UTC m=+9803.369134798" observedRunningTime="2026-03-13 12:47:09.911336962 +0000 UTC m=+9803.933867105" watchObservedRunningTime="2026-03-13 12:47:09.916101569 +0000 UTC m=+9803.938631702" Mar 13 12:47:09 crc kubenswrapper[4632]: I0313 12:47:09.966344 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-2jqnk_7de02b7f-4e1c-4ba1-9659-c864e9080092/registry-server/0.log" Mar 13 12:47:10 crc kubenswrapper[4632]: I0313 12:47:10.218507 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:47:10 crc kubenswrapper[4632]: I0313 12:47:10.218572 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:47:10 crc kubenswrapper[4632]: I0313 12:47:10.407047 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-bbc5b68f9-4m8kf_0a9d48f4-d68b-4ef9-826e-ed619c761405/manager/0.log" Mar 13 12:47:10 crc kubenswrapper[4632]: I0313 12:47:10.974538 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-574d45c66c-qkr9n_e66fe20e-05b5-42cd-ac1d-bc4eaee4c8e5/manager/0.log" Mar 13 12:47:11 crc kubenswrapper[4632]: I0313 12:47:11.276240 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wcgkr" podUID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerName="registry-server" probeResult="failure" output=< Mar 13 12:47:11 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:47:11 crc kubenswrapper[4632]: > Mar 13 12:47:11 crc kubenswrapper[4632]: I0313 12:47:11.283256 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-2lzt8_daba1153-3b28-4234-8dd0-ec20160abbfe/operator/0.log" Mar 13 12:47:11 crc kubenswrapper[4632]: I0313 12:47:11.558423 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-677c674df7-qbfg2_2d8a9f3a-6631-4c1e-8381-3bc313837ca0/manager/0.log" Mar 13 12:47:11 crc kubenswrapper[4632]: I0313 12:47:11.824736 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-677bd678f7-wj9qs_68c5eb80-4214-42c5-a08d-de6012969621/manager/0.log" Mar 13 12:47:12 crc kubenswrapper[4632]: I0313 12:47:12.131198 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-6cd66dbd4b-nt7np_ee081327-4c3f-4c0a-9085-71085c6487b5/manager/0.log" Mar 13 12:47:12 crc kubenswrapper[4632]: I0313 12:47:12.257137 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-jwrgq_7bab78c8-7dac-48dc-a426-ccd4ae00a428/manager/0.log" Mar 13 12:47:12 crc kubenswrapper[4632]: I0313 12:47:12.408415 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-85c677895b-thbc4_3fdb377f-5a78-4687-82e1-50718514290d/manager/0.log" Mar 13 12:47:12 crc kubenswrapper[4632]: I0313 12:47:12.502302 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6dd88c6f67-kv8b2_e0d1d349-d63d-498b-ae15-3121f9ae73f8/manager/0.log" Mar 13 12:47:18 crc kubenswrapper[4632]: I0313 12:47:18.289894 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-677bd678f7-wj9qs_68c5eb80-4214-42c5-a08d-de6012969621/manager/1.log" Mar 13 12:47:21 crc kubenswrapper[4632]: I0313 12:47:21.269244 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wcgkr" podUID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerName="registry-server" probeResult="failure" output=< Mar 13 12:47:21 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:47:21 crc kubenswrapper[4632]: > Mar 13 12:47:31 crc kubenswrapper[4632]: I0313 12:47:31.275643 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wcgkr" podUID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerName="registry-server" probeResult="failure" output=< Mar 13 12:47:31 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:47:31 crc kubenswrapper[4632]: > Mar 13 12:47:39 crc kubenswrapper[4632]: I0313 12:47:39.810107 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-pvwll_2332524f-f990-4ef2-90b3-8b90c389d873/control-plane-machine-set-operator/0.log" Mar 13 12:47:40 crc kubenswrapper[4632]: I0313 12:47:40.347407 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c6jnc_275c3112-6912-49f8-9d3f-8147662fb99f/kube-rbac-proxy/0.log" Mar 13 12:47:40 crc kubenswrapper[4632]: I0313 12:47:40.348769 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-c6jnc_275c3112-6912-49f8-9d3f-8147662fb99f/machine-api-operator/0.log" Mar 13 12:47:41 crc kubenswrapper[4632]: I0313 12:47:41.283430 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wcgkr" podUID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerName="registry-server" probeResult="failure" output=< Mar 13 12:47:41 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:47:41 crc kubenswrapper[4632]: > Mar 13 12:47:51 crc kubenswrapper[4632]: I0313 12:47:51.596910 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-wcgkr" podUID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerName="registry-server" probeResult="failure" output=< Mar 13 12:47:51 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:47:51 crc kubenswrapper[4632]: > Mar 13 12:47:57 crc kubenswrapper[4632]: I0313 12:47:57.340618 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-kh4n9_43729a96-008f-4af6-ba0d-d52f2f179c0b/cert-manager-controller/0.log" Mar 13 12:47:57 crc kubenswrapper[4632]: I0313 12:47:57.598231 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-xg2df_348f2814-4e97-4ec5-bcbb-35a868955687/cert-manager-cainjector/0.log" Mar 13 12:47:57 crc kubenswrapper[4632]: I0313 12:47:57.667583 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-tjkbb_a0d52d98-fe87-4bc8-890e-5c5efb1f30d6/cert-manager-webhook/0.log" Mar 13 12:48:00 crc kubenswrapper[4632]: I0313 12:48:00.302845 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:48:00 crc kubenswrapper[4632]: I0313 12:48:00.406264 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:48:00 crc kubenswrapper[4632]: I0313 12:48:00.464411 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556768-qx4vs"] Mar 13 12:48:00 crc kubenswrapper[4632]: I0313 12:48:00.483366 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556768-qx4vs" Mar 13 12:48:00 crc kubenswrapper[4632]: I0313 12:48:00.488156 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556768-qx4vs"] Mar 13 12:48:00 crc kubenswrapper[4632]: I0313 12:48:00.503203 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:48:00 crc kubenswrapper[4632]: I0313 12:48:00.503169 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:48:00 crc kubenswrapper[4632]: I0313 12:48:00.503426 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:48:00 crc kubenswrapper[4632]: I0313 12:48:00.607695 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wcgkr"] Mar 13 12:48:00 crc kubenswrapper[4632]: I0313 12:48:00.630283 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsv78\" (UniqueName: \"kubernetes.io/projected/7c0a936c-9a75-4f0b-81b1-fb7f74d9911f-kube-api-access-xsv78\") pod \"auto-csr-approver-29556768-qx4vs\" (UID: \"7c0a936c-9a75-4f0b-81b1-fb7f74d9911f\") " pod="openshift-infra/auto-csr-approver-29556768-qx4vs" Mar 13 12:48:00 crc kubenswrapper[4632]: I0313 12:48:00.732012 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsv78\" (UniqueName: \"kubernetes.io/projected/7c0a936c-9a75-4f0b-81b1-fb7f74d9911f-kube-api-access-xsv78\") pod \"auto-csr-approver-29556768-qx4vs\" (UID: \"7c0a936c-9a75-4f0b-81b1-fb7f74d9911f\") " pod="openshift-infra/auto-csr-approver-29556768-qx4vs" Mar 13 12:48:01 crc kubenswrapper[4632]: I0313 12:48:01.365799 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsv78\" (UniqueName: \"kubernetes.io/projected/7c0a936c-9a75-4f0b-81b1-fb7f74d9911f-kube-api-access-xsv78\") pod \"auto-csr-approver-29556768-qx4vs\" (UID: \"7c0a936c-9a75-4f0b-81b1-fb7f74d9911f\") " pod="openshift-infra/auto-csr-approver-29556768-qx4vs" Mar 13 12:48:01 crc kubenswrapper[4632]: I0313 12:48:01.425310 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556768-qx4vs" Mar 13 12:48:01 crc kubenswrapper[4632]: I0313 12:48:01.426187 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-wcgkr" podUID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerName="registry-server" containerID="cri-o://537f35c912c9c8057ecc3ec80663f0a0d3c386360eb374b59f3e50a3f8bd59ee" gracePeriod=2 Mar 13 12:48:02 crc kubenswrapper[4632]: I0313 12:48:02.437410 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wcgkr" event={"ID":"964e423d-b9e8-4f29-af5d-84b106ae8159","Type":"ContainerDied","Data":"537f35c912c9c8057ecc3ec80663f0a0d3c386360eb374b59f3e50a3f8bd59ee"} Mar 13 12:48:02 crc kubenswrapper[4632]: I0313 12:48:02.438248 4632 generic.go:334] "Generic (PLEG): container finished" podID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerID="537f35c912c9c8057ecc3ec80663f0a0d3c386360eb374b59f3e50a3f8bd59ee" exitCode=0 Mar 13 12:48:02 crc kubenswrapper[4632]: I0313 12:48:02.941584 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556768-qx4vs"] Mar 13 12:48:03 crc kubenswrapper[4632]: I0313 12:48:03.286334 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 12:48:03 crc kubenswrapper[4632]: I0313 12:48:03.449892 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556768-qx4vs" event={"ID":"7c0a936c-9a75-4f0b-81b1-fb7f74d9911f","Type":"ContainerStarted","Data":"d34eb978076dffd0023033a08fbd76563d781881681a241dbff3f3a50aa7ac78"} Mar 13 12:48:03 crc kubenswrapper[4632]: I0313 12:48:03.454241 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-wcgkr" event={"ID":"964e423d-b9e8-4f29-af5d-84b106ae8159","Type":"ContainerDied","Data":"38d0595c24aa0a1cd0e1355e6ab1c56e237ac066c9da90382b7cbb6e3fba6db4"} Mar 13 12:48:03 crc kubenswrapper[4632]: I0313 12:48:03.454862 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="38d0595c24aa0a1cd0e1355e6ab1c56e237ac066c9da90382b7cbb6e3fba6db4" Mar 13 12:48:03 crc kubenswrapper[4632]: I0313 12:48:03.485535 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:48:03 crc kubenswrapper[4632]: I0313 12:48:03.491874 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4frc4\" (UniqueName: \"kubernetes.io/projected/964e423d-b9e8-4f29-af5d-84b106ae8159-kube-api-access-4frc4\") pod \"964e423d-b9e8-4f29-af5d-84b106ae8159\" (UID: \"964e423d-b9e8-4f29-af5d-84b106ae8159\") " Mar 13 12:48:03 crc kubenswrapper[4632]: I0313 12:48:03.491986 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/964e423d-b9e8-4f29-af5d-84b106ae8159-utilities\") pod \"964e423d-b9e8-4f29-af5d-84b106ae8159\" (UID: \"964e423d-b9e8-4f29-af5d-84b106ae8159\") " Mar 13 12:48:03 crc kubenswrapper[4632]: I0313 12:48:03.492081 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/964e423d-b9e8-4f29-af5d-84b106ae8159-catalog-content\") pod \"964e423d-b9e8-4f29-af5d-84b106ae8159\" (UID: \"964e423d-b9e8-4f29-af5d-84b106ae8159\") " Mar 13 12:48:03 crc kubenswrapper[4632]: I0313 12:48:03.494499 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/964e423d-b9e8-4f29-af5d-84b106ae8159-utilities" (OuterVolumeSpecName: "utilities") pod "964e423d-b9e8-4f29-af5d-84b106ae8159" (UID: "964e423d-b9e8-4f29-af5d-84b106ae8159"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:48:03 crc kubenswrapper[4632]: I0313 12:48:03.510177 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/964e423d-b9e8-4f29-af5d-84b106ae8159-kube-api-access-4frc4" (OuterVolumeSpecName: "kube-api-access-4frc4") pod "964e423d-b9e8-4f29-af5d-84b106ae8159" (UID: "964e423d-b9e8-4f29-af5d-84b106ae8159"). InnerVolumeSpecName "kube-api-access-4frc4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:48:03 crc kubenswrapper[4632]: I0313 12:48:03.598584 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4frc4\" (UniqueName: \"kubernetes.io/projected/964e423d-b9e8-4f29-af5d-84b106ae8159-kube-api-access-4frc4\") on node \"crc\" DevicePath \"\"" Mar 13 12:48:03 crc kubenswrapper[4632]: I0313 12:48:03.598824 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/964e423d-b9e8-4f29-af5d-84b106ae8159-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:48:03 crc kubenswrapper[4632]: I0313 12:48:03.657731 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/964e423d-b9e8-4f29-af5d-84b106ae8159-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "964e423d-b9e8-4f29-af5d-84b106ae8159" (UID: "964e423d-b9e8-4f29-af5d-84b106ae8159"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:48:03 crc kubenswrapper[4632]: I0313 12:48:03.700999 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/964e423d-b9e8-4f29-af5d-84b106ae8159-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:48:04 crc kubenswrapper[4632]: I0313 12:48:04.462315 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-wcgkr" Mar 13 12:48:04 crc kubenswrapper[4632]: I0313 12:48:04.504074 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-wcgkr"] Mar 13 12:48:04 crc kubenswrapper[4632]: I0313 12:48:04.517652 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-wcgkr"] Mar 13 12:48:05 crc kubenswrapper[4632]: I0313 12:48:05.473688 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556768-qx4vs" event={"ID":"7c0a936c-9a75-4f0b-81b1-fb7f74d9911f","Type":"ContainerStarted","Data":"fbdf7412c66e2fa539b75629b05076618e8fad2c845d05a467fd575d619baa55"} Mar 13 12:48:05 crc kubenswrapper[4632]: I0313 12:48:05.494457 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556768-qx4vs" podStartSLOduration=4.593300179 podStartE2EDuration="5.493558994s" podCreationTimestamp="2026-03-13 12:48:00 +0000 UTC" firstStartedPulling="2026-03-13 12:48:03.280838457 +0000 UTC m=+9857.303368580" lastFinishedPulling="2026-03-13 12:48:04.181097262 +0000 UTC m=+9858.203627395" observedRunningTime="2026-03-13 12:48:05.489484404 +0000 UTC m=+9859.512014537" watchObservedRunningTime="2026-03-13 12:48:05.493558994 +0000 UTC m=+9859.516089137" Mar 13 12:48:06 crc kubenswrapper[4632]: I0313 12:48:06.058140 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="964e423d-b9e8-4f29-af5d-84b106ae8159" path="/var/lib/kubelet/pods/964e423d-b9e8-4f29-af5d-84b106ae8159/volumes" Mar 13 12:48:06 crc kubenswrapper[4632]: I0313 12:48:06.485146 4632 generic.go:334] "Generic (PLEG): container finished" podID="7c0a936c-9a75-4f0b-81b1-fb7f74d9911f" containerID="fbdf7412c66e2fa539b75629b05076618e8fad2c845d05a467fd575d619baa55" exitCode=0 Mar 13 12:48:06 crc kubenswrapper[4632]: I0313 12:48:06.485208 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556768-qx4vs" event={"ID":"7c0a936c-9a75-4f0b-81b1-fb7f74d9911f","Type":"ContainerDied","Data":"fbdf7412c66e2fa539b75629b05076618e8fad2c845d05a467fd575d619baa55"} Mar 13 12:48:07 crc kubenswrapper[4632]: I0313 12:48:07.903816 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556768-qx4vs" Mar 13 12:48:08 crc kubenswrapper[4632]: I0313 12:48:08.074357 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsv78\" (UniqueName: \"kubernetes.io/projected/7c0a936c-9a75-4f0b-81b1-fb7f74d9911f-kube-api-access-xsv78\") pod \"7c0a936c-9a75-4f0b-81b1-fb7f74d9911f\" (UID: \"7c0a936c-9a75-4f0b-81b1-fb7f74d9911f\") " Mar 13 12:48:08 crc kubenswrapper[4632]: I0313 12:48:08.095912 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c0a936c-9a75-4f0b-81b1-fb7f74d9911f-kube-api-access-xsv78" (OuterVolumeSpecName: "kube-api-access-xsv78") pod "7c0a936c-9a75-4f0b-81b1-fb7f74d9911f" (UID: "7c0a936c-9a75-4f0b-81b1-fb7f74d9911f"). InnerVolumeSpecName "kube-api-access-xsv78". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:48:08 crc kubenswrapper[4632]: I0313 12:48:08.176733 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsv78\" (UniqueName: \"kubernetes.io/projected/7c0a936c-9a75-4f0b-81b1-fb7f74d9911f-kube-api-access-xsv78\") on node \"crc\" DevicePath \"\"" Mar 13 12:48:08 crc kubenswrapper[4632]: I0313 12:48:08.504208 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556768-qx4vs" event={"ID":"7c0a936c-9a75-4f0b-81b1-fb7f74d9911f","Type":"ContainerDied","Data":"d34eb978076dffd0023033a08fbd76563d781881681a241dbff3f3a50aa7ac78"} Mar 13 12:48:08 crc kubenswrapper[4632]: I0313 12:48:08.504546 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d34eb978076dffd0023033a08fbd76563d781881681a241dbff3f3a50aa7ac78" Mar 13 12:48:08 crc kubenswrapper[4632]: I0313 12:48:08.504281 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556768-qx4vs" Mar 13 12:48:08 crc kubenswrapper[4632]: I0313 12:48:08.579926 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556762-mjtxh"] Mar 13 12:48:08 crc kubenswrapper[4632]: I0313 12:48:08.589524 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556762-mjtxh"] Mar 13 12:48:10 crc kubenswrapper[4632]: I0313 12:48:10.057545 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae" path="/var/lib/kubelet/pods/cf13c17d-1ea6-4a0e-bfbd-e3bfc8d453ae/volumes" Mar 13 12:48:10 crc kubenswrapper[4632]: I0313 12:48:10.460917 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:48:10 crc kubenswrapper[4632]: I0313 12:48:10.461669 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:48:16 crc kubenswrapper[4632]: I0313 12:48:16.492471 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-kzrvn_1ca5cae6-5549-492a-a257-745bb41d3574/nmstate-console-plugin/0.log" Mar 13 12:48:16 crc kubenswrapper[4632]: I0313 12:48:16.794970 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-mpfnk_33445a2b-7fa8-4198-a60a-09caeb69b8ed/nmstate-handler/0.log" Mar 13 12:48:16 crc kubenswrapper[4632]: I0313 12:48:16.990905 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-lnfrw_0c63c4bc-5c1a-4af0-b255-eb418d8a02cd/kube-rbac-proxy/0.log" Mar 13 12:48:17 crc kubenswrapper[4632]: I0313 12:48:17.032507 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-lnfrw_0c63c4bc-5c1a-4af0-b255-eb418d8a02cd/nmstate-metrics/0.log" Mar 13 12:48:17 crc kubenswrapper[4632]: I0313 12:48:17.159502 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-bzmdv_3b679db2-06cc-4796-945a-5ced45b39053/nmstate-operator/0.log" Mar 13 12:48:17 crc kubenswrapper[4632]: I0313 12:48:17.283645 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-gcngd_9bf11778-d854-4c97-acd1-ed4822ee5f47/nmstate-webhook/0.log" Mar 13 12:48:30 crc kubenswrapper[4632]: I0313 12:48:30.594131 4632 scope.go:117] "RemoveContainer" containerID="8dfadb29bc36e882b8d8ebf6016fec294107233b5f8602de74595b7d612d371c" Mar 13 12:48:40 crc kubenswrapper[4632]: I0313 12:48:40.463153 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:48:40 crc kubenswrapper[4632]: I0313 12:48:40.467579 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:48:53 crc kubenswrapper[4632]: I0313 12:48:53.628247 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-62bwr_277ddd7f-fd9c-4b27-9563-c904f1dffd40/controller/0.log" Mar 13 12:48:53 crc kubenswrapper[4632]: I0313 12:48:53.653932 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-62bwr_277ddd7f-fd9c-4b27-9563-c904f1dffd40/kube-rbac-proxy/0.log" Mar 13 12:48:53 crc kubenswrapper[4632]: I0313 12:48:53.921397 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/cp-frr-files/0.log" Mar 13 12:48:54 crc kubenswrapper[4632]: I0313 12:48:54.154237 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/cp-frr-files/0.log" Mar 13 12:48:54 crc kubenswrapper[4632]: I0313 12:48:54.178001 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/cp-metrics/0.log" Mar 13 12:48:54 crc kubenswrapper[4632]: I0313 12:48:54.205110 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/cp-reloader/0.log" Mar 13 12:48:54 crc kubenswrapper[4632]: I0313 12:48:54.214783 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/cp-reloader/0.log" Mar 13 12:48:54 crc kubenswrapper[4632]: I0313 12:48:54.710685 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/cp-metrics/0.log" Mar 13 12:48:54 crc kubenswrapper[4632]: I0313 12:48:54.754788 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/cp-frr-files/0.log" Mar 13 12:48:54 crc kubenswrapper[4632]: I0313 12:48:54.787283 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/cp-reloader/0.log" Mar 13 12:48:54 crc kubenswrapper[4632]: I0313 12:48:54.855837 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/cp-metrics/0.log" Mar 13 12:48:55 crc kubenswrapper[4632]: I0313 12:48:55.081317 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/cp-frr-files/0.log" Mar 13 12:48:55 crc kubenswrapper[4632]: I0313 12:48:55.169279 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/cp-reloader/0.log" Mar 13 12:48:55 crc kubenswrapper[4632]: I0313 12:48:55.232813 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/controller/1.log" Mar 13 12:48:55 crc kubenswrapper[4632]: I0313 12:48:55.310028 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/cp-metrics/0.log" Mar 13 12:48:55 crc kubenswrapper[4632]: I0313 12:48:55.469839 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/controller/0.log" Mar 13 12:48:55 crc kubenswrapper[4632]: I0313 12:48:55.635915 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/frr-metrics/0.log" Mar 13 12:48:55 crc kubenswrapper[4632]: I0313 12:48:55.883806 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/kube-rbac-proxy/0.log" Mar 13 12:48:55 crc kubenswrapper[4632]: I0313 12:48:55.995441 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/kube-rbac-proxy-frr/0.log" Mar 13 12:48:56 crc kubenswrapper[4632]: I0313 12:48:56.369821 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/reloader/0.log" Mar 13 12:48:56 crc kubenswrapper[4632]: I0313 12:48:56.775950 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-9zbh8_b33bccd8-6f28-4ffe-9500-069a52aab5df/frr-k8s-webhook-server/0.log" Mar 13 12:48:56 crc kubenswrapper[4632]: I0313 12:48:56.782750 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-9zbh8_b33bccd8-6f28-4ffe-9500-069a52aab5df/frr-k8s-webhook-server/1.log" Mar 13 12:48:57 crc kubenswrapper[4632]: I0313 12:48:57.275921 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-ffdcc767b-qxvlq_e62d674f-5b2c-4788-85a3-95b51621dbef/manager/0.log" Mar 13 12:48:57 crc kubenswrapper[4632]: I0313 12:48:57.621626 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-6c7bf5ddc5-v6t5l_712b2002-4fce-4983-926a-99a4b2dc7a8c/webhook-server/0.log" Mar 13 12:48:57 crc kubenswrapper[4632]: I0313 12:48:57.931766 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tztd9_8f51973a-596d-40dc-9b5b-b2c95a60ea0c/kube-rbac-proxy/0.log" Mar 13 12:48:57 crc kubenswrapper[4632]: I0313 12:48:57.974576 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/frr/1.log" Mar 13 12:48:58 crc kubenswrapper[4632]: I0313 12:48:58.528718 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tztd9_8f51973a-596d-40dc-9b5b-b2c95a60ea0c/speaker/1.log" Mar 13 12:48:58 crc kubenswrapper[4632]: I0313 12:48:58.862732 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-lvlxj_85b58bb0-63f5-4c85-8759-ce28d2c7db58/frr/0.log" Mar 13 12:48:58 crc kubenswrapper[4632]: I0313 12:48:58.885215 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-tztd9_8f51973a-596d-40dc-9b5b-b2c95a60ea0c/speaker/0.log" Mar 13 12:49:10 crc kubenswrapper[4632]: I0313 12:49:10.460857 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:49:10 crc kubenswrapper[4632]: I0313 12:49:10.461329 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:49:10 crc kubenswrapper[4632]: I0313 12:49:10.462849 4632 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" Mar 13 12:49:10 crc kubenswrapper[4632]: I0313 12:49:10.466132 4632 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086"} pod="openshift-machine-config-operator/machine-config-daemon-zkscb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 12:49:10 crc kubenswrapper[4632]: I0313 12:49:10.466214 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" containerID="cri-o://f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" gracePeriod=600 Mar 13 12:49:10 crc kubenswrapper[4632]: E0313 12:49:10.608650 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:49:11 crc kubenswrapper[4632]: I0313 12:49:11.131972 4632 generic.go:334] "Generic (PLEG): container finished" podID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" exitCode=0 Mar 13 12:49:11 crc kubenswrapper[4632]: I0313 12:49:11.132306 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerDied","Data":"f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086"} Mar 13 12:49:11 crc kubenswrapper[4632]: I0313 12:49:11.133059 4632 scope.go:117] "RemoveContainer" containerID="ac0aa587db0bc6f14a810b1c0a407933497eafb76c5051481d9814592d0380b3" Mar 13 12:49:11 crc kubenswrapper[4632]: I0313 12:49:11.133481 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:49:11 crc kubenswrapper[4632]: E0313 12:49:11.133754 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:49:16 crc kubenswrapper[4632]: I0313 12:49:16.321881 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg_2e270cfe-55fc-4855-87ff-4313a0ad319c/util/0.log" Mar 13 12:49:16 crc kubenswrapper[4632]: I0313 12:49:16.602175 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg_2e270cfe-55fc-4855-87ff-4313a0ad319c/util/0.log" Mar 13 12:49:16 crc kubenswrapper[4632]: I0313 12:49:16.622254 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg_2e270cfe-55fc-4855-87ff-4313a0ad319c/pull/0.log" Mar 13 12:49:16 crc kubenswrapper[4632]: I0313 12:49:16.675583 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg_2e270cfe-55fc-4855-87ff-4313a0ad319c/pull/0.log" Mar 13 12:49:16 crc kubenswrapper[4632]: I0313 12:49:16.891193 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg_2e270cfe-55fc-4855-87ff-4313a0ad319c/util/0.log" Mar 13 12:49:16 crc kubenswrapper[4632]: I0313 12:49:16.945130 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg_2e270cfe-55fc-4855-87ff-4313a0ad319c/pull/0.log" Mar 13 12:49:16 crc kubenswrapper[4632]: I0313 12:49:16.970364 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874pdtfg_2e270cfe-55fc-4855-87ff-4313a0ad319c/extract/0.log" Mar 13 12:49:17 crc kubenswrapper[4632]: I0313 12:49:17.156687 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj_8c1e4d78-3f38-48b5-b157-a1a076f31b76/util/0.log" Mar 13 12:49:17 crc kubenswrapper[4632]: I0313 12:49:17.431413 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj_8c1e4d78-3f38-48b5-b157-a1a076f31b76/util/0.log" Mar 13 12:49:17 crc kubenswrapper[4632]: I0313 12:49:17.500222 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj_8c1e4d78-3f38-48b5-b157-a1a076f31b76/pull/0.log" Mar 13 12:49:17 crc kubenswrapper[4632]: I0313 12:49:17.500674 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj_8c1e4d78-3f38-48b5-b157-a1a076f31b76/pull/0.log" Mar 13 12:49:17 crc kubenswrapper[4632]: I0313 12:49:17.782904 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj_8c1e4d78-3f38-48b5-b157-a1a076f31b76/util/0.log" Mar 13 12:49:17 crc kubenswrapper[4632]: I0313 12:49:17.839569 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj_8c1e4d78-3f38-48b5-b157-a1a076f31b76/pull/0.log" Mar 13 12:49:17 crc kubenswrapper[4632]: I0313 12:49:17.871513 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1n2dcj_8c1e4d78-3f38-48b5-b157-a1a076f31b76/extract/0.log" Mar 13 12:49:18 crc kubenswrapper[4632]: I0313 12:49:18.009043 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7ksc5_0fa3faab-9e82-4fde-afff-3de6939a17d1/extract-utilities/0.log" Mar 13 12:49:18 crc kubenswrapper[4632]: I0313 12:49:18.299557 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7ksc5_0fa3faab-9e82-4fde-afff-3de6939a17d1/extract-utilities/0.log" Mar 13 12:49:18 crc kubenswrapper[4632]: I0313 12:49:18.325003 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7ksc5_0fa3faab-9e82-4fde-afff-3de6939a17d1/extract-content/0.log" Mar 13 12:49:18 crc kubenswrapper[4632]: I0313 12:49:18.325018 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7ksc5_0fa3faab-9e82-4fde-afff-3de6939a17d1/extract-content/0.log" Mar 13 12:49:18 crc kubenswrapper[4632]: I0313 12:49:18.504286 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7ksc5_0fa3faab-9e82-4fde-afff-3de6939a17d1/extract-utilities/0.log" Mar 13 12:49:18 crc kubenswrapper[4632]: I0313 12:49:18.557308 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7ksc5_0fa3faab-9e82-4fde-afff-3de6939a17d1/extract-content/0.log" Mar 13 12:49:18 crc kubenswrapper[4632]: I0313 12:49:18.955748 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c87c2_7bbc76a6-d812-41c7-a63b-09f6fdb37405/extract-utilities/0.log" Mar 13 12:49:19 crc kubenswrapper[4632]: I0313 12:49:19.155654 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c87c2_7bbc76a6-d812-41c7-a63b-09f6fdb37405/extract-utilities/0.log" Mar 13 12:49:19 crc kubenswrapper[4632]: I0313 12:49:19.213287 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c87c2_7bbc76a6-d812-41c7-a63b-09f6fdb37405/extract-content/0.log" Mar 13 12:49:19 crc kubenswrapper[4632]: I0313 12:49:19.401114 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c87c2_7bbc76a6-d812-41c7-a63b-09f6fdb37405/extract-content/0.log" Mar 13 12:49:19 crc kubenswrapper[4632]: I0313 12:49:19.676072 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c87c2_7bbc76a6-d812-41c7-a63b-09f6fdb37405/extract-utilities/0.log" Mar 13 12:49:19 crc kubenswrapper[4632]: I0313 12:49:19.687851 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c87c2_7bbc76a6-d812-41c7-a63b-09f6fdb37405/extract-content/0.log" Mar 13 12:49:19 crc kubenswrapper[4632]: I0313 12:49:19.944439 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-7ksc5_0fa3faab-9e82-4fde-afff-3de6939a17d1/registry-server/0.log" Mar 13 12:49:20 crc kubenswrapper[4632]: I0313 12:49:19.995090 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-d9n25_023be687-a773-401c-981b-e3d7136f53b6/marketplace-operator/0.log" Mar 13 12:49:20 crc kubenswrapper[4632]: I0313 12:49:20.460368 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-c87c2_7bbc76a6-d812-41c7-a63b-09f6fdb37405/registry-server/0.log" Mar 13 12:49:20 crc kubenswrapper[4632]: I0313 12:49:20.470462 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gdt8x_f7f61b75-16bf-4c5a-be30-c88d155c203f/extract-utilities/0.log" Mar 13 12:49:20 crc kubenswrapper[4632]: I0313 12:49:20.998284 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gdt8x_f7f61b75-16bf-4c5a-be30-c88d155c203f/extract-utilities/0.log" Mar 13 12:49:21 crc kubenswrapper[4632]: I0313 12:49:21.008168 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gdt8x_f7f61b75-16bf-4c5a-be30-c88d155c203f/extract-content/0.log" Mar 13 12:49:21 crc kubenswrapper[4632]: I0313 12:49:21.044377 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gdt8x_f7f61b75-16bf-4c5a-be30-c88d155c203f/extract-content/0.log" Mar 13 12:49:21 crc kubenswrapper[4632]: I0313 12:49:21.205122 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gdt8x_f7f61b75-16bf-4c5a-be30-c88d155c203f/extract-utilities/0.log" Mar 13 12:49:21 crc kubenswrapper[4632]: I0313 12:49:21.257334 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gdt8x_f7f61b75-16bf-4c5a-be30-c88d155c203f/extract-content/0.log" Mar 13 12:49:21 crc kubenswrapper[4632]: I0313 12:49:21.527329 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-gdt8x_f7f61b75-16bf-4c5a-be30-c88d155c203f/registry-server/0.log" Mar 13 12:49:21 crc kubenswrapper[4632]: I0313 12:49:21.596371 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vwgfr_f5cc71d2-1901-4778-8e20-93646cfc1a85/extract-utilities/0.log" Mar 13 12:49:21 crc kubenswrapper[4632]: I0313 12:49:21.823865 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vwgfr_f5cc71d2-1901-4778-8e20-93646cfc1a85/extract-content/0.log" Mar 13 12:49:21 crc kubenswrapper[4632]: I0313 12:49:21.842252 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vwgfr_f5cc71d2-1901-4778-8e20-93646cfc1a85/extract-utilities/0.log" Mar 13 12:49:21 crc kubenswrapper[4632]: I0313 12:49:21.877350 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vwgfr_f5cc71d2-1901-4778-8e20-93646cfc1a85/extract-content/0.log" Mar 13 12:49:22 crc kubenswrapper[4632]: I0313 12:49:22.073498 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vwgfr_f5cc71d2-1901-4778-8e20-93646cfc1a85/extract-utilities/0.log" Mar 13 12:49:22 crc kubenswrapper[4632]: I0313 12:49:22.079469 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vwgfr_f5cc71d2-1901-4778-8e20-93646cfc1a85/extract-content/0.log" Mar 13 12:49:23 crc kubenswrapper[4632]: I0313 12:49:23.645645 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vwgfr_f5cc71d2-1901-4778-8e20-93646cfc1a85/registry-server/0.log" Mar 13 12:49:24 crc kubenswrapper[4632]: I0313 12:49:24.045628 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:49:24 crc kubenswrapper[4632]: E0313 12:49:24.046175 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:49:38 crc kubenswrapper[4632]: I0313 12:49:38.055964 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:49:38 crc kubenswrapper[4632]: E0313 12:49:38.056806 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:49:49 crc kubenswrapper[4632]: I0313 12:49:49.044697 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:49:49 crc kubenswrapper[4632]: E0313 12:49:49.045424 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:49:50 crc kubenswrapper[4632]: E0313 12:49:50.109786 4632 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.182:37506->38.102.83.182:37465: read tcp 38.102.83.182:37506->38.102.83.182:37465: read: connection reset by peer Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.617120 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556770-clndp"] Mar 13 12:50:00 crc kubenswrapper[4632]: E0313 12:50:00.623152 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerName="extract-utilities" Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.623190 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerName="extract-utilities" Mar 13 12:50:00 crc kubenswrapper[4632]: E0313 12:50:00.623219 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerName="registry-server" Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.623226 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerName="registry-server" Mar 13 12:50:00 crc kubenswrapper[4632]: E0313 12:50:00.623239 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c0a936c-9a75-4f0b-81b1-fb7f74d9911f" containerName="oc" Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.623245 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c0a936c-9a75-4f0b-81b1-fb7f74d9911f" containerName="oc" Mar 13 12:50:00 crc kubenswrapper[4632]: E0313 12:50:00.623272 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerName="extract-content" Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.623278 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerName="extract-content" Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.625552 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="964e423d-b9e8-4f29-af5d-84b106ae8159" containerName="registry-server" Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.625585 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c0a936c-9a75-4f0b-81b1-fb7f74d9911f" containerName="oc" Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.633739 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556770-clndp" Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.651074 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.651074 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.651083 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.739389 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556770-clndp"] Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.781364 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xgqb\" (UniqueName: \"kubernetes.io/projected/40d97b55-9e4c-4d9a-962c-4030dc7dd36b-kube-api-access-6xgqb\") pod \"auto-csr-approver-29556770-clndp\" (UID: \"40d97b55-9e4c-4d9a-962c-4030dc7dd36b\") " pod="openshift-infra/auto-csr-approver-29556770-clndp" Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.882748 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xgqb\" (UniqueName: \"kubernetes.io/projected/40d97b55-9e4c-4d9a-962c-4030dc7dd36b-kube-api-access-6xgqb\") pod \"auto-csr-approver-29556770-clndp\" (UID: \"40d97b55-9e4c-4d9a-962c-4030dc7dd36b\") " pod="openshift-infra/auto-csr-approver-29556770-clndp" Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.932314 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xgqb\" (UniqueName: \"kubernetes.io/projected/40d97b55-9e4c-4d9a-962c-4030dc7dd36b-kube-api-access-6xgqb\") pod \"auto-csr-approver-29556770-clndp\" (UID: \"40d97b55-9e4c-4d9a-962c-4030dc7dd36b\") " pod="openshift-infra/auto-csr-approver-29556770-clndp" Mar 13 12:50:00 crc kubenswrapper[4632]: I0313 12:50:00.964411 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556770-clndp" Mar 13 12:50:02 crc kubenswrapper[4632]: I0313 12:50:02.044893 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:50:02 crc kubenswrapper[4632]: E0313 12:50:02.046389 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:50:02 crc kubenswrapper[4632]: I0313 12:50:02.644927 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556770-clndp"] Mar 13 12:50:03 crc kubenswrapper[4632]: I0313 12:50:03.617382 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556770-clndp" event={"ID":"40d97b55-9e4c-4d9a-962c-4030dc7dd36b","Type":"ContainerStarted","Data":"17f3bfa72055148fc6070a539aaf1077a5482350320ffbe2a1d181f4064e1eb0"} Mar 13 12:50:05 crc kubenswrapper[4632]: I0313 12:50:05.644976 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556770-clndp" event={"ID":"40d97b55-9e4c-4d9a-962c-4030dc7dd36b","Type":"ContainerStarted","Data":"3a1743912c3055d81c796b10e35eb2de0472d47edbe9b1cc8de66ceb54f1f127"} Mar 13 12:50:05 crc kubenswrapper[4632]: I0313 12:50:05.677810 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556770-clndp" podStartSLOduration=4.535975405 podStartE2EDuration="5.676281276s" podCreationTimestamp="2026-03-13 12:50:00 +0000 UTC" firstStartedPulling="2026-03-13 12:50:02.693812234 +0000 UTC m=+9976.716342367" lastFinishedPulling="2026-03-13 12:50:03.834118105 +0000 UTC m=+9977.856648238" observedRunningTime="2026-03-13 12:50:05.670982216 +0000 UTC m=+9979.693512349" watchObservedRunningTime="2026-03-13 12:50:05.676281276 +0000 UTC m=+9979.698811419" Mar 13 12:50:06 crc kubenswrapper[4632]: I0313 12:50:06.656051 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556770-clndp" event={"ID":"40d97b55-9e4c-4d9a-962c-4030dc7dd36b","Type":"ContainerDied","Data":"3a1743912c3055d81c796b10e35eb2de0472d47edbe9b1cc8de66ceb54f1f127"} Mar 13 12:50:06 crc kubenswrapper[4632]: I0313 12:50:06.658371 4632 generic.go:334] "Generic (PLEG): container finished" podID="40d97b55-9e4c-4d9a-962c-4030dc7dd36b" containerID="3a1743912c3055d81c796b10e35eb2de0472d47edbe9b1cc8de66ceb54f1f127" exitCode=0 Mar 13 12:50:08 crc kubenswrapper[4632]: I0313 12:50:08.076567 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556770-clndp" Mar 13 12:50:08 crc kubenswrapper[4632]: I0313 12:50:08.142201 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xgqb\" (UniqueName: \"kubernetes.io/projected/40d97b55-9e4c-4d9a-962c-4030dc7dd36b-kube-api-access-6xgqb\") pod \"40d97b55-9e4c-4d9a-962c-4030dc7dd36b\" (UID: \"40d97b55-9e4c-4d9a-962c-4030dc7dd36b\") " Mar 13 12:50:08 crc kubenswrapper[4632]: I0313 12:50:08.175118 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40d97b55-9e4c-4d9a-962c-4030dc7dd36b-kube-api-access-6xgqb" (OuterVolumeSpecName: "kube-api-access-6xgqb") pod "40d97b55-9e4c-4d9a-962c-4030dc7dd36b" (UID: "40d97b55-9e4c-4d9a-962c-4030dc7dd36b"). InnerVolumeSpecName "kube-api-access-6xgqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:50:08 crc kubenswrapper[4632]: I0313 12:50:08.244828 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xgqb\" (UniqueName: \"kubernetes.io/projected/40d97b55-9e4c-4d9a-962c-4030dc7dd36b-kube-api-access-6xgqb\") on node \"crc\" DevicePath \"\"" Mar 13 12:50:08 crc kubenswrapper[4632]: I0313 12:50:08.679018 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556770-clndp" Mar 13 12:50:08 crc kubenswrapper[4632]: I0313 12:50:08.678932 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556770-clndp" event={"ID":"40d97b55-9e4c-4d9a-962c-4030dc7dd36b","Type":"ContainerDied","Data":"17f3bfa72055148fc6070a539aaf1077a5482350320ffbe2a1d181f4064e1eb0"} Mar 13 12:50:08 crc kubenswrapper[4632]: I0313 12:50:08.680685 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17f3bfa72055148fc6070a539aaf1077a5482350320ffbe2a1d181f4064e1eb0" Mar 13 12:50:09 crc kubenswrapper[4632]: I0313 12:50:09.175927 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556764-9lxh8"] Mar 13 12:50:09 crc kubenswrapper[4632]: I0313 12:50:09.185734 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556764-9lxh8"] Mar 13 12:50:10 crc kubenswrapper[4632]: I0313 12:50:10.059366 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="437f55ff-c573-4944-a680-6ac2d168cb0f" path="/var/lib/kubelet/pods/437f55ff-c573-4944-a680-6ac2d168cb0f/volumes" Mar 13 12:50:16 crc kubenswrapper[4632]: I0313 12:50:16.044220 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:50:16 crc kubenswrapper[4632]: E0313 12:50:16.045660 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:50:27 crc kubenswrapper[4632]: I0313 12:50:27.046570 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:50:27 crc kubenswrapper[4632]: E0313 12:50:27.047872 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:50:30 crc kubenswrapper[4632]: I0313 12:50:30.939549 4632 scope.go:117] "RemoveContainer" containerID="aa02f726269acb1a95d7a68005cbfbbe4f481bb481f4612470d155ee5bde6649" Mar 13 12:50:39 crc kubenswrapper[4632]: I0313 12:50:39.044422 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:50:39 crc kubenswrapper[4632]: E0313 12:50:39.045250 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:50:50 crc kubenswrapper[4632]: I0313 12:50:50.045075 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:50:50 crc kubenswrapper[4632]: E0313 12:50:50.047325 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:51:01 crc kubenswrapper[4632]: I0313 12:51:01.045696 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:51:01 crc kubenswrapper[4632]: E0313 12:51:01.046616 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:51:13 crc kubenswrapper[4632]: I0313 12:51:13.046033 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:51:13 crc kubenswrapper[4632]: E0313 12:51:13.046819 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:51:27 crc kubenswrapper[4632]: I0313 12:51:27.044665 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:51:27 crc kubenswrapper[4632]: E0313 12:51:27.045307 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:51:31 crc kubenswrapper[4632]: I0313 12:51:31.130526 4632 scope.go:117] "RemoveContainer" containerID="f9576f389f75db79c6cd02f685bff29c0e4ed007b62591df661e2a8ee57c8ce2" Mar 13 12:51:41 crc kubenswrapper[4632]: I0313 12:51:41.045732 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:51:41 crc kubenswrapper[4632]: E0313 12:51:41.046650 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:51:56 crc kubenswrapper[4632]: I0313 12:51:56.044870 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:51:56 crc kubenswrapper[4632]: E0313 12:51:56.045927 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:52:00 crc kubenswrapper[4632]: I0313 12:52:00.171600 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556772-dvdjm"] Mar 13 12:52:00 crc kubenswrapper[4632]: E0313 12:52:00.172656 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40d97b55-9e4c-4d9a-962c-4030dc7dd36b" containerName="oc" Mar 13 12:52:00 crc kubenswrapper[4632]: I0313 12:52:00.172675 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="40d97b55-9e4c-4d9a-962c-4030dc7dd36b" containerName="oc" Mar 13 12:52:00 crc kubenswrapper[4632]: I0313 12:52:00.172887 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="40d97b55-9e4c-4d9a-962c-4030dc7dd36b" containerName="oc" Mar 13 12:52:00 crc kubenswrapper[4632]: I0313 12:52:00.173597 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556772-dvdjm" Mar 13 12:52:00 crc kubenswrapper[4632]: I0313 12:52:00.177194 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:52:00 crc kubenswrapper[4632]: I0313 12:52:00.177732 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:52:00 crc kubenswrapper[4632]: I0313 12:52:00.178006 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:52:00 crc kubenswrapper[4632]: I0313 12:52:00.181602 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556772-dvdjm"] Mar 13 12:52:00 crc kubenswrapper[4632]: I0313 12:52:00.367309 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmdn2\" (UniqueName: \"kubernetes.io/projected/576ef56c-ba4f-4def-89ac-cdef2e378fca-kube-api-access-wmdn2\") pod \"auto-csr-approver-29556772-dvdjm\" (UID: \"576ef56c-ba4f-4def-89ac-cdef2e378fca\") " pod="openshift-infra/auto-csr-approver-29556772-dvdjm" Mar 13 12:52:00 crc kubenswrapper[4632]: I0313 12:52:00.469882 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmdn2\" (UniqueName: \"kubernetes.io/projected/576ef56c-ba4f-4def-89ac-cdef2e378fca-kube-api-access-wmdn2\") pod \"auto-csr-approver-29556772-dvdjm\" (UID: \"576ef56c-ba4f-4def-89ac-cdef2e378fca\") " pod="openshift-infra/auto-csr-approver-29556772-dvdjm" Mar 13 12:52:00 crc kubenswrapper[4632]: I0313 12:52:00.504168 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmdn2\" (UniqueName: \"kubernetes.io/projected/576ef56c-ba4f-4def-89ac-cdef2e378fca-kube-api-access-wmdn2\") pod \"auto-csr-approver-29556772-dvdjm\" (UID: \"576ef56c-ba4f-4def-89ac-cdef2e378fca\") " pod="openshift-infra/auto-csr-approver-29556772-dvdjm" Mar 13 12:52:00 crc kubenswrapper[4632]: I0313 12:52:00.796742 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556772-dvdjm" Mar 13 12:52:01 crc kubenswrapper[4632]: I0313 12:52:01.307485 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556772-dvdjm"] Mar 13 12:52:01 crc kubenswrapper[4632]: I0313 12:52:01.882703 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556772-dvdjm" event={"ID":"576ef56c-ba4f-4def-89ac-cdef2e378fca","Type":"ContainerStarted","Data":"ac34c41e4a723fd479a80798f16607c4694c8f642f8c95c8503d5d4d0b6c9f9f"} Mar 13 12:52:03 crc kubenswrapper[4632]: I0313 12:52:03.903208 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556772-dvdjm" event={"ID":"576ef56c-ba4f-4def-89ac-cdef2e378fca","Type":"ContainerStarted","Data":"2288f2f8e699c9708c03d351054c645fb4cb8dd9cc206c490b95db693eefe035"} Mar 13 12:52:03 crc kubenswrapper[4632]: I0313 12:52:03.922898 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556772-dvdjm" podStartSLOduration=2.539646749 podStartE2EDuration="3.922875079s" podCreationTimestamp="2026-03-13 12:52:00 +0000 UTC" firstStartedPulling="2026-03-13 12:52:01.303055858 +0000 UTC m=+10095.325585991" lastFinishedPulling="2026-03-13 12:52:02.686284188 +0000 UTC m=+10096.708814321" observedRunningTime="2026-03-13 12:52:03.913623843 +0000 UTC m=+10097.936153986" watchObservedRunningTime="2026-03-13 12:52:03.922875079 +0000 UTC m=+10097.945405212" Mar 13 12:52:04 crc kubenswrapper[4632]: I0313 12:52:04.916186 4632 generic.go:334] "Generic (PLEG): container finished" podID="576ef56c-ba4f-4def-89ac-cdef2e378fca" containerID="2288f2f8e699c9708c03d351054c645fb4cb8dd9cc206c490b95db693eefe035" exitCode=0 Mar 13 12:52:04 crc kubenswrapper[4632]: I0313 12:52:04.916273 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556772-dvdjm" event={"ID":"576ef56c-ba4f-4def-89ac-cdef2e378fca","Type":"ContainerDied","Data":"2288f2f8e699c9708c03d351054c645fb4cb8dd9cc206c490b95db693eefe035"} Mar 13 12:52:06 crc kubenswrapper[4632]: I0313 12:52:06.400798 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556772-dvdjm" Mar 13 12:52:06 crc kubenswrapper[4632]: I0313 12:52:06.415669 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmdn2\" (UniqueName: \"kubernetes.io/projected/576ef56c-ba4f-4def-89ac-cdef2e378fca-kube-api-access-wmdn2\") pod \"576ef56c-ba4f-4def-89ac-cdef2e378fca\" (UID: \"576ef56c-ba4f-4def-89ac-cdef2e378fca\") " Mar 13 12:52:06 crc kubenswrapper[4632]: I0313 12:52:06.423201 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/576ef56c-ba4f-4def-89ac-cdef2e378fca-kube-api-access-wmdn2" (OuterVolumeSpecName: "kube-api-access-wmdn2") pod "576ef56c-ba4f-4def-89ac-cdef2e378fca" (UID: "576ef56c-ba4f-4def-89ac-cdef2e378fca"). InnerVolumeSpecName "kube-api-access-wmdn2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:52:06 crc kubenswrapper[4632]: I0313 12:52:06.518981 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmdn2\" (UniqueName: \"kubernetes.io/projected/576ef56c-ba4f-4def-89ac-cdef2e378fca-kube-api-access-wmdn2\") on node \"crc\" DevicePath \"\"" Mar 13 12:52:06 crc kubenswrapper[4632]: I0313 12:52:06.951842 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556772-dvdjm" event={"ID":"576ef56c-ba4f-4def-89ac-cdef2e378fca","Type":"ContainerDied","Data":"ac34c41e4a723fd479a80798f16607c4694c8f642f8c95c8503d5d4d0b6c9f9f"} Mar 13 12:52:06 crc kubenswrapper[4632]: I0313 12:52:06.951905 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac34c41e4a723fd479a80798f16607c4694c8f642f8c95c8503d5d4d0b6c9f9f" Mar 13 12:52:06 crc kubenswrapper[4632]: I0313 12:52:06.951914 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556772-dvdjm" Mar 13 12:52:07 crc kubenswrapper[4632]: I0313 12:52:07.030007 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556766-lklf4"] Mar 13 12:52:07 crc kubenswrapper[4632]: I0313 12:52:07.051062 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556766-lklf4"] Mar 13 12:52:08 crc kubenswrapper[4632]: I0313 12:52:08.061037 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e38323af-ae58-48ce-979e-c8905218b4fe" path="/var/lib/kubelet/pods/e38323af-ae58-48ce-979e-c8905218b4fe/volumes" Mar 13 12:52:11 crc kubenswrapper[4632]: I0313 12:52:11.044634 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:52:11 crc kubenswrapper[4632]: E0313 12:52:11.045567 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:52:18 crc kubenswrapper[4632]: I0313 12:52:18.109838 4632 generic.go:334] "Generic (PLEG): container finished" podID="252f97d9-adeb-4cce-858d-eb0bdb151871" containerID="ce4c408f0cb872f87b4909caa3af1013273fb7803c492e30e5b30666a913955d" exitCode=0 Mar 13 12:52:18 crc kubenswrapper[4632]: I0313 12:52:18.110425 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-jfn52/must-gather-9gqfn" event={"ID":"252f97d9-adeb-4cce-858d-eb0bdb151871","Type":"ContainerDied","Data":"ce4c408f0cb872f87b4909caa3af1013273fb7803c492e30e5b30666a913955d"} Mar 13 12:52:18 crc kubenswrapper[4632]: I0313 12:52:18.111521 4632 scope.go:117] "RemoveContainer" containerID="ce4c408f0cb872f87b4909caa3af1013273fb7803c492e30e5b30666a913955d" Mar 13 12:52:18 crc kubenswrapper[4632]: I0313 12:52:18.520381 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jfn52_must-gather-9gqfn_252f97d9-adeb-4cce-858d-eb0bdb151871/gather/0.log" Mar 13 12:52:24 crc kubenswrapper[4632]: I0313 12:52:24.044926 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:52:24 crc kubenswrapper[4632]: E0313 12:52:24.046285 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:52:25 crc kubenswrapper[4632]: I0313 12:52:25.814252 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4hj7k"] Mar 13 12:52:25 crc kubenswrapper[4632]: E0313 12:52:25.816482 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="576ef56c-ba4f-4def-89ac-cdef2e378fca" containerName="oc" Mar 13 12:52:25 crc kubenswrapper[4632]: I0313 12:52:25.816505 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="576ef56c-ba4f-4def-89ac-cdef2e378fca" containerName="oc" Mar 13 12:52:25 crc kubenswrapper[4632]: I0313 12:52:25.816859 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="576ef56c-ba4f-4def-89ac-cdef2e378fca" containerName="oc" Mar 13 12:52:25 crc kubenswrapper[4632]: I0313 12:52:25.827588 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:25 crc kubenswrapper[4632]: I0313 12:52:25.849015 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4hj7k"] Mar 13 12:52:25 crc kubenswrapper[4632]: I0313 12:52:25.963043 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24088399-8751-4389-b28b-1bca8ff6f809-catalog-content\") pod \"community-operators-4hj7k\" (UID: \"24088399-8751-4389-b28b-1bca8ff6f809\") " pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:25 crc kubenswrapper[4632]: I0313 12:52:25.963255 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mc94g\" (UniqueName: \"kubernetes.io/projected/24088399-8751-4389-b28b-1bca8ff6f809-kube-api-access-mc94g\") pod \"community-operators-4hj7k\" (UID: \"24088399-8751-4389-b28b-1bca8ff6f809\") " pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:25 crc kubenswrapper[4632]: I0313 12:52:25.963422 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24088399-8751-4389-b28b-1bca8ff6f809-utilities\") pod \"community-operators-4hj7k\" (UID: \"24088399-8751-4389-b28b-1bca8ff6f809\") " pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:26 crc kubenswrapper[4632]: I0313 12:52:26.065795 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24088399-8751-4389-b28b-1bca8ff6f809-utilities\") pod \"community-operators-4hj7k\" (UID: \"24088399-8751-4389-b28b-1bca8ff6f809\") " pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:26 crc kubenswrapper[4632]: I0313 12:52:26.066674 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24088399-8751-4389-b28b-1bca8ff6f809-utilities\") pod \"community-operators-4hj7k\" (UID: \"24088399-8751-4389-b28b-1bca8ff6f809\") " pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:26 crc kubenswrapper[4632]: I0313 12:52:26.066874 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24088399-8751-4389-b28b-1bca8ff6f809-catalog-content\") pod \"community-operators-4hj7k\" (UID: \"24088399-8751-4389-b28b-1bca8ff6f809\") " pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:26 crc kubenswrapper[4632]: I0313 12:52:26.067733 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mc94g\" (UniqueName: \"kubernetes.io/projected/24088399-8751-4389-b28b-1bca8ff6f809-kube-api-access-mc94g\") pod \"community-operators-4hj7k\" (UID: \"24088399-8751-4389-b28b-1bca8ff6f809\") " pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:26 crc kubenswrapper[4632]: I0313 12:52:26.071314 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24088399-8751-4389-b28b-1bca8ff6f809-catalog-content\") pod \"community-operators-4hj7k\" (UID: \"24088399-8751-4389-b28b-1bca8ff6f809\") " pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:26 crc kubenswrapper[4632]: I0313 12:52:26.457752 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mc94g\" (UniqueName: \"kubernetes.io/projected/24088399-8751-4389-b28b-1bca8ff6f809-kube-api-access-mc94g\") pod \"community-operators-4hj7k\" (UID: \"24088399-8751-4389-b28b-1bca8ff6f809\") " pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:26 crc kubenswrapper[4632]: I0313 12:52:26.470281 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:27 crc kubenswrapper[4632]: I0313 12:52:27.107425 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4hj7k"] Mar 13 12:52:27 crc kubenswrapper[4632]: I0313 12:52:27.212286 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hj7k" event={"ID":"24088399-8751-4389-b28b-1bca8ff6f809","Type":"ContainerStarted","Data":"c31bc30aa7a32f866165e740bb17b10d2dfb78b39bb7977502e5547cf122fe4f"} Mar 13 12:52:28 crc kubenswrapper[4632]: I0313 12:52:28.222102 4632 generic.go:334] "Generic (PLEG): container finished" podID="24088399-8751-4389-b28b-1bca8ff6f809" containerID="4f8879384b19de20a7ba023ed6242f065059f220338be6e6dcb54651850e6173" exitCode=0 Mar 13 12:52:28 crc kubenswrapper[4632]: I0313 12:52:28.222161 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hj7k" event={"ID":"24088399-8751-4389-b28b-1bca8ff6f809","Type":"ContainerDied","Data":"4f8879384b19de20a7ba023ed6242f065059f220338be6e6dcb54651850e6173"} Mar 13 12:52:29 crc kubenswrapper[4632]: I0313 12:52:29.232759 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hj7k" event={"ID":"24088399-8751-4389-b28b-1bca8ff6f809","Type":"ContainerStarted","Data":"84f0f7be6e5f1526a1ffb017b5e0805dd39faaa07bca4f8b78e79b3871714f7e"} Mar 13 12:52:31 crc kubenswrapper[4632]: I0313 12:52:31.228440 4632 scope.go:117] "RemoveContainer" containerID="e2d392c178854d8d02c1d90a74a70ca0dce9ae28135802be619d355191eb7f40" Mar 13 12:52:32 crc kubenswrapper[4632]: I0313 12:52:32.253723 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-jfn52/must-gather-9gqfn"] Mar 13 12:52:32 crc kubenswrapper[4632]: I0313 12:52:32.253979 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-jfn52/must-gather-9gqfn" podUID="252f97d9-adeb-4cce-858d-eb0bdb151871" containerName="copy" containerID="cri-o://8c2cc6936b125830f781b6c13d49ab8294ed023e1773b93feeb9bf3339c7d42f" gracePeriod=2 Mar 13 12:52:32 crc kubenswrapper[4632]: I0313 12:52:32.263499 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-jfn52/must-gather-9gqfn"] Mar 13 12:52:32 crc kubenswrapper[4632]: I0313 12:52:32.264890 4632 generic.go:334] "Generic (PLEG): container finished" podID="24088399-8751-4389-b28b-1bca8ff6f809" containerID="84f0f7be6e5f1526a1ffb017b5e0805dd39faaa07bca4f8b78e79b3871714f7e" exitCode=0 Mar 13 12:52:32 crc kubenswrapper[4632]: I0313 12:52:32.264956 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hj7k" event={"ID":"24088399-8751-4389-b28b-1bca8ff6f809","Type":"ContainerDied","Data":"84f0f7be6e5f1526a1ffb017b5e0805dd39faaa07bca4f8b78e79b3871714f7e"} Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.074911 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jfn52_must-gather-9gqfn_252f97d9-adeb-4cce-858d-eb0bdb151871/copy/0.log" Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.076567 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/must-gather-9gqfn" Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.232471 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwxgx\" (UniqueName: \"kubernetes.io/projected/252f97d9-adeb-4cce-858d-eb0bdb151871-kube-api-access-kwxgx\") pod \"252f97d9-adeb-4cce-858d-eb0bdb151871\" (UID: \"252f97d9-adeb-4cce-858d-eb0bdb151871\") " Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.232803 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/252f97d9-adeb-4cce-858d-eb0bdb151871-must-gather-output\") pod \"252f97d9-adeb-4cce-858d-eb0bdb151871\" (UID: \"252f97d9-adeb-4cce-858d-eb0bdb151871\") " Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.264041 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/252f97d9-adeb-4cce-858d-eb0bdb151871-kube-api-access-kwxgx" (OuterVolumeSpecName: "kube-api-access-kwxgx") pod "252f97d9-adeb-4cce-858d-eb0bdb151871" (UID: "252f97d9-adeb-4cce-858d-eb0bdb151871"). InnerVolumeSpecName "kube-api-access-kwxgx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.282883 4632 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-jfn52_must-gather-9gqfn_252f97d9-adeb-4cce-858d-eb0bdb151871/copy/0.log" Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.283420 4632 generic.go:334] "Generic (PLEG): container finished" podID="252f97d9-adeb-4cce-858d-eb0bdb151871" containerID="8c2cc6936b125830f781b6c13d49ab8294ed023e1773b93feeb9bf3339c7d42f" exitCode=143 Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.283497 4632 scope.go:117] "RemoveContainer" containerID="8c2cc6936b125830f781b6c13d49ab8294ed023e1773b93feeb9bf3339c7d42f" Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.284185 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-jfn52/must-gather-9gqfn" Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.293248 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hj7k" event={"ID":"24088399-8751-4389-b28b-1bca8ff6f809","Type":"ContainerStarted","Data":"4e0eda93026c95782b0f1087fe1c83f62b5c89bdde15c230bf8e0fb7a66c1f63"} Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.323283 4632 scope.go:117] "RemoveContainer" containerID="ce4c408f0cb872f87b4909caa3af1013273fb7803c492e30e5b30666a913955d" Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.329711 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4hj7k" podStartSLOduration=3.7149638659999997 podStartE2EDuration="8.32969037s" podCreationTimestamp="2026-03-13 12:52:25 +0000 UTC" firstStartedPulling="2026-03-13 12:52:28.223983826 +0000 UTC m=+10122.246513959" lastFinishedPulling="2026-03-13 12:52:32.83871033 +0000 UTC m=+10126.861240463" observedRunningTime="2026-03-13 12:52:33.324060631 +0000 UTC m=+10127.346590764" watchObservedRunningTime="2026-03-13 12:52:33.32969037 +0000 UTC m=+10127.352220503" Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.338529 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwxgx\" (UniqueName: \"kubernetes.io/projected/252f97d9-adeb-4cce-858d-eb0bdb151871-kube-api-access-kwxgx\") on node \"crc\" DevicePath \"\"" Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.368565 4632 scope.go:117] "RemoveContainer" containerID="8c2cc6936b125830f781b6c13d49ab8294ed023e1773b93feeb9bf3339c7d42f" Mar 13 12:52:33 crc kubenswrapper[4632]: E0313 12:52:33.375816 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c2cc6936b125830f781b6c13d49ab8294ed023e1773b93feeb9bf3339c7d42f\": container with ID starting with 8c2cc6936b125830f781b6c13d49ab8294ed023e1773b93feeb9bf3339c7d42f not found: ID does not exist" containerID="8c2cc6936b125830f781b6c13d49ab8294ed023e1773b93feeb9bf3339c7d42f" Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.375883 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c2cc6936b125830f781b6c13d49ab8294ed023e1773b93feeb9bf3339c7d42f"} err="failed to get container status \"8c2cc6936b125830f781b6c13d49ab8294ed023e1773b93feeb9bf3339c7d42f\": rpc error: code = NotFound desc = could not find container \"8c2cc6936b125830f781b6c13d49ab8294ed023e1773b93feeb9bf3339c7d42f\": container with ID starting with 8c2cc6936b125830f781b6c13d49ab8294ed023e1773b93feeb9bf3339c7d42f not found: ID does not exist" Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.375921 4632 scope.go:117] "RemoveContainer" containerID="ce4c408f0cb872f87b4909caa3af1013273fb7803c492e30e5b30666a913955d" Mar 13 12:52:33 crc kubenswrapper[4632]: E0313 12:52:33.376684 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce4c408f0cb872f87b4909caa3af1013273fb7803c492e30e5b30666a913955d\": container with ID starting with ce4c408f0cb872f87b4909caa3af1013273fb7803c492e30e5b30666a913955d not found: ID does not exist" containerID="ce4c408f0cb872f87b4909caa3af1013273fb7803c492e30e5b30666a913955d" Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.376720 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce4c408f0cb872f87b4909caa3af1013273fb7803c492e30e5b30666a913955d"} err="failed to get container status \"ce4c408f0cb872f87b4909caa3af1013273fb7803c492e30e5b30666a913955d\": rpc error: code = NotFound desc = could not find container \"ce4c408f0cb872f87b4909caa3af1013273fb7803c492e30e5b30666a913955d\": container with ID starting with ce4c408f0cb872f87b4909caa3af1013273fb7803c492e30e5b30666a913955d not found: ID does not exist" Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.384740 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/252f97d9-adeb-4cce-858d-eb0bdb151871-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "252f97d9-adeb-4cce-858d-eb0bdb151871" (UID: "252f97d9-adeb-4cce-858d-eb0bdb151871"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:52:33 crc kubenswrapper[4632]: I0313 12:52:33.441017 4632 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/252f97d9-adeb-4cce-858d-eb0bdb151871-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 13 12:52:34 crc kubenswrapper[4632]: I0313 12:52:34.055776 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="252f97d9-adeb-4cce-858d-eb0bdb151871" path="/var/lib/kubelet/pods/252f97d9-adeb-4cce-858d-eb0bdb151871/volumes" Mar 13 12:52:36 crc kubenswrapper[4632]: I0313 12:52:36.470447 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:36 crc kubenswrapper[4632]: I0313 12:52:36.470979 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:37 crc kubenswrapper[4632]: I0313 12:52:37.044415 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:52:37 crc kubenswrapper[4632]: E0313 12:52:37.044922 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:52:37 crc kubenswrapper[4632]: I0313 12:52:37.527737 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-4hj7k" podUID="24088399-8751-4389-b28b-1bca8ff6f809" containerName="registry-server" probeResult="failure" output=< Mar 13 12:52:37 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:52:37 crc kubenswrapper[4632]: > Mar 13 12:52:47 crc kubenswrapper[4632]: I0313 12:52:47.519981 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-4hj7k" podUID="24088399-8751-4389-b28b-1bca8ff6f809" containerName="registry-server" probeResult="failure" output=< Mar 13 12:52:47 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:52:47 crc kubenswrapper[4632]: > Mar 13 12:52:50 crc kubenswrapper[4632]: I0313 12:52:50.044831 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:52:50 crc kubenswrapper[4632]: E0313 12:52:50.045576 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:52:56 crc kubenswrapper[4632]: I0313 12:52:56.525956 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:56 crc kubenswrapper[4632]: I0313 12:52:56.597331 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:56 crc kubenswrapper[4632]: I0313 12:52:56.778778 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4hj7k"] Mar 13 12:52:58 crc kubenswrapper[4632]: I0313 12:52:58.549161 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4hj7k" podUID="24088399-8751-4389-b28b-1bca8ff6f809" containerName="registry-server" containerID="cri-o://4e0eda93026c95782b0f1087fe1c83f62b5c89bdde15c230bf8e0fb7a66c1f63" gracePeriod=2 Mar 13 12:52:59 crc kubenswrapper[4632]: I0313 12:52:59.594671 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hj7k" event={"ID":"24088399-8751-4389-b28b-1bca8ff6f809","Type":"ContainerDied","Data":"4e0eda93026c95782b0f1087fe1c83f62b5c89bdde15c230bf8e0fb7a66c1f63"} Mar 13 12:52:59 crc kubenswrapper[4632]: I0313 12:52:59.597761 4632 generic.go:334] "Generic (PLEG): container finished" podID="24088399-8751-4389-b28b-1bca8ff6f809" containerID="4e0eda93026c95782b0f1087fe1c83f62b5c89bdde15c230bf8e0fb7a66c1f63" exitCode=0 Mar 13 12:52:59 crc kubenswrapper[4632]: I0313 12:52:59.597824 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4hj7k" event={"ID":"24088399-8751-4389-b28b-1bca8ff6f809","Type":"ContainerDied","Data":"c31bc30aa7a32f866165e740bb17b10d2dfb78b39bb7977502e5547cf122fe4f"} Mar 13 12:52:59 crc kubenswrapper[4632]: I0313 12:52:59.597851 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c31bc30aa7a32f866165e740bb17b10d2dfb78b39bb7977502e5547cf122fe4f" Mar 13 12:52:59 crc kubenswrapper[4632]: I0313 12:52:59.599520 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:52:59 crc kubenswrapper[4632]: I0313 12:52:59.734709 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24088399-8751-4389-b28b-1bca8ff6f809-catalog-content\") pod \"24088399-8751-4389-b28b-1bca8ff6f809\" (UID: \"24088399-8751-4389-b28b-1bca8ff6f809\") " Mar 13 12:52:59 crc kubenswrapper[4632]: I0313 12:52:59.734859 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mc94g\" (UniqueName: \"kubernetes.io/projected/24088399-8751-4389-b28b-1bca8ff6f809-kube-api-access-mc94g\") pod \"24088399-8751-4389-b28b-1bca8ff6f809\" (UID: \"24088399-8751-4389-b28b-1bca8ff6f809\") " Mar 13 12:52:59 crc kubenswrapper[4632]: I0313 12:52:59.734933 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24088399-8751-4389-b28b-1bca8ff6f809-utilities\") pod \"24088399-8751-4389-b28b-1bca8ff6f809\" (UID: \"24088399-8751-4389-b28b-1bca8ff6f809\") " Mar 13 12:52:59 crc kubenswrapper[4632]: I0313 12:52:59.735537 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24088399-8751-4389-b28b-1bca8ff6f809-utilities" (OuterVolumeSpecName: "utilities") pod "24088399-8751-4389-b28b-1bca8ff6f809" (UID: "24088399-8751-4389-b28b-1bca8ff6f809"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:52:59 crc kubenswrapper[4632]: I0313 12:52:59.750521 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/24088399-8751-4389-b28b-1bca8ff6f809-kube-api-access-mc94g" (OuterVolumeSpecName: "kube-api-access-mc94g") pod "24088399-8751-4389-b28b-1bca8ff6f809" (UID: "24088399-8751-4389-b28b-1bca8ff6f809"). InnerVolumeSpecName "kube-api-access-mc94g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:52:59 crc kubenswrapper[4632]: I0313 12:52:59.804998 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/24088399-8751-4389-b28b-1bca8ff6f809-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "24088399-8751-4389-b28b-1bca8ff6f809" (UID: "24088399-8751-4389-b28b-1bca8ff6f809"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:52:59 crc kubenswrapper[4632]: I0313 12:52:59.840208 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/24088399-8751-4389-b28b-1bca8ff6f809-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:52:59 crc kubenswrapper[4632]: I0313 12:52:59.840546 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/24088399-8751-4389-b28b-1bca8ff6f809-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:52:59 crc kubenswrapper[4632]: I0313 12:52:59.840638 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mc94g\" (UniqueName: \"kubernetes.io/projected/24088399-8751-4389-b28b-1bca8ff6f809-kube-api-access-mc94g\") on node \"crc\" DevicePath \"\"" Mar 13 12:53:00 crc kubenswrapper[4632]: I0313 12:53:00.606653 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4hj7k" Mar 13 12:53:00 crc kubenswrapper[4632]: I0313 12:53:00.634560 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4hj7k"] Mar 13 12:53:00 crc kubenswrapper[4632]: I0313 12:53:00.642894 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4hj7k"] Mar 13 12:53:02 crc kubenswrapper[4632]: I0313 12:53:02.058968 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="24088399-8751-4389-b28b-1bca8ff6f809" path="/var/lib/kubelet/pods/24088399-8751-4389-b28b-1bca8ff6f809/volumes" Mar 13 12:53:05 crc kubenswrapper[4632]: I0313 12:53:05.045442 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:53:05 crc kubenswrapper[4632]: E0313 12:53:05.046779 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:53:17 crc kubenswrapper[4632]: I0313 12:53:17.044160 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:53:17 crc kubenswrapper[4632]: E0313 12:53:17.044793 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:53:31 crc kubenswrapper[4632]: I0313 12:53:31.045214 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:53:31 crc kubenswrapper[4632]: E0313 12:53:31.046602 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:53:31 crc kubenswrapper[4632]: I0313 12:53:31.308091 4632 scope.go:117] "RemoveContainer" containerID="4814d54398dddc7491e0a4c9f868011d9742ed90e2d05a245b23e35c28be791e" Mar 13 12:53:31 crc kubenswrapper[4632]: I0313 12:53:31.342130 4632 scope.go:117] "RemoveContainer" containerID="537f35c912c9c8057ecc3ec80663f0a0d3c386360eb374b59f3e50a3f8bd59ee" Mar 13 12:53:31 crc kubenswrapper[4632]: I0313 12:53:31.378101 4632 scope.go:117] "RemoveContainer" containerID="219da7f0798a57241a68dc972a8e6cf63665a59509f96ff776be2e82e493c3c5" Mar 13 12:53:46 crc kubenswrapper[4632]: I0313 12:53:46.045599 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:53:46 crc kubenswrapper[4632]: E0313 12:53:46.046661 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:53:57 crc kubenswrapper[4632]: I0313 12:53:57.044345 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:53:57 crc kubenswrapper[4632]: E0313 12:53:57.046655 4632 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-zkscb_openshift-machine-config-operator(d77b18a7-7ad9-4bf5-bff5-da45878af7f4)\"" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.285257 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556774-hwghb"] Mar 13 12:54:00 crc kubenswrapper[4632]: E0313 12:54:00.289616 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24088399-8751-4389-b28b-1bca8ff6f809" containerName="extract-utilities" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.289645 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="24088399-8751-4389-b28b-1bca8ff6f809" containerName="extract-utilities" Mar 13 12:54:00 crc kubenswrapper[4632]: E0313 12:54:00.289679 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24088399-8751-4389-b28b-1bca8ff6f809" containerName="extract-content" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.289687 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="24088399-8751-4389-b28b-1bca8ff6f809" containerName="extract-content" Mar 13 12:54:00 crc kubenswrapper[4632]: E0313 12:54:00.289701 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="252f97d9-adeb-4cce-858d-eb0bdb151871" containerName="gather" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.289707 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="252f97d9-adeb-4cce-858d-eb0bdb151871" containerName="gather" Mar 13 12:54:00 crc kubenswrapper[4632]: E0313 12:54:00.289730 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="252f97d9-adeb-4cce-858d-eb0bdb151871" containerName="copy" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.289736 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="252f97d9-adeb-4cce-858d-eb0bdb151871" containerName="copy" Mar 13 12:54:00 crc kubenswrapper[4632]: E0313 12:54:00.289757 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="24088399-8751-4389-b28b-1bca8ff6f809" containerName="registry-server" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.289763 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="24088399-8751-4389-b28b-1bca8ff6f809" containerName="registry-server" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.290712 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="252f97d9-adeb-4cce-858d-eb0bdb151871" containerName="gather" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.290743 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="252f97d9-adeb-4cce-858d-eb0bdb151871" containerName="copy" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.290756 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="24088399-8751-4389-b28b-1bca8ff6f809" containerName="registry-server" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.297493 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556774-hwghb" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.308883 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.308885 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.308890 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.355184 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556774-hwghb"] Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.390101 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6hnl\" (UniqueName: \"kubernetes.io/projected/f3f96182-2be7-4262-9ba2-7b363f07fd2d-kube-api-access-r6hnl\") pod \"auto-csr-approver-29556774-hwghb\" (UID: \"f3f96182-2be7-4262-9ba2-7b363f07fd2d\") " pod="openshift-infra/auto-csr-approver-29556774-hwghb" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.492549 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6hnl\" (UniqueName: \"kubernetes.io/projected/f3f96182-2be7-4262-9ba2-7b363f07fd2d-kube-api-access-r6hnl\") pod \"auto-csr-approver-29556774-hwghb\" (UID: \"f3f96182-2be7-4262-9ba2-7b363f07fd2d\") " pod="openshift-infra/auto-csr-approver-29556774-hwghb" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.532335 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6hnl\" (UniqueName: \"kubernetes.io/projected/f3f96182-2be7-4262-9ba2-7b363f07fd2d-kube-api-access-r6hnl\") pod \"auto-csr-approver-29556774-hwghb\" (UID: \"f3f96182-2be7-4262-9ba2-7b363f07fd2d\") " pod="openshift-infra/auto-csr-approver-29556774-hwghb" Mar 13 12:54:00 crc kubenswrapper[4632]: I0313 12:54:00.622512 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556774-hwghb" Mar 13 12:54:01 crc kubenswrapper[4632]: I0313 12:54:01.050680 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556774-hwghb"] Mar 13 12:54:01 crc kubenswrapper[4632]: W0313 12:54:01.066803 4632 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3f96182_2be7_4262_9ba2_7b363f07fd2d.slice/crio-26cf7dbf74e22752caadec7e37b845a467437edabf9d757be9c601c40aef85e4 WatchSource:0}: Error finding container 26cf7dbf74e22752caadec7e37b845a467437edabf9d757be9c601c40aef85e4: Status 404 returned error can't find the container with id 26cf7dbf74e22752caadec7e37b845a467437edabf9d757be9c601c40aef85e4 Mar 13 12:54:01 crc kubenswrapper[4632]: I0313 12:54:01.088117 4632 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 12:54:01 crc kubenswrapper[4632]: I0313 12:54:01.176151 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556774-hwghb" event={"ID":"f3f96182-2be7-4262-9ba2-7b363f07fd2d","Type":"ContainerStarted","Data":"26cf7dbf74e22752caadec7e37b845a467437edabf9d757be9c601c40aef85e4"} Mar 13 12:54:04 crc kubenswrapper[4632]: I0313 12:54:04.204422 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556774-hwghb" event={"ID":"f3f96182-2be7-4262-9ba2-7b363f07fd2d","Type":"ContainerStarted","Data":"d2d7c4023cf6c21c65c9b9134fc1ea2d1154414b6f5373f5fd1987f228f3c9f6"} Mar 13 12:54:04 crc kubenswrapper[4632]: I0313 12:54:04.235566 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556774-hwghb" podStartSLOduration=2.9074145639999998 podStartE2EDuration="4.234407034s" podCreationTimestamp="2026-03-13 12:54:00 +0000 UTC" firstStartedPulling="2026-03-13 12:54:01.083038656 +0000 UTC m=+10215.105568789" lastFinishedPulling="2026-03-13 12:54:02.410031126 +0000 UTC m=+10216.432561259" observedRunningTime="2026-03-13 12:54:04.219758194 +0000 UTC m=+10218.242288357" watchObservedRunningTime="2026-03-13 12:54:04.234407034 +0000 UTC m=+10218.256937177" Mar 13 12:54:05 crc kubenswrapper[4632]: I0313 12:54:05.216636 4632 generic.go:334] "Generic (PLEG): container finished" podID="f3f96182-2be7-4262-9ba2-7b363f07fd2d" containerID="d2d7c4023cf6c21c65c9b9134fc1ea2d1154414b6f5373f5fd1987f228f3c9f6" exitCode=0 Mar 13 12:54:05 crc kubenswrapper[4632]: I0313 12:54:05.216681 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556774-hwghb" event={"ID":"f3f96182-2be7-4262-9ba2-7b363f07fd2d","Type":"ContainerDied","Data":"d2d7c4023cf6c21c65c9b9134fc1ea2d1154414b6f5373f5fd1987f228f3c9f6"} Mar 13 12:54:06 crc kubenswrapper[4632]: I0313 12:54:06.602418 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556774-hwghb" Mar 13 12:54:06 crc kubenswrapper[4632]: I0313 12:54:06.718373 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6hnl\" (UniqueName: \"kubernetes.io/projected/f3f96182-2be7-4262-9ba2-7b363f07fd2d-kube-api-access-r6hnl\") pod \"f3f96182-2be7-4262-9ba2-7b363f07fd2d\" (UID: \"f3f96182-2be7-4262-9ba2-7b363f07fd2d\") " Mar 13 12:54:06 crc kubenswrapper[4632]: I0313 12:54:06.732591 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3f96182-2be7-4262-9ba2-7b363f07fd2d-kube-api-access-r6hnl" (OuterVolumeSpecName: "kube-api-access-r6hnl") pod "f3f96182-2be7-4262-9ba2-7b363f07fd2d" (UID: "f3f96182-2be7-4262-9ba2-7b363f07fd2d"). InnerVolumeSpecName "kube-api-access-r6hnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:54:06 crc kubenswrapper[4632]: I0313 12:54:06.820628 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6hnl\" (UniqueName: \"kubernetes.io/projected/f3f96182-2be7-4262-9ba2-7b363f07fd2d-kube-api-access-r6hnl\") on node \"crc\" DevicePath \"\"" Mar 13 12:54:07 crc kubenswrapper[4632]: I0313 12:54:07.239969 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556774-hwghb" event={"ID":"f3f96182-2be7-4262-9ba2-7b363f07fd2d","Type":"ContainerDied","Data":"26cf7dbf74e22752caadec7e37b845a467437edabf9d757be9c601c40aef85e4"} Mar 13 12:54:07 crc kubenswrapper[4632]: I0313 12:54:07.240034 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26cf7dbf74e22752caadec7e37b845a467437edabf9d757be9c601c40aef85e4" Mar 13 12:54:07 crc kubenswrapper[4632]: I0313 12:54:07.240057 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556774-hwghb" Mar 13 12:54:07 crc kubenswrapper[4632]: I0313 12:54:07.343178 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556768-qx4vs"] Mar 13 12:54:07 crc kubenswrapper[4632]: I0313 12:54:07.356417 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556768-qx4vs"] Mar 13 12:54:08 crc kubenswrapper[4632]: I0313 12:54:08.070959 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c0a936c-9a75-4f0b-81b1-fb7f74d9911f" path="/var/lib/kubelet/pods/7c0a936c-9a75-4f0b-81b1-fb7f74d9911f/volumes" Mar 13 12:54:12 crc kubenswrapper[4632]: I0313 12:54:12.044983 4632 scope.go:117] "RemoveContainer" containerID="f99b8bf20e90f85804d1361e895aa3d02f6ce45057066d1347ea4edee62e1086" Mar 13 12:54:13 crc kubenswrapper[4632]: I0313 12:54:13.306362 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" event={"ID":"d77b18a7-7ad9-4bf5-bff5-da45878af7f4","Type":"ContainerStarted","Data":"07bc2cd5f1b1c505fce3a8916e93f01bdf10c685e4d30c4c6edb1e67e635fe5f"} Mar 13 12:54:31 crc kubenswrapper[4632]: I0313 12:54:31.482849 4632 scope.go:117] "RemoveContainer" containerID="fbdf7412c66e2fa539b75629b05076618e8fad2c845d05a467fd575d619baa55" Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.144298 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8wm2q"] Mar 13 12:55:10 crc kubenswrapper[4632]: E0313 12:55:10.149672 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3f96182-2be7-4262-9ba2-7b363f07fd2d" containerName="oc" Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.149825 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3f96182-2be7-4262-9ba2-7b363f07fd2d" containerName="oc" Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.150674 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3f96182-2be7-4262-9ba2-7b363f07fd2d" containerName="oc" Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.158457 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.165596 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8wm2q"] Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.342089 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156b689a-7e5a-4335-b861-6470b6c336e9-catalog-content\") pod \"certified-operators-8wm2q\" (UID: \"156b689a-7e5a-4335-b861-6470b6c336e9\") " pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.342478 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156b689a-7e5a-4335-b861-6470b6c336e9-utilities\") pod \"certified-operators-8wm2q\" (UID: \"156b689a-7e5a-4335-b861-6470b6c336e9\") " pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.342552 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gclsw\" (UniqueName: \"kubernetes.io/projected/156b689a-7e5a-4335-b861-6470b6c336e9-kube-api-access-gclsw\") pod \"certified-operators-8wm2q\" (UID: \"156b689a-7e5a-4335-b861-6470b6c336e9\") " pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.444510 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156b689a-7e5a-4335-b861-6470b6c336e9-utilities\") pod \"certified-operators-8wm2q\" (UID: \"156b689a-7e5a-4335-b861-6470b6c336e9\") " pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.444671 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gclsw\" (UniqueName: \"kubernetes.io/projected/156b689a-7e5a-4335-b861-6470b6c336e9-kube-api-access-gclsw\") pod \"certified-operators-8wm2q\" (UID: \"156b689a-7e5a-4335-b861-6470b6c336e9\") " pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.444830 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156b689a-7e5a-4335-b861-6470b6c336e9-catalog-content\") pod \"certified-operators-8wm2q\" (UID: \"156b689a-7e5a-4335-b861-6470b6c336e9\") " pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.445481 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156b689a-7e5a-4335-b861-6470b6c336e9-utilities\") pod \"certified-operators-8wm2q\" (UID: \"156b689a-7e5a-4335-b861-6470b6c336e9\") " pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.446754 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156b689a-7e5a-4335-b861-6470b6c336e9-catalog-content\") pod \"certified-operators-8wm2q\" (UID: \"156b689a-7e5a-4335-b861-6470b6c336e9\") " pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.469831 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gclsw\" (UniqueName: \"kubernetes.io/projected/156b689a-7e5a-4335-b861-6470b6c336e9-kube-api-access-gclsw\") pod \"certified-operators-8wm2q\" (UID: \"156b689a-7e5a-4335-b861-6470b6c336e9\") " pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:10 crc kubenswrapper[4632]: I0313 12:55:10.479964 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:11 crc kubenswrapper[4632]: I0313 12:55:11.110864 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8wm2q"] Mar 13 12:55:11 crc kubenswrapper[4632]: I0313 12:55:11.995743 4632 generic.go:334] "Generic (PLEG): container finished" podID="156b689a-7e5a-4335-b861-6470b6c336e9" containerID="8dab11ebf37d779f2da3a122c2b995fd25ff3a689b642c34db1453ee35e1b169" exitCode=0 Mar 13 12:55:11 crc kubenswrapper[4632]: I0313 12:55:11.997235 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8wm2q" event={"ID":"156b689a-7e5a-4335-b861-6470b6c336e9","Type":"ContainerDied","Data":"8dab11ebf37d779f2da3a122c2b995fd25ff3a689b642c34db1453ee35e1b169"} Mar 13 12:55:11 crc kubenswrapper[4632]: I0313 12:55:11.997378 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8wm2q" event={"ID":"156b689a-7e5a-4335-b861-6470b6c336e9","Type":"ContainerStarted","Data":"feb59d2fc60689fa5a353a08e74284afe95caab4744d06f1c02c1c8737813065"} Mar 13 12:55:13 crc kubenswrapper[4632]: I0313 12:55:13.016205 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8wm2q" event={"ID":"156b689a-7e5a-4335-b861-6470b6c336e9","Type":"ContainerStarted","Data":"d7b53ae310ac7f474cee531840648431c705e87927ace402b17a25d79a19a361"} Mar 13 12:55:15 crc kubenswrapper[4632]: I0313 12:55:15.034917 4632 generic.go:334] "Generic (PLEG): container finished" podID="156b689a-7e5a-4335-b861-6470b6c336e9" containerID="d7b53ae310ac7f474cee531840648431c705e87927ace402b17a25d79a19a361" exitCode=0 Mar 13 12:55:15 crc kubenswrapper[4632]: I0313 12:55:15.034979 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8wm2q" event={"ID":"156b689a-7e5a-4335-b861-6470b6c336e9","Type":"ContainerDied","Data":"d7b53ae310ac7f474cee531840648431c705e87927ace402b17a25d79a19a361"} Mar 13 12:55:17 crc kubenswrapper[4632]: I0313 12:55:17.069817 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8wm2q" event={"ID":"156b689a-7e5a-4335-b861-6470b6c336e9","Type":"ContainerStarted","Data":"a56230e1aca4eab0b09b3cd9ce5c2f2212d66ef93f13ffa3c09b64aa2806a748"} Mar 13 12:55:17 crc kubenswrapper[4632]: I0313 12:55:17.093815 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8wm2q" podStartSLOduration=3.532117583 podStartE2EDuration="7.09374426s" podCreationTimestamp="2026-03-13 12:55:10 +0000 UTC" firstStartedPulling="2026-03-13 12:55:12.001057944 +0000 UTC m=+10286.023588077" lastFinishedPulling="2026-03-13 12:55:15.562684581 +0000 UTC m=+10289.585214754" observedRunningTime="2026-03-13 12:55:17.088371538 +0000 UTC m=+10291.110901681" watchObservedRunningTime="2026-03-13 12:55:17.09374426 +0000 UTC m=+10291.116274393" Mar 13 12:55:20 crc kubenswrapper[4632]: I0313 12:55:20.480781 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:20 crc kubenswrapper[4632]: I0313 12:55:20.481331 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:21 crc kubenswrapper[4632]: I0313 12:55:21.588000 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-8wm2q" podUID="156b689a-7e5a-4335-b861-6470b6c336e9" containerName="registry-server" probeResult="failure" output=< Mar 13 12:55:21 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:55:21 crc kubenswrapper[4632]: > Mar 13 12:55:31 crc kubenswrapper[4632]: I0313 12:55:31.537593 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-8wm2q" podUID="156b689a-7e5a-4335-b861-6470b6c336e9" containerName="registry-server" probeResult="failure" output=< Mar 13 12:55:31 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:55:31 crc kubenswrapper[4632]: > Mar 13 12:55:40 crc kubenswrapper[4632]: I0313 12:55:40.545449 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:40 crc kubenswrapper[4632]: I0313 12:55:40.606613 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:41 crc kubenswrapper[4632]: I0313 12:55:41.332138 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8wm2q"] Mar 13 12:55:42 crc kubenswrapper[4632]: I0313 12:55:42.362046 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8wm2q" podUID="156b689a-7e5a-4335-b861-6470b6c336e9" containerName="registry-server" containerID="cri-o://a56230e1aca4eab0b09b3cd9ce5c2f2212d66ef93f13ffa3c09b64aa2806a748" gracePeriod=2 Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.330865 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.379709 4632 generic.go:334] "Generic (PLEG): container finished" podID="156b689a-7e5a-4335-b861-6470b6c336e9" containerID="a56230e1aca4eab0b09b3cd9ce5c2f2212d66ef93f13ffa3c09b64aa2806a748" exitCode=0 Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.379756 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8wm2q" event={"ID":"156b689a-7e5a-4335-b861-6470b6c336e9","Type":"ContainerDied","Data":"a56230e1aca4eab0b09b3cd9ce5c2f2212d66ef93f13ffa3c09b64aa2806a748"} Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.379780 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8wm2q" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.379804 4632 scope.go:117] "RemoveContainer" containerID="a56230e1aca4eab0b09b3cd9ce5c2f2212d66ef93f13ffa3c09b64aa2806a748" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.379788 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8wm2q" event={"ID":"156b689a-7e5a-4335-b861-6470b6c336e9","Type":"ContainerDied","Data":"feb59d2fc60689fa5a353a08e74284afe95caab4744d06f1c02c1c8737813065"} Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.419993 4632 scope.go:117] "RemoveContainer" containerID="d7b53ae310ac7f474cee531840648431c705e87927ace402b17a25d79a19a361" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.437997 4632 scope.go:117] "RemoveContainer" containerID="8dab11ebf37d779f2da3a122c2b995fd25ff3a689b642c34db1453ee35e1b169" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.462448 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gclsw\" (UniqueName: \"kubernetes.io/projected/156b689a-7e5a-4335-b861-6470b6c336e9-kube-api-access-gclsw\") pod \"156b689a-7e5a-4335-b861-6470b6c336e9\" (UID: \"156b689a-7e5a-4335-b861-6470b6c336e9\") " Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.462533 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156b689a-7e5a-4335-b861-6470b6c336e9-catalog-content\") pod \"156b689a-7e5a-4335-b861-6470b6c336e9\" (UID: \"156b689a-7e5a-4335-b861-6470b6c336e9\") " Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.462573 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156b689a-7e5a-4335-b861-6470b6c336e9-utilities\") pod \"156b689a-7e5a-4335-b861-6470b6c336e9\" (UID: \"156b689a-7e5a-4335-b861-6470b6c336e9\") " Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.463428 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156b689a-7e5a-4335-b861-6470b6c336e9-utilities" (OuterVolumeSpecName: "utilities") pod "156b689a-7e5a-4335-b861-6470b6c336e9" (UID: "156b689a-7e5a-4335-b861-6470b6c336e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.472414 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/156b689a-7e5a-4335-b861-6470b6c336e9-kube-api-access-gclsw" (OuterVolumeSpecName: "kube-api-access-gclsw") pod "156b689a-7e5a-4335-b861-6470b6c336e9" (UID: "156b689a-7e5a-4335-b861-6470b6c336e9"). InnerVolumeSpecName "kube-api-access-gclsw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.488126 4632 scope.go:117] "RemoveContainer" containerID="a56230e1aca4eab0b09b3cd9ce5c2f2212d66ef93f13ffa3c09b64aa2806a748" Mar 13 12:55:43 crc kubenswrapper[4632]: E0313 12:55:43.492143 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a56230e1aca4eab0b09b3cd9ce5c2f2212d66ef93f13ffa3c09b64aa2806a748\": container with ID starting with a56230e1aca4eab0b09b3cd9ce5c2f2212d66ef93f13ffa3c09b64aa2806a748 not found: ID does not exist" containerID="a56230e1aca4eab0b09b3cd9ce5c2f2212d66ef93f13ffa3c09b64aa2806a748" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.492212 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a56230e1aca4eab0b09b3cd9ce5c2f2212d66ef93f13ffa3c09b64aa2806a748"} err="failed to get container status \"a56230e1aca4eab0b09b3cd9ce5c2f2212d66ef93f13ffa3c09b64aa2806a748\": rpc error: code = NotFound desc = could not find container \"a56230e1aca4eab0b09b3cd9ce5c2f2212d66ef93f13ffa3c09b64aa2806a748\": container with ID starting with a56230e1aca4eab0b09b3cd9ce5c2f2212d66ef93f13ffa3c09b64aa2806a748 not found: ID does not exist" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.492244 4632 scope.go:117] "RemoveContainer" containerID="d7b53ae310ac7f474cee531840648431c705e87927ace402b17a25d79a19a361" Mar 13 12:55:43 crc kubenswrapper[4632]: E0313 12:55:43.492682 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7b53ae310ac7f474cee531840648431c705e87927ace402b17a25d79a19a361\": container with ID starting with d7b53ae310ac7f474cee531840648431c705e87927ace402b17a25d79a19a361 not found: ID does not exist" containerID="d7b53ae310ac7f474cee531840648431c705e87927ace402b17a25d79a19a361" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.492717 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7b53ae310ac7f474cee531840648431c705e87927ace402b17a25d79a19a361"} err="failed to get container status \"d7b53ae310ac7f474cee531840648431c705e87927ace402b17a25d79a19a361\": rpc error: code = NotFound desc = could not find container \"d7b53ae310ac7f474cee531840648431c705e87927ace402b17a25d79a19a361\": container with ID starting with d7b53ae310ac7f474cee531840648431c705e87927ace402b17a25d79a19a361 not found: ID does not exist" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.492735 4632 scope.go:117] "RemoveContainer" containerID="8dab11ebf37d779f2da3a122c2b995fd25ff3a689b642c34db1453ee35e1b169" Mar 13 12:55:43 crc kubenswrapper[4632]: E0313 12:55:43.493050 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dab11ebf37d779f2da3a122c2b995fd25ff3a689b642c34db1453ee35e1b169\": container with ID starting with 8dab11ebf37d779f2da3a122c2b995fd25ff3a689b642c34db1453ee35e1b169 not found: ID does not exist" containerID="8dab11ebf37d779f2da3a122c2b995fd25ff3a689b642c34db1453ee35e1b169" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.493073 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dab11ebf37d779f2da3a122c2b995fd25ff3a689b642c34db1453ee35e1b169"} err="failed to get container status \"8dab11ebf37d779f2da3a122c2b995fd25ff3a689b642c34db1453ee35e1b169\": rpc error: code = NotFound desc = could not find container \"8dab11ebf37d779f2da3a122c2b995fd25ff3a689b642c34db1453ee35e1b169\": container with ID starting with 8dab11ebf37d779f2da3a122c2b995fd25ff3a689b642c34db1453ee35e1b169 not found: ID does not exist" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.538513 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/156b689a-7e5a-4335-b861-6470b6c336e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "156b689a-7e5a-4335-b861-6470b6c336e9" (UID: "156b689a-7e5a-4335-b861-6470b6c336e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.567831 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gclsw\" (UniqueName: \"kubernetes.io/projected/156b689a-7e5a-4335-b861-6470b6c336e9-kube-api-access-gclsw\") on node \"crc\" DevicePath \"\"" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.567866 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/156b689a-7e5a-4335-b861-6470b6c336e9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.567875 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/156b689a-7e5a-4335-b861-6470b6c336e9-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.727601 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8wm2q"] Mar 13 12:55:43 crc kubenswrapper[4632]: I0313 12:55:43.737714 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8wm2q"] Mar 13 12:55:44 crc kubenswrapper[4632]: I0313 12:55:44.079047 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="156b689a-7e5a-4335-b861-6470b6c336e9" path="/var/lib/kubelet/pods/156b689a-7e5a-4335-b861-6470b6c336e9/volumes" Mar 13 12:56:00 crc kubenswrapper[4632]: I0313 12:56:00.168729 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556776-spgrm"] Mar 13 12:56:00 crc kubenswrapper[4632]: E0313 12:56:00.170542 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156b689a-7e5a-4335-b861-6470b6c336e9" containerName="registry-server" Mar 13 12:56:00 crc kubenswrapper[4632]: I0313 12:56:00.170564 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="156b689a-7e5a-4335-b861-6470b6c336e9" containerName="registry-server" Mar 13 12:56:00 crc kubenswrapper[4632]: E0313 12:56:00.170614 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156b689a-7e5a-4335-b861-6470b6c336e9" containerName="extract-content" Mar 13 12:56:00 crc kubenswrapper[4632]: I0313 12:56:00.170621 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="156b689a-7e5a-4335-b861-6470b6c336e9" containerName="extract-content" Mar 13 12:56:00 crc kubenswrapper[4632]: E0313 12:56:00.170633 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="156b689a-7e5a-4335-b861-6470b6c336e9" containerName="extract-utilities" Mar 13 12:56:00 crc kubenswrapper[4632]: I0313 12:56:00.170640 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="156b689a-7e5a-4335-b861-6470b6c336e9" containerName="extract-utilities" Mar 13 12:56:00 crc kubenswrapper[4632]: I0313 12:56:00.170850 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="156b689a-7e5a-4335-b861-6470b6c336e9" containerName="registry-server" Mar 13 12:56:00 crc kubenswrapper[4632]: I0313 12:56:00.172774 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556776-spgrm" Mar 13 12:56:00 crc kubenswrapper[4632]: I0313 12:56:00.180923 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 12:56:00 crc kubenswrapper[4632]: I0313 12:56:00.181100 4632 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 12:56:00 crc kubenswrapper[4632]: I0313 12:56:00.181196 4632 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-kdp2p" Mar 13 12:56:00 crc kubenswrapper[4632]: I0313 12:56:00.201562 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556776-spgrm"] Mar 13 12:56:00 crc kubenswrapper[4632]: I0313 12:56:00.308020 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrh69\" (UniqueName: \"kubernetes.io/projected/549f75b2-26cb-4f82-a218-67acaaa1b9e3-kube-api-access-zrh69\") pod \"auto-csr-approver-29556776-spgrm\" (UID: \"549f75b2-26cb-4f82-a218-67acaaa1b9e3\") " pod="openshift-infra/auto-csr-approver-29556776-spgrm" Mar 13 12:56:00 crc kubenswrapper[4632]: I0313 12:56:00.410562 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrh69\" (UniqueName: \"kubernetes.io/projected/549f75b2-26cb-4f82-a218-67acaaa1b9e3-kube-api-access-zrh69\") pod \"auto-csr-approver-29556776-spgrm\" (UID: \"549f75b2-26cb-4f82-a218-67acaaa1b9e3\") " pod="openshift-infra/auto-csr-approver-29556776-spgrm" Mar 13 12:56:00 crc kubenswrapper[4632]: I0313 12:56:00.856688 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrh69\" (UniqueName: \"kubernetes.io/projected/549f75b2-26cb-4f82-a218-67acaaa1b9e3-kube-api-access-zrh69\") pod \"auto-csr-approver-29556776-spgrm\" (UID: \"549f75b2-26cb-4f82-a218-67acaaa1b9e3\") " pod="openshift-infra/auto-csr-approver-29556776-spgrm" Mar 13 12:56:01 crc kubenswrapper[4632]: I0313 12:56:01.095570 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556776-spgrm" Mar 13 12:56:01 crc kubenswrapper[4632]: I0313 12:56:01.622529 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556776-spgrm"] Mar 13 12:56:02 crc kubenswrapper[4632]: I0313 12:56:02.634792 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556776-spgrm" event={"ID":"549f75b2-26cb-4f82-a218-67acaaa1b9e3","Type":"ContainerStarted","Data":"4ea53b1a7c43ee5fd12084b1457f72787b5b467fa56de561bec07c5b1e454988"} Mar 13 12:56:03 crc kubenswrapper[4632]: I0313 12:56:03.648915 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556776-spgrm" event={"ID":"549f75b2-26cb-4f82-a218-67acaaa1b9e3","Type":"ContainerStarted","Data":"7a0a83ef3d6f48ba18a3a27e062b3429f8776838b6789c4b6baea6ff1a4cd532"} Mar 13 12:56:03 crc kubenswrapper[4632]: I0313 12:56:03.668445 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556776-spgrm" podStartSLOduration=2.85809333 podStartE2EDuration="3.668422489s" podCreationTimestamp="2026-03-13 12:56:00 +0000 UTC" firstStartedPulling="2026-03-13 12:56:01.643628792 +0000 UTC m=+10335.666158965" lastFinishedPulling="2026-03-13 12:56:02.453957991 +0000 UTC m=+10336.476488124" observedRunningTime="2026-03-13 12:56:03.663470217 +0000 UTC m=+10337.686000340" watchObservedRunningTime="2026-03-13 12:56:03.668422489 +0000 UTC m=+10337.690952622" Mar 13 12:56:05 crc kubenswrapper[4632]: I0313 12:56:05.670291 4632 generic.go:334] "Generic (PLEG): container finished" podID="549f75b2-26cb-4f82-a218-67acaaa1b9e3" containerID="7a0a83ef3d6f48ba18a3a27e062b3429f8776838b6789c4b6baea6ff1a4cd532" exitCode=0 Mar 13 12:56:05 crc kubenswrapper[4632]: I0313 12:56:05.670396 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556776-spgrm" event={"ID":"549f75b2-26cb-4f82-a218-67acaaa1b9e3","Type":"ContainerDied","Data":"7a0a83ef3d6f48ba18a3a27e062b3429f8776838b6789c4b6baea6ff1a4cd532"} Mar 13 12:56:07 crc kubenswrapper[4632]: I0313 12:56:07.164006 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556776-spgrm" Mar 13 12:56:07 crc kubenswrapper[4632]: I0313 12:56:07.296301 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrh69\" (UniqueName: \"kubernetes.io/projected/549f75b2-26cb-4f82-a218-67acaaa1b9e3-kube-api-access-zrh69\") pod \"549f75b2-26cb-4f82-a218-67acaaa1b9e3\" (UID: \"549f75b2-26cb-4f82-a218-67acaaa1b9e3\") " Mar 13 12:56:07 crc kubenswrapper[4632]: I0313 12:56:07.305345 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/549f75b2-26cb-4f82-a218-67acaaa1b9e3-kube-api-access-zrh69" (OuterVolumeSpecName: "kube-api-access-zrh69") pod "549f75b2-26cb-4f82-a218-67acaaa1b9e3" (UID: "549f75b2-26cb-4f82-a218-67acaaa1b9e3"). InnerVolumeSpecName "kube-api-access-zrh69". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:56:07 crc kubenswrapper[4632]: I0313 12:56:07.403142 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrh69\" (UniqueName: \"kubernetes.io/projected/549f75b2-26cb-4f82-a218-67acaaa1b9e3-kube-api-access-zrh69\") on node \"crc\" DevicePath \"\"" Mar 13 12:56:07 crc kubenswrapper[4632]: I0313 12:56:07.713259 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556776-spgrm" event={"ID":"549f75b2-26cb-4f82-a218-67acaaa1b9e3","Type":"ContainerDied","Data":"4ea53b1a7c43ee5fd12084b1457f72787b5b467fa56de561bec07c5b1e454988"} Mar 13 12:56:07 crc kubenswrapper[4632]: I0313 12:56:07.713323 4632 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4ea53b1a7c43ee5fd12084b1457f72787b5b467fa56de561bec07c5b1e454988" Mar 13 12:56:07 crc kubenswrapper[4632]: I0313 12:56:07.713406 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556776-spgrm" Mar 13 12:56:07 crc kubenswrapper[4632]: I0313 12:56:07.787148 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556770-clndp"] Mar 13 12:56:07 crc kubenswrapper[4632]: I0313 12:56:07.798219 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556770-clndp"] Mar 13 12:56:08 crc kubenswrapper[4632]: I0313 12:56:08.068768 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40d97b55-9e4c-4d9a-962c-4030dc7dd36b" path="/var/lib/kubelet/pods/40d97b55-9e4c-4d9a-962c-4030dc7dd36b/volumes" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.154144 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wwhf9"] Mar 13 12:56:15 crc kubenswrapper[4632]: E0313 12:56:15.155067 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="549f75b2-26cb-4f82-a218-67acaaa1b9e3" containerName="oc" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.155081 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="549f75b2-26cb-4f82-a218-67acaaa1b9e3" containerName="oc" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.155297 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="549f75b2-26cb-4f82-a218-67acaaa1b9e3" containerName="oc" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.157847 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.180661 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwhf9"] Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.293834 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dc7801b-a904-42ac-a8e4-02599a43e787-catalog-content\") pod \"redhat-marketplace-wwhf9\" (UID: \"1dc7801b-a904-42ac-a8e4-02599a43e787\") " pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.293907 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp4jx\" (UniqueName: \"kubernetes.io/projected/1dc7801b-a904-42ac-a8e4-02599a43e787-kube-api-access-wp4jx\") pod \"redhat-marketplace-wwhf9\" (UID: \"1dc7801b-a904-42ac-a8e4-02599a43e787\") " pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.294042 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dc7801b-a904-42ac-a8e4-02599a43e787-utilities\") pod \"redhat-marketplace-wwhf9\" (UID: \"1dc7801b-a904-42ac-a8e4-02599a43e787\") " pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.396134 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dc7801b-a904-42ac-a8e4-02599a43e787-catalog-content\") pod \"redhat-marketplace-wwhf9\" (UID: \"1dc7801b-a904-42ac-a8e4-02599a43e787\") " pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.396236 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp4jx\" (UniqueName: \"kubernetes.io/projected/1dc7801b-a904-42ac-a8e4-02599a43e787-kube-api-access-wp4jx\") pod \"redhat-marketplace-wwhf9\" (UID: \"1dc7801b-a904-42ac-a8e4-02599a43e787\") " pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.396329 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dc7801b-a904-42ac-a8e4-02599a43e787-utilities\") pod \"redhat-marketplace-wwhf9\" (UID: \"1dc7801b-a904-42ac-a8e4-02599a43e787\") " pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.396649 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dc7801b-a904-42ac-a8e4-02599a43e787-catalog-content\") pod \"redhat-marketplace-wwhf9\" (UID: \"1dc7801b-a904-42ac-a8e4-02599a43e787\") " pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.396870 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dc7801b-a904-42ac-a8e4-02599a43e787-utilities\") pod \"redhat-marketplace-wwhf9\" (UID: \"1dc7801b-a904-42ac-a8e4-02599a43e787\") " pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.430389 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp4jx\" (UniqueName: \"kubernetes.io/projected/1dc7801b-a904-42ac-a8e4-02599a43e787-kube-api-access-wp4jx\") pod \"redhat-marketplace-wwhf9\" (UID: \"1dc7801b-a904-42ac-a8e4-02599a43e787\") " pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.478491 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:15 crc kubenswrapper[4632]: I0313 12:56:15.808435 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwhf9"] Mar 13 12:56:16 crc kubenswrapper[4632]: I0313 12:56:16.818198 4632 generic.go:334] "Generic (PLEG): container finished" podID="1dc7801b-a904-42ac-a8e4-02599a43e787" containerID="d02ca469feda329ba8cfeef0af360635a4fdac24dcc38577abbc928f873e33ba" exitCode=0 Mar 13 12:56:16 crc kubenswrapper[4632]: I0313 12:56:16.818318 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwhf9" event={"ID":"1dc7801b-a904-42ac-a8e4-02599a43e787","Type":"ContainerDied","Data":"d02ca469feda329ba8cfeef0af360635a4fdac24dcc38577abbc928f873e33ba"} Mar 13 12:56:16 crc kubenswrapper[4632]: I0313 12:56:16.818520 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwhf9" event={"ID":"1dc7801b-a904-42ac-a8e4-02599a43e787","Type":"ContainerStarted","Data":"651a2a7181ea09e8fd2c5827d756706cc7c40dcd69e459a11ac096ef093b12ef"} Mar 13 12:56:17 crc kubenswrapper[4632]: I0313 12:56:17.832363 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwhf9" event={"ID":"1dc7801b-a904-42ac-a8e4-02599a43e787","Type":"ContainerStarted","Data":"804a3456daf22141fbdc0415c7d68b1fcf76e64fe47d38454a2cdba5fbb3fb68"} Mar 13 12:56:19 crc kubenswrapper[4632]: I0313 12:56:19.856097 4632 generic.go:334] "Generic (PLEG): container finished" podID="1dc7801b-a904-42ac-a8e4-02599a43e787" containerID="804a3456daf22141fbdc0415c7d68b1fcf76e64fe47d38454a2cdba5fbb3fb68" exitCode=0 Mar 13 12:56:19 crc kubenswrapper[4632]: I0313 12:56:19.856179 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwhf9" event={"ID":"1dc7801b-a904-42ac-a8e4-02599a43e787","Type":"ContainerDied","Data":"804a3456daf22141fbdc0415c7d68b1fcf76e64fe47d38454a2cdba5fbb3fb68"} Mar 13 12:56:20 crc kubenswrapper[4632]: I0313 12:56:20.873112 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwhf9" event={"ID":"1dc7801b-a904-42ac-a8e4-02599a43e787","Type":"ContainerStarted","Data":"1f8a7d6a071ff8a13935e5d10768c1d8b064693a71d7f9487aac2d2e31fab110"} Mar 13 12:56:20 crc kubenswrapper[4632]: I0313 12:56:20.935273 4632 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wwhf9" podStartSLOduration=2.465306369 podStartE2EDuration="5.935251795s" podCreationTimestamp="2026-03-13 12:56:15 +0000 UTC" firstStartedPulling="2026-03-13 12:56:16.820032502 +0000 UTC m=+10350.842562635" lastFinishedPulling="2026-03-13 12:56:20.289977928 +0000 UTC m=+10354.312508061" observedRunningTime="2026-03-13 12:56:20.902166544 +0000 UTC m=+10354.924696687" watchObservedRunningTime="2026-03-13 12:56:20.935251795 +0000 UTC m=+10354.957781928" Mar 13 12:56:25 crc kubenswrapper[4632]: I0313 12:56:25.479149 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:25 crc kubenswrapper[4632]: I0313 12:56:25.479599 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:26 crc kubenswrapper[4632]: I0313 12:56:26.526754 4632 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-wwhf9" podUID="1dc7801b-a904-42ac-a8e4-02599a43e787" containerName="registry-server" probeResult="failure" output=< Mar 13 12:56:26 crc kubenswrapper[4632]: timeout: failed to connect service ":50051" within 1s Mar 13 12:56:26 crc kubenswrapper[4632]: > Mar 13 12:56:31 crc kubenswrapper[4632]: I0313 12:56:31.753466 4632 scope.go:117] "RemoveContainer" containerID="3a1743912c3055d81c796b10e35eb2de0472d47edbe9b1cc8de66ceb54f1f127" Mar 13 12:56:35 crc kubenswrapper[4632]: I0313 12:56:35.550807 4632 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:35 crc kubenswrapper[4632]: I0313 12:56:35.619605 4632 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:35 crc kubenswrapper[4632]: I0313 12:56:35.792368 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwhf9"] Mar 13 12:56:37 crc kubenswrapper[4632]: I0313 12:56:37.083560 4632 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wwhf9" podUID="1dc7801b-a904-42ac-a8e4-02599a43e787" containerName="registry-server" containerID="cri-o://1f8a7d6a071ff8a13935e5d10768c1d8b064693a71d7f9487aac2d2e31fab110" gracePeriod=2 Mar 13 12:56:37 crc kubenswrapper[4632]: I0313 12:56:37.591061 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:37 crc kubenswrapper[4632]: I0313 12:56:37.742865 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dc7801b-a904-42ac-a8e4-02599a43e787-utilities\") pod \"1dc7801b-a904-42ac-a8e4-02599a43e787\" (UID: \"1dc7801b-a904-42ac-a8e4-02599a43e787\") " Mar 13 12:56:37 crc kubenswrapper[4632]: I0313 12:56:37.743063 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dc7801b-a904-42ac-a8e4-02599a43e787-catalog-content\") pod \"1dc7801b-a904-42ac-a8e4-02599a43e787\" (UID: \"1dc7801b-a904-42ac-a8e4-02599a43e787\") " Mar 13 12:56:37 crc kubenswrapper[4632]: I0313 12:56:37.743173 4632 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wp4jx\" (UniqueName: \"kubernetes.io/projected/1dc7801b-a904-42ac-a8e4-02599a43e787-kube-api-access-wp4jx\") pod \"1dc7801b-a904-42ac-a8e4-02599a43e787\" (UID: \"1dc7801b-a904-42ac-a8e4-02599a43e787\") " Mar 13 12:56:37 crc kubenswrapper[4632]: I0313 12:56:37.745999 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dc7801b-a904-42ac-a8e4-02599a43e787-utilities" (OuterVolumeSpecName: "utilities") pod "1dc7801b-a904-42ac-a8e4-02599a43e787" (UID: "1dc7801b-a904-42ac-a8e4-02599a43e787"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:56:37 crc kubenswrapper[4632]: I0313 12:56:37.753575 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1dc7801b-a904-42ac-a8e4-02599a43e787-kube-api-access-wp4jx" (OuterVolumeSpecName: "kube-api-access-wp4jx") pod "1dc7801b-a904-42ac-a8e4-02599a43e787" (UID: "1dc7801b-a904-42ac-a8e4-02599a43e787"). InnerVolumeSpecName "kube-api-access-wp4jx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 12:56:37 crc kubenswrapper[4632]: I0313 12:56:37.811534 4632 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1dc7801b-a904-42ac-a8e4-02599a43e787-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1dc7801b-a904-42ac-a8e4-02599a43e787" (UID: "1dc7801b-a904-42ac-a8e4-02599a43e787"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 12:56:37 crc kubenswrapper[4632]: I0313 12:56:37.846306 4632 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1dc7801b-a904-42ac-a8e4-02599a43e787-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 12:56:37 crc kubenswrapper[4632]: I0313 12:56:37.846728 4632 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1dc7801b-a904-42ac-a8e4-02599a43e787-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 12:56:37 crc kubenswrapper[4632]: I0313 12:56:37.846931 4632 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wp4jx\" (UniqueName: \"kubernetes.io/projected/1dc7801b-a904-42ac-a8e4-02599a43e787-kube-api-access-wp4jx\") on node \"crc\" DevicePath \"\"" Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.095135 4632 generic.go:334] "Generic (PLEG): container finished" podID="1dc7801b-a904-42ac-a8e4-02599a43e787" containerID="1f8a7d6a071ff8a13935e5d10768c1d8b064693a71d7f9487aac2d2e31fab110" exitCode=0 Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.096518 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwhf9" event={"ID":"1dc7801b-a904-42ac-a8e4-02599a43e787","Type":"ContainerDied","Data":"1f8a7d6a071ff8a13935e5d10768c1d8b064693a71d7f9487aac2d2e31fab110"} Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.096840 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wwhf9" event={"ID":"1dc7801b-a904-42ac-a8e4-02599a43e787","Type":"ContainerDied","Data":"651a2a7181ea09e8fd2c5827d756706cc7c40dcd69e459a11ac096ef093b12ef"} Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.099689 4632 scope.go:117] "RemoveContainer" containerID="1f8a7d6a071ff8a13935e5d10768c1d8b064693a71d7f9487aac2d2e31fab110" Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.100279 4632 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wwhf9" Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.136778 4632 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwhf9"] Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.144489 4632 scope.go:117] "RemoveContainer" containerID="804a3456daf22141fbdc0415c7d68b1fcf76e64fe47d38454a2cdba5fbb3fb68" Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.149715 4632 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wwhf9"] Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.167424 4632 scope.go:117] "RemoveContainer" containerID="d02ca469feda329ba8cfeef0af360635a4fdac24dcc38577abbc928f873e33ba" Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.238205 4632 scope.go:117] "RemoveContainer" containerID="1f8a7d6a071ff8a13935e5d10768c1d8b064693a71d7f9487aac2d2e31fab110" Mar 13 12:56:38 crc kubenswrapper[4632]: E0313 12:56:38.238825 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f8a7d6a071ff8a13935e5d10768c1d8b064693a71d7f9487aac2d2e31fab110\": container with ID starting with 1f8a7d6a071ff8a13935e5d10768c1d8b064693a71d7f9487aac2d2e31fab110 not found: ID does not exist" containerID="1f8a7d6a071ff8a13935e5d10768c1d8b064693a71d7f9487aac2d2e31fab110" Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.238868 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f8a7d6a071ff8a13935e5d10768c1d8b064693a71d7f9487aac2d2e31fab110"} err="failed to get container status \"1f8a7d6a071ff8a13935e5d10768c1d8b064693a71d7f9487aac2d2e31fab110\": rpc error: code = NotFound desc = could not find container \"1f8a7d6a071ff8a13935e5d10768c1d8b064693a71d7f9487aac2d2e31fab110\": container with ID starting with 1f8a7d6a071ff8a13935e5d10768c1d8b064693a71d7f9487aac2d2e31fab110 not found: ID does not exist" Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.238893 4632 scope.go:117] "RemoveContainer" containerID="804a3456daf22141fbdc0415c7d68b1fcf76e64fe47d38454a2cdba5fbb3fb68" Mar 13 12:56:38 crc kubenswrapper[4632]: E0313 12:56:38.239466 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"804a3456daf22141fbdc0415c7d68b1fcf76e64fe47d38454a2cdba5fbb3fb68\": container with ID starting with 804a3456daf22141fbdc0415c7d68b1fcf76e64fe47d38454a2cdba5fbb3fb68 not found: ID does not exist" containerID="804a3456daf22141fbdc0415c7d68b1fcf76e64fe47d38454a2cdba5fbb3fb68" Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.239491 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"804a3456daf22141fbdc0415c7d68b1fcf76e64fe47d38454a2cdba5fbb3fb68"} err="failed to get container status \"804a3456daf22141fbdc0415c7d68b1fcf76e64fe47d38454a2cdba5fbb3fb68\": rpc error: code = NotFound desc = could not find container \"804a3456daf22141fbdc0415c7d68b1fcf76e64fe47d38454a2cdba5fbb3fb68\": container with ID starting with 804a3456daf22141fbdc0415c7d68b1fcf76e64fe47d38454a2cdba5fbb3fb68 not found: ID does not exist" Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.239507 4632 scope.go:117] "RemoveContainer" containerID="d02ca469feda329ba8cfeef0af360635a4fdac24dcc38577abbc928f873e33ba" Mar 13 12:56:38 crc kubenswrapper[4632]: E0313 12:56:38.239925 4632 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d02ca469feda329ba8cfeef0af360635a4fdac24dcc38577abbc928f873e33ba\": container with ID starting with d02ca469feda329ba8cfeef0af360635a4fdac24dcc38577abbc928f873e33ba not found: ID does not exist" containerID="d02ca469feda329ba8cfeef0af360635a4fdac24dcc38577abbc928f873e33ba" Mar 13 12:56:38 crc kubenswrapper[4632]: I0313 12:56:38.239990 4632 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d02ca469feda329ba8cfeef0af360635a4fdac24dcc38577abbc928f873e33ba"} err="failed to get container status \"d02ca469feda329ba8cfeef0af360635a4fdac24dcc38577abbc928f873e33ba\": rpc error: code = NotFound desc = could not find container \"d02ca469feda329ba8cfeef0af360635a4fdac24dcc38577abbc928f873e33ba\": container with ID starting with d02ca469feda329ba8cfeef0af360635a4fdac24dcc38577abbc928f873e33ba not found: ID does not exist" Mar 13 12:56:40 crc kubenswrapper[4632]: I0313 12:56:40.069232 4632 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1dc7801b-a904-42ac-a8e4-02599a43e787" path="/var/lib/kubelet/pods/1dc7801b-a904-42ac-a8e4-02599a43e787/volumes" Mar 13 12:56:40 crc kubenswrapper[4632]: I0313 12:56:40.460857 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:56:40 crc kubenswrapper[4632]: I0313 12:56:40.460998 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:57:10 crc kubenswrapper[4632]: I0313 12:57:10.460729 4632 patch_prober.go:28] interesting pod/machine-config-daemon-zkscb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 12:57:10 crc kubenswrapper[4632]: I0313 12:57:10.461348 4632 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-zkscb" podUID="d77b18a7-7ad9-4bf5-bff5-da45878af7f4" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.160700 4632 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-plp7p"] Mar 13 12:57:17 crc kubenswrapper[4632]: E0313 12:57:17.161481 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dc7801b-a904-42ac-a8e4-02599a43e787" containerName="extract-content" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.161494 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dc7801b-a904-42ac-a8e4-02599a43e787" containerName="extract-content" Mar 13 12:57:17 crc kubenswrapper[4632]: E0313 12:57:17.161530 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dc7801b-a904-42ac-a8e4-02599a43e787" containerName="extract-utilities" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.161536 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dc7801b-a904-42ac-a8e4-02599a43e787" containerName="extract-utilities" Mar 13 12:57:17 crc kubenswrapper[4632]: E0313 12:57:17.161548 4632 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1dc7801b-a904-42ac-a8e4-02599a43e787" containerName="registry-server" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.161554 4632 state_mem.go:107] "Deleted CPUSet assignment" podUID="1dc7801b-a904-42ac-a8e4-02599a43e787" containerName="registry-server" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.161739 4632 memory_manager.go:354] "RemoveStaleState removing state" podUID="1dc7801b-a904-42ac-a8e4-02599a43e787" containerName="registry-server" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.163192 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-plp7p" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.179616 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-plp7p"] Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.316194 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95kmr\" (UniqueName: \"kubernetes.io/projected/776af28c-b8fa-4956-92cc-4e8742b45977-kube-api-access-95kmr\") pod \"redhat-operators-plp7p\" (UID: \"776af28c-b8fa-4956-92cc-4e8742b45977\") " pod="openshift-marketplace/redhat-operators-plp7p" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.316245 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/776af28c-b8fa-4956-92cc-4e8742b45977-catalog-content\") pod \"redhat-operators-plp7p\" (UID: \"776af28c-b8fa-4956-92cc-4e8742b45977\") " pod="openshift-marketplace/redhat-operators-plp7p" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.316280 4632 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/776af28c-b8fa-4956-92cc-4e8742b45977-utilities\") pod \"redhat-operators-plp7p\" (UID: \"776af28c-b8fa-4956-92cc-4e8742b45977\") " pod="openshift-marketplace/redhat-operators-plp7p" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.418472 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/776af28c-b8fa-4956-92cc-4e8742b45977-utilities\") pod \"redhat-operators-plp7p\" (UID: \"776af28c-b8fa-4956-92cc-4e8742b45977\") " pod="openshift-marketplace/redhat-operators-plp7p" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.418754 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95kmr\" (UniqueName: \"kubernetes.io/projected/776af28c-b8fa-4956-92cc-4e8742b45977-kube-api-access-95kmr\") pod \"redhat-operators-plp7p\" (UID: \"776af28c-b8fa-4956-92cc-4e8742b45977\") " pod="openshift-marketplace/redhat-operators-plp7p" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.418812 4632 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/776af28c-b8fa-4956-92cc-4e8742b45977-catalog-content\") pod \"redhat-operators-plp7p\" (UID: \"776af28c-b8fa-4956-92cc-4e8742b45977\") " pod="openshift-marketplace/redhat-operators-plp7p" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.419327 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/776af28c-b8fa-4956-92cc-4e8742b45977-catalog-content\") pod \"redhat-operators-plp7p\" (UID: \"776af28c-b8fa-4956-92cc-4e8742b45977\") " pod="openshift-marketplace/redhat-operators-plp7p" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.419809 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/776af28c-b8fa-4956-92cc-4e8742b45977-utilities\") pod \"redhat-operators-plp7p\" (UID: \"776af28c-b8fa-4956-92cc-4e8742b45977\") " pod="openshift-marketplace/redhat-operators-plp7p" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.445891 4632 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95kmr\" (UniqueName: \"kubernetes.io/projected/776af28c-b8fa-4956-92cc-4e8742b45977-kube-api-access-95kmr\") pod \"redhat-operators-plp7p\" (UID: \"776af28c-b8fa-4956-92cc-4e8742b45977\") " pod="openshift-marketplace/redhat-operators-plp7p" Mar 13 12:57:17 crc kubenswrapper[4632]: I0313 12:57:17.534682 4632 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-plp7p" Mar 13 12:57:18 crc kubenswrapper[4632]: I0313 12:57:18.033247 4632 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-plp7p"] Mar 13 12:57:18 crc kubenswrapper[4632]: I0313 12:57:18.492707 4632 generic.go:334] "Generic (PLEG): container finished" podID="776af28c-b8fa-4956-92cc-4e8742b45977" containerID="dcbd515594a90692f579837b518369662d87fdcc1100cc624dd7853a1262878f" exitCode=0 Mar 13 12:57:18 crc kubenswrapper[4632]: I0313 12:57:18.492783 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-plp7p" event={"ID":"776af28c-b8fa-4956-92cc-4e8742b45977","Type":"ContainerDied","Data":"dcbd515594a90692f579837b518369662d87fdcc1100cc624dd7853a1262878f"} Mar 13 12:57:18 crc kubenswrapper[4632]: I0313 12:57:18.492834 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-plp7p" event={"ID":"776af28c-b8fa-4956-92cc-4e8742b45977","Type":"ContainerStarted","Data":"353de2a55edf635455d1dd8c35a1d255a22a39f9ba2c6050920d89923c1a17af"} Mar 13 12:57:20 crc kubenswrapper[4632]: I0313 12:57:20.513548 4632 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-plp7p" event={"ID":"776af28c-b8fa-4956-92cc-4e8742b45977","Type":"ContainerStarted","Data":"2381a4bf987cf45aa5b719a9c28119038c5eb308dc2dcdc26fa692cff58869e9"} var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515155004676024456 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015155004677017374 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015154760033016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015154760033015461 5ustar corecore